Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
855
| labels
stringlengths 4
721
| body
stringlengths 1
261k
| index
stringclasses 13
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
240k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
47,743
| 2,984,925,863
|
IssuesEvent
|
2015-07-18 13:52:09
|
AmatCoder/mednaffe
|
https://api.github.com/repos/AmatCoder/mednaffe
|
reopened
|
[BUG] XInput maps always Axis - deadzones required
|
bug Priority-High
|
Microsoft gamepads are crappy so they need deadzone for axis and triggers, Xinput.h defines them:
#define XINPUT_GAMEPAD_LEFT_THUMB_DEADZONE 7849
#define XINPUT_GAMEPAD_RIGHT_THUMB_DEADZONE 8689
#define XINPUT_GAMEPAD_TRIGGER_THRESHOLD 30
I have mapping problem because of this ie. axis movement is reported as 'key pressed' in PSX Input mapping.
|
1.0
|
[BUG] XInput maps always Axis - deadzones required - Microsoft gamepads are crappy so they need deadzone for axis and triggers, Xinput.h defines them:
#define XINPUT_GAMEPAD_LEFT_THUMB_DEADZONE 7849
#define XINPUT_GAMEPAD_RIGHT_THUMB_DEADZONE 8689
#define XINPUT_GAMEPAD_TRIGGER_THRESHOLD 30
I have mapping problem because of this ie. axis movement is reported as 'key pressed' in PSX Input mapping.
|
priority
|
xinput maps always axis deadzones required microsoft gamepads are crappy so they need deadzone for axis and triggers xinput h defines them define xinput gamepad left thumb deadzone define xinput gamepad right thumb deadzone define xinput gamepad trigger threshold i have mapping problem because of this ie axis movement is reported as key pressed in psx input mapping
| 1
|
221,707
| 7,394,393,840
|
IssuesEvent
|
2018-03-17 10:32:38
|
glutanimate/image-occlusion-enhanced
|
https://api.github.com/repos/glutanimate/image-occlusion-enhanced
|
closed
|
Switching to a different note type from I/O causing the first field to be hidden
|
anki2.1 high priority regression
|
Steps to reproduce:
1. Open Anki's "Add" screen
2. Change the note type to "Image Occlusion Enhanced"
3. Switch to a different note type (e.g. basic)
Expected behavior (present under Anki 2.0):
- Image Occlusion Enhanced note type: first field ("ID (hidden)" is hidden from user)
- Basic note type: All fields visible
Observed behavior (present under Anki 2.1):
- Image Occlusion Enhanced note type: first field ("ID (hidden)" is hidden from user)
- Basic note type: **First field hidden**
Might be a good opportunity to discuss whether Anki should offer a way for add-on authors to mark specific note type fields as hidden (could be useful for any note type that is generated programmatically and contains fields that should not be user-editable).
|
1.0
|
Switching to a different note type from I/O causing the first field to be hidden - Steps to reproduce:
1. Open Anki's "Add" screen
2. Change the note type to "Image Occlusion Enhanced"
3. Switch to a different note type (e.g. basic)
Expected behavior (present under Anki 2.0):
- Image Occlusion Enhanced note type: first field ("ID (hidden)" is hidden from user)
- Basic note type: All fields visible
Observed behavior (present under Anki 2.1):
- Image Occlusion Enhanced note type: first field ("ID (hidden)" is hidden from user)
- Basic note type: **First field hidden**
Might be a good opportunity to discuss whether Anki should offer a way for add-on authors to mark specific note type fields as hidden (could be useful for any note type that is generated programmatically and contains fields that should not be user-editable).
|
priority
|
switching to a different note type from i o causing the first field to be hidden steps to reproduce open anki s add screen change the note type to image occlusion enhanced switch to a different note type e g basic expected behavior present under anki image occlusion enhanced note type first field id hidden is hidden from user basic note type all fields visible observed behavior present under anki image occlusion enhanced note type first field id hidden is hidden from user basic note type first field hidden might be a good opportunity to discuss whether anki should offer a way for add on authors to mark specific note type fields as hidden could be useful for any note type that is generated programmatically and contains fields that should not be user editable
| 1
|
24,836
| 2,673,783,993
|
IssuesEvent
|
2015-03-24 21:10:55
|
pufexi/multiorder
|
https://api.github.com/repos/pufexi/multiorder
|
closed
|
!!! zadruhe !!!! Číslo zásilky (tracking)
|
high priority
|
Tohle udelame rovnou s tim importem tech cisel zasilek... at se s tim nemazem na vickrat.
V priloze zasilam zasilkyPodaniExport.csv.png , smaz to png a mas CSV , v tom je na prvnim miste (Excel sloupec A) cislo zasilky a na sloupci W 1413996795 , to je variabilni symbo, pro nas cislo objednavky, cele v UTF-8. Podle tohodle to budes parovat.
Samotne cislo zasilky bych dal za stat, tedy "Frýdek-Místek, CZ , BA6950905034M" a bylo by prokliknutelne na https://www.postaonline.cz/trackandtrace/-/zasilka/cislo?parcelNumbers=BA6950905034M v novem okne. Netreba zadne editace nikde, stava se zridka a editoval bych v DB pres phpmyadmina, ostatne takhle jsem zvyklej nejaky extemy doladit.
Pak nekde musi byt ten formular, kam budu vkladat ten soubor pro import... udelej to treba doleva jak jsou ty Objednavky otevrene, tak treba dvojta cara ci neco a dat tam "Import tracking Čpost"
Pri importu se zaroven odesle email zakaznikovi, to si tam nejak priprav, ale ocenime to v jinem Issue, kde budem resit notifikaci zakaznika.
VIZ SOUBOR: zasilkypodaniexport.csv
|
1.0
|
!!! zadruhe !!!! Číslo zásilky (tracking) - Tohle udelame rovnou s tim importem tech cisel zasilek... at se s tim nemazem na vickrat.
V priloze zasilam zasilkyPodaniExport.csv.png , smaz to png a mas CSV , v tom je na prvnim miste (Excel sloupec A) cislo zasilky a na sloupci W 1413996795 , to je variabilni symbo, pro nas cislo objednavky, cele v UTF-8. Podle tohodle to budes parovat.
Samotne cislo zasilky bych dal za stat, tedy "Frýdek-Místek, CZ , BA6950905034M" a bylo by prokliknutelne na https://www.postaonline.cz/trackandtrace/-/zasilka/cislo?parcelNumbers=BA6950905034M v novem okne. Netreba zadne editace nikde, stava se zridka a editoval bych v DB pres phpmyadmina, ostatne takhle jsem zvyklej nejaky extemy doladit.
Pak nekde musi byt ten formular, kam budu vkladat ten soubor pro import... udelej to treba doleva jak jsou ty Objednavky otevrene, tak treba dvojta cara ci neco a dat tam "Import tracking Čpost"
Pri importu se zaroven odesle email zakaznikovi, to si tam nejak priprav, ale ocenime to v jinem Issue, kde budem resit notifikaci zakaznika.
VIZ SOUBOR: zasilkypodaniexport.csv
|
priority
|
zadruhe číslo zásilky tracking tohle udelame rovnou s tim importem tech cisel zasilek at se s tim nemazem na vickrat v priloze zasilam zasilkypodaniexport csv png smaz to png a mas csv v tom je na prvnim miste excel sloupec a cislo zasilky a na sloupci w to je variabilni symbo pro nas cislo objednavky cele v utf podle tohodle to budes parovat samotne cislo zasilky bych dal za stat tedy frýdek místek cz a bylo by prokliknutelne na v novem okne netreba zadne editace nikde stava se zridka a editoval bych v db pres phpmyadmina ostatne takhle jsem zvyklej nejaky extemy doladit pak nekde musi byt ten formular kam budu vkladat ten soubor pro import udelej to treba doleva jak jsou ty objednavky otevrene tak treba dvojta cara ci neco a dat tam import tracking čpost pri importu se zaroven odesle email zakaznikovi to si tam nejak priprav ale ocenime to v jinem issue kde budem resit notifikaci zakaznika viz soubor zasilkypodaniexport csv
| 1
|
539,050
| 15,782,695,630
|
IssuesEvent
|
2021-04-01 13:07:17
|
eksperimental/ex_doc
|
https://api.github.com/repos/eksperimental/ex_doc
|
closed
|
add "(type)" to link title in menu
|
Priority:High enhancement
|
search for "has_key?/2"
and two `Dict.has_key?/2` will be displayed, without being able to know what is the difference.
MAybe we should put in parenthesis the type of anything that is not a macro or a function. (ie, callbacks)
|
1.0
|
add "(type)" to link title in menu - search for "has_key?/2"
and two `Dict.has_key?/2` will be displayed, without being able to know what is the difference.
MAybe we should put in parenthesis the type of anything that is not a macro or a function. (ie, callbacks)
|
priority
|
add type to link title in menu search for has key and two dict has key will be displayed without being able to know what is the difference maybe we should put in parenthesis the type of anything that is not a macro or a function ie callbacks
| 1
|
308,732
| 9,449,220,143
|
IssuesEvent
|
2019-04-16 00:52:33
|
smacademic/project-cgkm
|
https://api.github.com/repos/smacademic/project-cgkm
|
closed
|
Arc is missing due dates
|
priority - high severity - major type - planned feature
|
Arcs do not currently have due dates. This would be a simple addition in the arc.dart file and the databasehelper.dart file.
|
1.0
|
Arc is missing due dates - Arcs do not currently have due dates. This would be a simple addition in the arc.dart file and the databasehelper.dart file.
|
priority
|
arc is missing due dates arcs do not currently have due dates this would be a simple addition in the arc dart file and the databasehelper dart file
| 1
|
465,200
| 13,358,551,349
|
IssuesEvent
|
2020-08-31 11:54:55
|
DIAGNijmegen/website-content
|
https://api.github.com/repos/DIAGNijmegen/website-content
|
closed
|
Build redirects for transition from old to new website
|
Priority: High enhancement
|
- [x] People pages
- [x] Publication pages for people
- [x] Publications
- [x] Main pages
|
1.0
|
Build redirects for transition from old to new website - - [x] People pages
- [x] Publication pages for people
- [x] Publications
- [x] Main pages
|
priority
|
build redirects for transition from old to new website people pages publication pages for people publications main pages
| 1
|
135,336
| 5,246,900,348
|
IssuesEvent
|
2017-02-01 11:07:40
|
DigitalCampus/oppia-mobile-android
|
https://api.github.com/repos/DigitalCampus/oppia-mobile-android
|
opened
|
App crashing issues on password reset and video download
|
bug High priority
|
On both these tasks, the app crashes. Very similar to the login carshing we had the other day - so suspect that it may be the same cause/issue
|
1.0
|
App crashing issues on password reset and video download - On both these tasks, the app crashes. Very similar to the login carshing we had the other day - so suspect that it may be the same cause/issue
|
priority
|
app crashing issues on password reset and video download on both these tasks the app crashes very similar to the login carshing we had the other day so suspect that it may be the same cause issue
| 1
|
613,919
| 19,101,539,576
|
IssuesEvent
|
2021-11-29 23:22:06
|
CMPUT301F21T21/detes
|
https://api.github.com/repos/CMPUT301F21T21/detes
|
closed
|
US 05.01.01 - Habit Following and Sharing
|
Final checkpoint High Risk Medium Priority Updated
|
As a doer, I want to ask another doer to follow all their **public** habits.
**Clarification:** The user would like to ask another user to follow their progress on their publicly viewable habits
Story Points: 4
|
1.0
|
US 05.01.01 - Habit Following and Sharing - As a doer, I want to ask another doer to follow all their **public** habits.
**Clarification:** The user would like to ask another user to follow their progress on their publicly viewable habits
Story Points: 4
|
priority
|
us habit following and sharing as a doer i want to ask another doer to follow all their public habits clarification the user would like to ask another user to follow their progress on their publicly viewable habits story points
| 1
|
399,200
| 11,744,496,462
|
IssuesEvent
|
2020-03-12 07:51:54
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
support.mozilla.org - see bug description
|
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important
|
<!-- @browser: Firefox Mobile 75.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/50052 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://support.mozilla.org/zh-CN/kb/firefox-preview-upgrade-faqs
**Browser / Version**: Firefox Mobile 75.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: js和ua
**Steps to Reproduce**:
能否增加自定义网站ua与JavaScript禁止开启控制
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
support.mozilla.org - see bug description - <!-- @browser: Firefox Mobile 75.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/50052 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://support.mozilla.org/zh-CN/kb/firefox-preview-upgrade-faqs
**Browser / Version**: Firefox Mobile 75.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: js和ua
**Steps to Reproduce**:
能否增加自定义网站ua与JavaScript禁止开启控制
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
priority
|
support mozilla org see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description js和ua steps to reproduce 能否增加自定义网站ua与javascript禁止开启控制 browser configuration none from with ❤️
| 1
|
511,805
| 14,882,023,910
|
IssuesEvent
|
2021-01-20 11:17:45
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
opened
|
[0.9.2 staging-1907] Creating civics articles in Amendments breaks all Capitol articles , not only delete one.
|
Category: Laws Priority: High
|
Step to reproduce:
- create a new world, ratify constitution:

- check court to see tha I can create laws:

- place amedndments and start to create another article:

- /civics winelection. So now all articles instead of the new one were broken:



- I only can create Elected titles:

In government window I have all articles are active, and removed article is active one too:

|
1.0
|
[0.9.2 staging-1907] Creating civics articles in Amendments breaks all Capitol articles , not only delete one. - Step to reproduce:
- create a new world, ratify constitution:

- check court to see tha I can create laws:

- place amedndments and start to create another article:

- /civics winelection. So now all articles instead of the new one were broken:



- I only can create Elected titles:

In government window I have all articles are active, and removed article is active one too:

|
priority
|
creating civics articles in amendments breaks all capitol articles not only delete one step to reproduce create a new world ratify constitution check court to see tha i can create laws place amedndments and start to create another article civics winelection so now all articles instead of the new one were broken i only can create elected titles in government window i have all articles are active and removed article is active one too
| 1
|
798,169
| 28,238,656,753
|
IssuesEvent
|
2023-04-06 04:30:01
|
AY2223S2-CS2113-T15-4/tp
|
https://api.github.com/repos/AY2223S2-CS2113-T15-4/tp
|
closed
|
[PE-D][Tester C] Trying to update the question to one with a whitespace results in wrong error message
|
type.Bug priority.High severity.Medium
|
Throwing an error is correct in this case, however, the name of the error is misleading.
Steps to reproduce: key in `add /q What is the worst fruit? /a Durian` and then key in `update fruit`, press enter and then key in `1 /q ` with a whitespace.
Expected: Unable to update the question to one that has an empty input!
Actual:

This can mislead the user thinking that the selected index is wrong, but '1' is actually in fact correct.
<!--session: 1680252405098-9e85e652-f0de-43e6-b248-5a883a5bc5ba-->
<!--Version: Web v3.4.7-->
-------------
Labels: `severity.VeryLow` `type.FunctionalityBug`
original: denzelcjy/ped#4
|
1.0
|
[PE-D][Tester C] Trying to update the question to one with a whitespace results in wrong error message - Throwing an error is correct in this case, however, the name of the error is misleading.
Steps to reproduce: key in `add /q What is the worst fruit? /a Durian` and then key in `update fruit`, press enter and then key in `1 /q ` with a whitespace.
Expected: Unable to update the question to one that has an empty input!
Actual:

This can mislead the user thinking that the selected index is wrong, but '1' is actually in fact correct.
<!--session: 1680252405098-9e85e652-f0de-43e6-b248-5a883a5bc5ba-->
<!--Version: Web v3.4.7-->
-------------
Labels: `severity.VeryLow` `type.FunctionalityBug`
original: denzelcjy/ped#4
|
priority
|
trying to update the question to one with a whitespace results in wrong error message throwing an error is correct in this case however the name of the error is misleading steps to reproduce key in add q what is the worst fruit a durian and then key in update fruit press enter and then key in q with a whitespace expected unable to update the question to one that has an empty input actual this can mislead the user thinking that the selected index is wrong but is actually in fact correct labels severity verylow type functionalitybug original denzelcjy ped
| 1
|
191,197
| 6,826,811,446
|
IssuesEvent
|
2017-11-08 15:15:33
|
GRIS-UdeM/ServerGRIS
|
https://api.github.com/repos/GRIS-UdeM/ServerGRIS
|
closed
|
Ajouter HRTF
|
enhancement High priority
|
Pour mémoire, suite à la réunion du 11 septembre, ajouter le HRTF avant la date possible de sortie, le 1er novembre.
|
1.0
|
Ajouter HRTF - Pour mémoire, suite à la réunion du 11 septembre, ajouter le HRTF avant la date possible de sortie, le 1er novembre.
|
priority
|
ajouter hrtf pour mémoire suite à la réunion du septembre ajouter le hrtf avant la date possible de sortie le novembre
| 1
|
439,031
| 12,676,194,051
|
IssuesEvent
|
2020-06-19 04:19:45
|
ocaml/ocaml
|
https://api.github.com/repos/ocaml/ocaml
|
closed
|
ocamlmklib always adds -L (absolute) directories also the run-time linker path.
|
Stale bug high-priority tools
|
**Original bug ID:** 5943
**Reporter:** is
**Status:** acknowledged (set by @damiendoligez on 2013-06-19T18:34:33Z)
**Resolution:** open
**Priority:** high
**Severity:** minor
**Platform:** any
**OS:** any
**OS Version:** any
**Version:** 4.00.1
**Category:** tools (ocaml{lex,yacc,dep,debug,...})
**Tags:** patch
**Monitored by:** is @ygrek @hcarty
## Bug description
ocamlmklib contains this snippet:
else if starts_with s "-L" then
(c_Lopts := s :: !c_Lopts;
let l = chop_prefix s "-L" in
if not (Filename.is_relative l) then rpath := l :: !rpath)
This results in absolute paths always added to the run-time-path. This is wrong in any build environment where the object directory is accessed through an absolute path; when using -R, the wrong path is added along the right one.
Contrary, ELF linker tools always require explicit specification of the run-time path, even when the same.
I suggest removing
let l = chop_prefix s "-L" in
if not (Filename.is_relative l) then rpath := l :: !rpath)
If this behaviour is deemed necessary for backwards compatibility, the new one should at least be selectable by a global option to ocamlmklib.
## File attachments
- [patch-tools_ocamlmklib](https://gist.githubusercontent.com/vicuna/1562f94183302316bea164bbb7db9014/raw/c9a1b7bcfb09e724b4100fe8a4143def7b6ab471/patch-tools_ocamlmklib)
|
1.0
|
ocamlmklib always adds -L (absolute) directories also the run-time linker path. - **Original bug ID:** 5943
**Reporter:** is
**Status:** acknowledged (set by @damiendoligez on 2013-06-19T18:34:33Z)
**Resolution:** open
**Priority:** high
**Severity:** minor
**Platform:** any
**OS:** any
**OS Version:** any
**Version:** 4.00.1
**Category:** tools (ocaml{lex,yacc,dep,debug,...})
**Tags:** patch
**Monitored by:** is @ygrek @hcarty
## Bug description
ocamlmklib contains this snippet:
else if starts_with s "-L" then
(c_Lopts := s :: !c_Lopts;
let l = chop_prefix s "-L" in
if not (Filename.is_relative l) then rpath := l :: !rpath)
This results in absolute paths always added to the run-time-path. This is wrong in any build environment where the object directory is accessed through an absolute path; when using -R, the wrong path is added along the right one.
Contrary, ELF linker tools always require explicit specification of the run-time path, even when the same.
I suggest removing
let l = chop_prefix s "-L" in
if not (Filename.is_relative l) then rpath := l :: !rpath)
If this behaviour is deemed necessary for backwards compatibility, the new one should at least be selectable by a global option to ocamlmklib.
## File attachments
- [patch-tools_ocamlmklib](https://gist.githubusercontent.com/vicuna/1562f94183302316bea164bbb7db9014/raw/c9a1b7bcfb09e724b4100fe8a4143def7b6ab471/patch-tools_ocamlmklib)
|
priority
|
ocamlmklib always adds l absolute directories also the run time linker path original bug id reporter is status acknowledged set by damiendoligez on resolution open priority high severity minor platform any os any os version any version category tools ocaml lex yacc dep debug tags patch monitored by is ygrek hcarty bug description ocamlmklib contains this snippet else if starts with s l then c lopts s c lopts let l chop prefix s l in if not filename is relative l then rpath l rpath this results in absolute paths always added to the run time path this is wrong in any build environment where the object directory is accessed through an absolute path when using r the wrong path is added along the right one contrary elf linker tools always require explicit specification of the run time path even when the same i suggest removing let l chop prefix s l in if not filename is relative l then rpath l rpath if this behaviour is deemed necessary for backwards compatibility the new one should at least be selectable by a global option to ocamlmklib file attachments
| 1
|
333,315
| 10,120,459,467
|
IssuesEvent
|
2019-07-31 13:47:02
|
jncc/topcat
|
https://api.github.com/repos/jncc/topcat
|
closed
|
Providing users with links to Data Provider Agreements and License
|
high priority
|
For third party data, the Marine team is working towards the use of a _'Data Provider Agreement'_ which will capture specific restrictions information on the dataset as requested by the original owner/supplier. The aim is for this procedure to be rolled out across JNCC's marine team (and potentially JNCC-wide) in the coming months.
Whilst restrictions can be captured generally in the _usage_ section of the metadata, it would be worthwhile to have a standard place for metadata creators to link to the DPA for the dataset in question for auditing and information etc.
This doesn't necessarily have to be a new field, but could instead be written into the topcat/metadata protocol to instruct the user to include the information in an existing field (e.g. the Notes section).
|
1.0
|
Providing users with links to Data Provider Agreements and License - For third party data, the Marine team is working towards the use of a _'Data Provider Agreement'_ which will capture specific restrictions information on the dataset as requested by the original owner/supplier. The aim is for this procedure to be rolled out across JNCC's marine team (and potentially JNCC-wide) in the coming months.
Whilst restrictions can be captured generally in the _usage_ section of the metadata, it would be worthwhile to have a standard place for metadata creators to link to the DPA for the dataset in question for auditing and information etc.
This doesn't necessarily have to be a new field, but could instead be written into the topcat/metadata protocol to instruct the user to include the information in an existing field (e.g. the Notes section).
|
priority
|
providing users with links to data provider agreements and license for third party data the marine team is working towards the use of a data provider agreement which will capture specific restrictions information on the dataset as requested by the original owner supplier the aim is for this procedure to be rolled out across jncc s marine team and potentially jncc wide in the coming months whilst restrictions can be captured generally in the usage section of the metadata it would be worthwhile to have a standard place for metadata creators to link to the dpa for the dataset in question for auditing and information etc this doesn t necessarily have to be a new field but could instead be written into the topcat metadata protocol to instruct the user to include the information in an existing field e g the notes section
| 1
|
90,301
| 3,814,201,535
|
IssuesEvent
|
2016-03-28 11:38:01
|
Esri/coordinate-conversion-addin-dotnet
|
https://api.github.com/repos/Esri/coordinate-conversion-addin-dotnet
|
closed
|
Use the "Flash" Call Inherent to ArcMap for Flash button
|
4 - Verify priority - high
|
Customer Feedback from on-site visit on 18FEB2016. Currently the Flash button pans-to and creates a graphic at the coordinate location. What if the graphic is the same symbology as a feature at the same location? When clicking the "Flash" button, the user thought it would act just like the inherent "Flash" call in ArcMap.
This isn't identical to Issue #67, but does overlap when speaking about the graphic being placed in the Data Frame.
|
1.0
|
Use the "Flash" Call Inherent to ArcMap for Flash button - Customer Feedback from on-site visit on 18FEB2016. Currently the Flash button pans-to and creates a graphic at the coordinate location. What if the graphic is the same symbology as a feature at the same location? When clicking the "Flash" button, the user thought it would act just like the inherent "Flash" call in ArcMap.
This isn't identical to Issue #67, but does overlap when speaking about the graphic being placed in the Data Frame.
|
priority
|
use the flash call inherent to arcmap for flash button customer feedback from on site visit on currently the flash button pans to and creates a graphic at the coordinate location what if the graphic is the same symbology as a feature at the same location when clicking the flash button the user thought it would act just like the inherent flash call in arcmap this isn t identical to issue but does overlap when speaking about the graphic being placed in the data frame
| 1
|
27,815
| 2,696,335,559
|
IssuesEvent
|
2015-04-02 13:30:37
|
alexeyxo/protobuf-swift
|
https://api.github.com/repos/alexeyxo/protobuf-swift
|
closed
|
Errors while compiling: does not conform to protocol 'GeneratedMessageProtocol' and is not a member type of
|
bug high priority
|
I have several protobuf files which are actively used in a project with some other languages, and now need swift versions. I can build .pb.swift files successfully but when I add them to Xcode, I am receiving tons of bugs mainly composed of:
* Type 1 'Foo' does not conform to protocol 'GeneratedMessageProtocol'"
* 'Foo' is not a member type of 'Bar'
I prepared a minimal example showing this issue:
```
message PStatus {
optional int64 battery_usage = 1;
optional int64 max_battery_usage = 2;
message PCardVoltageStatus
{
message PCard
{
optional string card_name = 1;
optional string no = 2;
};
repeated PCard card = 1;
};
};
```
If you build .pb.swift file from this example and add to xcode, you should also get those errors.
|
1.0
|
Errors while compiling: does not conform to protocol 'GeneratedMessageProtocol' and is not a member type of - I have several protobuf files which are actively used in a project with some other languages, and now need swift versions. I can build .pb.swift files successfully but when I add them to Xcode, I am receiving tons of bugs mainly composed of:
* Type 1 'Foo' does not conform to protocol 'GeneratedMessageProtocol'"
* 'Foo' is not a member type of 'Bar'
I prepared a minimal example showing this issue:
```
message PStatus {
optional int64 battery_usage = 1;
optional int64 max_battery_usage = 2;
message PCardVoltageStatus
{
message PCard
{
optional string card_name = 1;
optional string no = 2;
};
repeated PCard card = 1;
};
};
```
If you build .pb.swift file from this example and add to xcode, you should also get those errors.
|
priority
|
errors while compiling does not conform to protocol generatedmessageprotocol and is not a member type of i have several protobuf files which are actively used in a project with some other languages and now need swift versions i can build pb swift files successfully but when i add them to xcode i am receiving tons of bugs mainly composed of type foo does not conform to protocol generatedmessageprotocol foo is not a member type of bar i prepared a minimal example showing this issue message pstatus optional battery usage optional max battery usage message pcardvoltagestatus message pcard optional string card name optional string no repeated pcard card if you build pb swift file from this example and add to xcode you should also get those errors
| 1
|
279,362
| 8,664,452,354
|
IssuesEvent
|
2018-11-28 20:12:49
|
supergiant/control
|
https://api.github.com/repos/supergiant/control
|
closed
|
2.0: API - After rebooting the master, the etcd database can not be found.
|
High Priority
|
**Short Summary:**
After rebooting the master, the master can not join the cluster due to the etcd db volume being stored in the /tmp directory.
**Steps to Reproduce:**
1. Spin up a cluster
2. ssh into the master and issue a reboot
**Expected Results:**
The box reboots and the cluster is working correctly
**Actual Results:**
The box reboots and the etcd database can not be found. It shows as started when checking systemctl status. However, nothing is happening. After restarting etcd with systemctl stop and restart, the etcd will not run.
Please see the attached screen shot to show that the volume is located at /tmp.
**Dev Info:** (add links to log files)
1. Output of `go version`: 1.10
2. Commit hash or release tag used (`git log`): bfa60ad35ccef78c4e51b19bd01eb2a1e06d2d3f
3. Number of Masters and Nodes: Any
4. Cloud Provider: DO and AWS

|
1.0
|
2.0: API - After rebooting the master, the etcd database can not be found. - **Short Summary:**
After rebooting the master, the master can not join the cluster due to the etcd db volume being stored in the /tmp directory.
**Steps to Reproduce:**
1. Spin up a cluster
2. ssh into the master and issue a reboot
**Expected Results:**
The box reboots and the cluster is working correctly
**Actual Results:**
The box reboots and the etcd database can not be found. It shows as started when checking systemctl status. However, nothing is happening. After restarting etcd with systemctl stop and restart, the etcd will not run.
Please see the attached screen shot to show that the volume is located at /tmp.
**Dev Info:** (add links to log files)
1. Output of `go version`: 1.10
2. Commit hash or release tag used (`git log`): bfa60ad35ccef78c4e51b19bd01eb2a1e06d2d3f
3. Number of Masters and Nodes: Any
4. Cloud Provider: DO and AWS

|
priority
|
api after rebooting the master the etcd database can not be found short summary after rebooting the master the master can not join the cluster due to the etcd db volume being stored in the tmp directory steps to reproduce spin up a cluster ssh into the master and issue a reboot expected results the box reboots and the cluster is working correctly actual results the box reboots and the etcd database can not be found it shows as started when checking systemctl status however nothing is happening after restarting etcd with systemctl stop and restart the etcd will not run please see the attached screen shot to show that the volume is located at tmp dev info add links to log files output of go version commit hash or release tag used git log number of masters and nodes any cloud provider do and aws
| 1
|
639,705
| 20,762,585,956
|
IssuesEvent
|
2022-03-15 17:28:23
|
project-pareto/project-pareto
|
https://api.github.com/repos/project-pareto/project-pareto
|
opened
|
Fix image location warnings in Sphinx/ReadTheDocs
|
Priority:High
|
- [ ] Fix warnings
- [ ] Set `fail_on_warning: true` in `.readthedocs.yaml`
|
1.0
|
Fix image location warnings in Sphinx/ReadTheDocs - - [ ] Fix warnings
- [ ] Set `fail_on_warning: true` in `.readthedocs.yaml`
|
priority
|
fix image location warnings in sphinx readthedocs fix warnings set fail on warning true in readthedocs yaml
| 1
|
717,068
| 24,659,795,236
|
IssuesEvent
|
2022-10-18 05:15:44
|
AY2223S1-CS2103T-T10-1/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-T10-1/tp
|
closed
|
As a student, I can edit module
|
type.Story priority.High
|
So that I do not have to delete module data and re-create in case I mess up.
|
1.0
|
As a student, I can edit module - So that I do not have to delete module data and re-create in case I mess up.
|
priority
|
as a student i can edit module so that i do not have to delete module data and re create in case i mess up
| 1
|
695,951
| 23,877,712,097
|
IssuesEvent
|
2022-09-07 20:49:20
|
ClassicLootManager/ClassicLootManager
|
https://api.github.com/repos/ClassicLootManager/ClassicLootManager
|
closed
|
Player rename not working due to GUID check
|
bug Priority::High
|
Existence check in AddProfile is wrong and checks the conditions improperly thus blocking rename support
|
1.0
|
Player rename not working due to GUID check - Existence check in AddProfile is wrong and checks the conditions improperly thus blocking rename support
|
priority
|
player rename not working due to guid check existence check in addprofile is wrong and checks the conditions improperly thus blocking rename support
| 1
|
74,282
| 3,437,296,982
|
IssuesEvent
|
2015-12-13 03:11:09
|
jakev/dtf
|
https://api.github.com/repos/jakev/dtf
|
closed
|
Export fails when modules lack version
|
bug priority-high
|
```
06:38:58 /DevTesting$ dtf pm export test.zip
Traceback (most recent call last):
File "/usr/local/bin/dtf", line 185, in <module>
sys.exit(main())
File "/usr/local/bin/dtf", line 149, in main
return pkg.launch_builtin_module('pm', sys.argv)
File "/usr/local/lib/python2.7/dist-packages/dtf/packages.py", line 145, in launch_builtin_module
return __launch_python_module(launch_path, cmd, args)
File "/usr/local/lib/python2.7/dist-packages/dtf/packages.py", line 97, in __launch_python_module
return mod_inst.run(args)
File "/usr/local/lib/python2.7/dist-packages/dtf/module.py", line 62, in run
result = getattr(self, 'execute')(args)
File "/usr/local/lib/python2.7/dist-packages/dtf/core/cmds/pm.py", line 723, in execute
rtn = self.do_export(args)
File "/usr/local/lib/python2.7/dist-packages/dtf/core/cmds/pm.py", line 235, in do_export
rtn = self.generate_export_xml(export_items, export_manifest)
File "/usr/local/lib/python2.7/dist-packages/dtf/core/cmds/pm.py", line 390, in generate_export_xml
item_xml.attrib['majorVersion'] = item.major_version
File "lxml.etree.pyx", line 2245, in lxml.etree._Attrib.__setitem__ (src/lxml/lxml.etree.c:58775)
File "apihelpers.pxi", line 547, in lxml.etree._setAttributeValue (src/lxml/lxml.etree.c:19025)
File "apihelpers.pxi", line 1393, in lxml.etree._utf8 (src/lxml/lxml.etree.c:26460)
TypeError: Argument must be bytes or unicode, got 'NoneType'
```
|
1.0
|
Export fails when modules lack version - ```
06:38:58 /DevTesting$ dtf pm export test.zip
Traceback (most recent call last):
File "/usr/local/bin/dtf", line 185, in <module>
sys.exit(main())
File "/usr/local/bin/dtf", line 149, in main
return pkg.launch_builtin_module('pm', sys.argv)
File "/usr/local/lib/python2.7/dist-packages/dtf/packages.py", line 145, in launch_builtin_module
return __launch_python_module(launch_path, cmd, args)
File "/usr/local/lib/python2.7/dist-packages/dtf/packages.py", line 97, in __launch_python_module
return mod_inst.run(args)
File "/usr/local/lib/python2.7/dist-packages/dtf/module.py", line 62, in run
result = getattr(self, 'execute')(args)
File "/usr/local/lib/python2.7/dist-packages/dtf/core/cmds/pm.py", line 723, in execute
rtn = self.do_export(args)
File "/usr/local/lib/python2.7/dist-packages/dtf/core/cmds/pm.py", line 235, in do_export
rtn = self.generate_export_xml(export_items, export_manifest)
File "/usr/local/lib/python2.7/dist-packages/dtf/core/cmds/pm.py", line 390, in generate_export_xml
item_xml.attrib['majorVersion'] = item.major_version
File "lxml.etree.pyx", line 2245, in lxml.etree._Attrib.__setitem__ (src/lxml/lxml.etree.c:58775)
File "apihelpers.pxi", line 547, in lxml.etree._setAttributeValue (src/lxml/lxml.etree.c:19025)
File "apihelpers.pxi", line 1393, in lxml.etree._utf8 (src/lxml/lxml.etree.c:26460)
TypeError: Argument must be bytes or unicode, got 'NoneType'
```
|
priority
|
export fails when modules lack version devtesting dtf pm export test zip traceback most recent call last file usr local bin dtf line in sys exit main file usr local bin dtf line in main return pkg launch builtin module pm sys argv file usr local lib dist packages dtf packages py line in launch builtin module return launch python module launch path cmd args file usr local lib dist packages dtf packages py line in launch python module return mod inst run args file usr local lib dist packages dtf module py line in run result getattr self execute args file usr local lib dist packages dtf core cmds pm py line in execute rtn self do export args file usr local lib dist packages dtf core cmds pm py line in do export rtn self generate export xml export items export manifest file usr local lib dist packages dtf core cmds pm py line in generate export xml item xml attrib item major version file lxml etree pyx line in lxml etree attrib setitem src lxml lxml etree c file apihelpers pxi line in lxml etree setattributevalue src lxml lxml etree c file apihelpers pxi line in lxml etree src lxml lxml etree c typeerror argument must be bytes or unicode got nonetype
| 1
|
420,091
| 12,233,183,891
|
IssuesEvent
|
2020-05-04 11:08:13
|
bedita/bedita
|
https://api.github.com/repos/bedita/bedita
|
closed
|
Links (model, table, entity)
|
Priority - High Topic - Core Topic - ORM
|
Provide data modelling in BEdita/Core for "links".
Links extend object. They should have also the following properties:
- http status
- last update date
|
1.0
|
Links (model, table, entity) - Provide data modelling in BEdita/Core for "links".
Links extend object. They should have also the following properties:
- http status
- last update date
|
priority
|
links model table entity provide data modelling in bedita core for links links extend object they should have also the following properties http status last update date
| 1
|
590,716
| 17,785,768,822
|
IssuesEvent
|
2021-08-31 10:50:59
|
GEWIS/gewisweb
|
https://api.github.com/repos/GEWIS/gewisweb
|
closed
|
`./web` is broken, which causes the cronjobs to fail
|
Type: Bug Priority: High For: Backend Status: Confirmed
|
`./web` fails because the script is unaware of any environment variables (primarily `APP_ENV`).
|
1.0
|
`./web` is broken, which causes the cronjobs to fail - `./web` fails because the script is unaware of any environment variables (primarily `APP_ENV`).
|
priority
|
web is broken which causes the cronjobs to fail web fails because the script is unaware of any environment variables primarily app env
| 1
|
121,116
| 4,805,127,192
|
IssuesEvent
|
2016-11-02 15:19:17
|
windupmicheal/Tackle-Trading
|
https://api.github.com/repos/windupmicheal/Tackle-Trading
|
closed
|
New header design + modification to stylesheet
|
1. High Priority!
|
Notice:
- The new all-white logo
- New menu font
- The location of LOGIN | SIGNUP so that the P in SIGNUP is virtically aligned with the magnifying glass of the search bar
- New Login/signup/logout font
- New "Active Menu" locator - the grey highlight oval surrounding the currently expanded menu
- The font and light-grey color of the SEARCH text
- The location of the magnifying glass in the SEARCH bar
You'll notice that the menu headers are using the "MUSEO SLAB" font, which is available from typekit.com. I have Noah's credentials for this. Ask me for them in private when you are ready.
According to a wpengine.com scan, we're using FIVE fonts on the site. That is two too many. I would like to adjust the stylesheet to use MUSEO SLAB font for all headers, and HELVETICA NEUE for paragraph text. The only other font that should be used is AERO, which is the logo font and the font for headers on the LEARN MORE page. No other fonts should be used on the site. Where OPEN SANS is used in the footer, let's use HELVETICA NEUE.
You should be able to pull graphic assets from this PDF file. If not, let me know what you need and I'll get it to you.
[new front page w video bg.pdf](https://github.com/windupmicheal/Tackle-Trading/files/546130/new.front.page.w.video.bg.pdf)
|
1.0
|
New header design + modification to stylesheet - Notice:
- The new all-white logo
- New menu font
- The location of LOGIN | SIGNUP so that the P in SIGNUP is virtically aligned with the magnifying glass of the search bar
- New Login/signup/logout font
- New "Active Menu" locator - the grey highlight oval surrounding the currently expanded menu
- The font and light-grey color of the SEARCH text
- The location of the magnifying glass in the SEARCH bar
You'll notice that the menu headers are using the "MUSEO SLAB" font, which is available from typekit.com. I have Noah's credentials for this. Ask me for them in private when you are ready.
According to a wpengine.com scan, we're using FIVE fonts on the site. That is two too many. I would like to adjust the stylesheet to use MUSEO SLAB font for all headers, and HELVETICA NEUE for paragraph text. The only other font that should be used is AERO, which is the logo font and the font for headers on the LEARN MORE page. No other fonts should be used on the site. Where OPEN SANS is used in the footer, let's use HELVETICA NEUE.
You should be able to pull graphic assets from this PDF file. If not, let me know what you need and I'll get it to you.
[new front page w video bg.pdf](https://github.com/windupmicheal/Tackle-Trading/files/546130/new.front.page.w.video.bg.pdf)
|
priority
|
new header design modification to stylesheet notice the new all white logo new menu font the location of login signup so that the p in signup is virtically aligned with the magnifying glass of the search bar new login signup logout font new active menu locator the grey highlight oval surrounding the currently expanded menu the font and light grey color of the search text the location of the magnifying glass in the search bar you ll notice that the menu headers are using the museo slab font which is available from typekit com i have noah s credentials for this ask me for them in private when you are ready according to a wpengine com scan we re using five fonts on the site that is two too many i would like to adjust the stylesheet to use museo slab font for all headers and helvetica neue for paragraph text the only other font that should be used is aero which is the logo font and the font for headers on the learn more page no other fonts should be used on the site where open sans is used in the footer let s use helvetica neue you should be able to pull graphic assets from this pdf file if not let me know what you need and i ll get it to you
| 1
|
115,838
| 4,682,939,155
|
IssuesEvent
|
2016-10-09 14:42:45
|
CS2103AUG2016-T15-C3/main
|
https://api.github.com/repos/CS2103AUG2016-T15-C3/main
|
opened
|
As a user I want to modify the information of a task
|
priority.high type.story
|
So that I can update the details, requirements and deadline of a task if they are changed
|
1.0
|
As a user I want to modify the information of a task - So that I can update the details, requirements and deadline of a task if they are changed
|
priority
|
as a user i want to modify the information of a task so that i can update the details requirements and deadline of a task if they are changed
| 1
|
409,100
| 11,956,787,907
|
IssuesEvent
|
2020-04-04 12:04:45
|
weso/shex-lite
|
https://api.github.com/repos/weso/shex-lite
|
closed
|
Travis CI is completly unstable for scala builds
|
affects: repository dificulty: low priority: high status: accepted type: bug
|
Travis CI has been unstable during all day for scala builds, this is related to that sometimes it misses some libraries, sometimes it misses other libraries and then the same test executed twice fails first time and passes second time....
In my opinion we should be looking for another CI tool.
|
1.0
|
Travis CI is completly unstable for scala builds - Travis CI has been unstable during all day for scala builds, this is related to that sometimes it misses some libraries, sometimes it misses other libraries and then the same test executed twice fails first time and passes second time....
In my opinion we should be looking for another CI tool.
|
priority
|
travis ci is completly unstable for scala builds travis ci has been unstable during all day for scala builds this is related to that sometimes it misses some libraries sometimes it misses other libraries and then the same test executed twice fails first time and passes second time in my opinion we should be looking for another ci tool
| 1
|
97,244
| 3,987,523,488
|
IssuesEvent
|
2016-05-09 04:29:18
|
rfbonett/CSCi435-ODBR
|
https://api.github.com/repos/rfbonett/CSCi435-ODBR
|
closed
|
Hierarchy Dump : Can no longer use Accessibility Service to get root Node
|
bug High Priority
|
We need a new way to get a handle to the root node's AccessibilityNodeInfo if we are to use the current implementation. Otherwise, we can use uiautomator dump via bash.
|
1.0
|
Hierarchy Dump : Can no longer use Accessibility Service to get root Node - We need a new way to get a handle to the root node's AccessibilityNodeInfo if we are to use the current implementation. Otherwise, we can use uiautomator dump via bash.
|
priority
|
hierarchy dump can no longer use accessibility service to get root node we need a new way to get a handle to the root node s accessibilitynodeinfo if we are to use the current implementation otherwise we can use uiautomator dump via bash
| 1
|
248,647
| 7,934,659,443
|
IssuesEvent
|
2018-07-08 21:51:45
|
BananiumLabs/AtomBlast.io
|
https://api.github.com/repos/BananiumLabs/AtomBlast.io
|
closed
|
Split Powerup into Atom and Compound
|
enhancement high priority
|
Currently the Powerup structure is created in a way that the item you pick up is the same item that you will use. However, this is not how our game will be structured.
Instead, we will split the functionality into two different classes, Atom and Compound.
Atom is what you pick up, so it will inherit most of the current Powerup functionality (spawning, pickup, etc). However, you cannot use Atoms by themselves.
Atoms must be turned into Compounds before they are useful. Compounds use Atoms as basic building blocks, just like in real life. Players will be able to select (or unlock??) Blueprints which will allow for different configurations of Compounds to be created and deployed.
We will likely create several slots in the HUD where players can select what several blueprints they want ingame out of a large collection of different blueprints. During the game, players cannot change what blueprints they can use- this will be done in the Main Menu and between games.
|
1.0
|
Split Powerup into Atom and Compound - Currently the Powerup structure is created in a way that the item you pick up is the same item that you will use. However, this is not how our game will be structured.
Instead, we will split the functionality into two different classes, Atom and Compound.
Atom is what you pick up, so it will inherit most of the current Powerup functionality (spawning, pickup, etc). However, you cannot use Atoms by themselves.
Atoms must be turned into Compounds before they are useful. Compounds use Atoms as basic building blocks, just like in real life. Players will be able to select (or unlock??) Blueprints which will allow for different configurations of Compounds to be created and deployed.
We will likely create several slots in the HUD where players can select what several blueprints they want ingame out of a large collection of different blueprints. During the game, players cannot change what blueprints they can use- this will be done in the Main Menu and between games.
|
priority
|
split powerup into atom and compound currently the powerup structure is created in a way that the item you pick up is the same item that you will use however this is not how our game will be structured instead we will split the functionality into two different classes atom and compound atom is what you pick up so it will inherit most of the current powerup functionality spawning pickup etc however you cannot use atoms by themselves atoms must be turned into compounds before they are useful compounds use atoms as basic building blocks just like in real life players will be able to select or unlock blueprints which will allow for different configurations of compounds to be created and deployed we will likely create several slots in the hud where players can select what several blueprints they want ingame out of a large collection of different blueprints during the game players cannot change what blueprints they can use this will be done in the main menu and between games
| 1
|
168,623
| 6,379,292,135
|
IssuesEvent
|
2017-08-02 14:29:00
|
canonical-websites/vanillaframework.io
|
https://api.github.com/repos/canonical-websites/vanillaframework.io
|
opened
|
Update link of CTA button 'Get started'
|
Priority: High
|
## Summary
Update 'Get started' CTA button to point to Vanilla documentation rather than anchor scrolling down one row on homepage.
## Current and expected result
Current result - 'Get started' anchor scrolls to 'Quick start' on same page.
Expected result - 'Get started' to send users to Vanilla documentation.
|
1.0
|
Update link of CTA button 'Get started' - ## Summary
Update 'Get started' CTA button to point to Vanilla documentation rather than anchor scrolling down one row on homepage.
## Current and expected result
Current result - 'Get started' anchor scrolls to 'Quick start' on same page.
Expected result - 'Get started' to send users to Vanilla documentation.
|
priority
|
update link of cta button get started summary update get started cta button to point to vanilla documentation rather than anchor scrolling down one row on homepage current and expected result current result get started anchor scrolls to quick start on same page expected result get started to send users to vanilla documentation
| 1
|
586,440
| 17,577,733,599
|
IssuesEvent
|
2021-08-15 23:16:33
|
PlanktonTeam/planktonr
|
https://api.github.com/repos/PlanktonTeam/planktonr
|
closed
|
Importing commonly used packages
|
enhancement high priority
|
I think I should switch to attaching the whole `dplyr` package to minimise the `dplyr::` calls.
This can be done just once (suggest a function in utils.R). I should also adopt this approach for magrittr.
|
1.0
|
Importing commonly used packages - I think I should switch to attaching the whole `dplyr` package to minimise the `dplyr::` calls.
This can be done just once (suggest a function in utils.R). I should also adopt this approach for magrittr.
|
priority
|
importing commonly used packages i think i should switch to attaching the whole dplyr package to minimise the dplyr calls this can be done just once suggest a function in utils r i should also adopt this approach for magrittr
| 1
|
352,894
| 10,546,929,425
|
IssuesEvent
|
2019-10-02 23:00:51
|
emory-libraries/ezpaarse-platforms
|
https://api.github.com/repos/emory-libraries/ezpaarse-platforms
|
closed
|
Nurimedia (DBpia & KRpia)
|
Add Parser High Priority Stakeholder Priority
|
### Example:star::star: :
http://www.dbpia.co.kr.proxy.library.emory.edu/
http://www.krpia.co.kr.proxy.library.emory.edu/
### Priority:
High
### Subscriber (Library):
Woodruff
### ezPAARSE
Analysis: N/A
Trello: N/A
|
2.0
|
Nurimedia (DBpia & KRpia) - ### Example:star::star: :
http://www.dbpia.co.kr.proxy.library.emory.edu/
http://www.krpia.co.kr.proxy.library.emory.edu/
### Priority:
High
### Subscriber (Library):
Woodruff
### ezPAARSE
Analysis: N/A
Trello: N/A
|
priority
|
nurimedia dbpia krpia example star star priority high subscriber library woodruff ezpaarse analysis n a trello n a
| 1
|
96,140
| 3,964,945,508
|
IssuesEvent
|
2016-05-03 04:59:37
|
meumobi/infomobi
|
https://api.github.com/repos/meumobi/infomobi
|
opened
|
Prevent vote if none option is selected
|
bug high-priority polls
|
see logs on api
```bash
[2016-05-03 07:04:44] sitebuilder.INFO: http request {"method":"POST","url":"/api/alcon.meumobi.com/items/57282f7a9a645da03e22a204/poll","data":[],"component":"api"}
```
|
1.0
|
Prevent vote if none option is selected - see logs on api
```bash
[2016-05-03 07:04:44] sitebuilder.INFO: http request {"method":"POST","url":"/api/alcon.meumobi.com/items/57282f7a9a645da03e22a204/poll","data":[],"component":"api"}
```
|
priority
|
prevent vote if none option is selected see logs on api bash sitebuilder info http request method post url api alcon meumobi com items poll data component api
| 1
|
747,855
| 26,101,183,410
|
IssuesEvent
|
2022-12-27 07:15:54
|
bounswe/bounswe2022group7
|
https://api.github.com/repos/bounswe/bounswe2022group7
|
closed
|
[FE] Display following users on profile page
|
Status: Completed Priority: High Type: Implementation Target: Frontend
|
When the following button on the profile page is clicked, the list of followed users should be displayed.
|
1.0
|
[FE] Display following users on profile page - When the following button on the profile page is clicked, the list of followed users should be displayed.
|
priority
|
display following users on profile page when the following button on the profile page is clicked the list of followed users should be displayed
| 1
|
774,403
| 27,195,452,142
|
IssuesEvent
|
2023-02-20 04:32:02
|
openmsupply/open-msupply
|
https://api.github.com/repos/openmsupply/open-msupply
|
closed
|
Postgres running out of memory
|
bug back-end sync v5 Priority: High
|
When running in development with postgres DB I cannot initialise modest sized sites. In my tests I am initialising and integrating about 180K records, at about 10 mins into integration I get this error and integration stops:
```
2022-12-01 17:57:26.443497 WARN service::sync::translation_and_integration - DBError { msg: "UNKNOWN", extra: "\"\\\"out of shared memory\\\"\"" } "6C2C13CE7735AB4CA5E002AF1B533E61" "requisition_line" service/src/sync/translation_and_integration.rs:125
```
Likely unrelated, but during initialisation I get many, many errors for unsupported records.
**Interestingly this is totally fine when running sqlite.**
Preliminary googling suggests this is likely due to having too many transactions within a transaction.
|
1.0
|
Postgres running out of memory - When running in development with postgres DB I cannot initialise modest sized sites. In my tests I am initialising and integrating about 180K records, at about 10 mins into integration I get this error and integration stops:
```
2022-12-01 17:57:26.443497 WARN service::sync::translation_and_integration - DBError { msg: "UNKNOWN", extra: "\"\\\"out of shared memory\\\"\"" } "6C2C13CE7735AB4CA5E002AF1B533E61" "requisition_line" service/src/sync/translation_and_integration.rs:125
```
Likely unrelated, but during initialisation I get many, many errors for unsupported records.
**Interestingly this is totally fine when running sqlite.**
Preliminary googling suggests this is likely due to having too many transactions within a transaction.
|
priority
|
postgres running out of memory when running in development with postgres db i cannot initialise modest sized sites in my tests i am initialising and integrating about records at about mins into integration i get this error and integration stops warn service sync translation and integration dberror msg unknown extra out of shared memory requisition line service src sync translation and integration rs likely unrelated but during initialisation i get many many errors for unsupported records interestingly this is totally fine when running sqlite preliminary googling suggests this is likely due to having too many transactions within a transaction
| 1
|
117,337
| 4,713,972,026
|
IssuesEvent
|
2016-10-14 22:00:54
|
mulesoft/api-workbench
|
https://api.github.com/repos/mulesoft/api-workbench
|
closed
|
DataType Examples are not parsed
|
bug priority:high
|
In attempting to use dataTypes in place of schemas I have defined a type with an example. The example in the type is ignored by the console or mocking service unless the example is explicitly defined in the response's specific media type.
I raised this issue on the api-designer repo as well, and it was confirmed a bug this morning.
Example:
Type Definition:
```RAML
types:
Member:
properties:
firstName: string
lastName: string
age: integer
examples:
Bob:
value:
firstName: "Bob"
lastName: "Slidell"
age: 42
Bill:
value:
firstName: "Bill"
lastName: "Lumberg"
age: 41
````
If a response is resource type the example is not included
````RAML
resourceTypes:
collection:
usage: Use this resourceType to represent a collection of <<resourcePathName|!singularize>> items
description: A collection of <<resourcePathName>>
get:
description: |
Get all <<resourcePathName>>,
optionally filtered
responses:
200:
body:
type: <<typeName|!singularize>>[]
````
Resource:
````RAML
/members:
type:
collection:
typeName: <<resourcePathName|!uppercamelcase>>
get:
````
This will show the Object type and its properties but not the example. If the resource type explicitly defines the example it will be included in the console and returned by the Mock service.
````RAML
resourceTypes:
collection:
usage: Use this resourceType to represent a collection of <<resourcePathName|!singularize>> items
description: A collection of <<resourcePathName>>
get:
description: |
Get all <<resourcePathName>>,
optionally filtered
responses:
200:
body:
type: <<typeName|!singularize>>[]
application/json:
examples:
Bob:
value:
firstName: "Bob"
lastName: "Slidell"
age: 42
Bill:
value:
firstName: "Bill"
lastName: "Lumberg"
age: 41
````
The description of Examples in the 1.0 spec seems to indicate this would be a supported usage. Full RAML below:
````RAML
#%RAML 1.0
title: TestRAML
baseUri: http://someapi.com/api
mediaType: application/json
resourceTypes:
collection:
usage: Use this resourceType to represent a collection of <<resourcePathName|!singularize>> items
description: A collection of <<resourcePathName>>
get:
description: |
Get all <<resourcePathName>>,
optionally filtered
responses:
200:
body:
type: <<typeName|!singularize>>[]
types:
Member:
properties:
firstName: string
lastName: string
age: integer
examples:
Bob:
value:
firstName: "Bob"
lastName: "Slidell"
age: 42
Bill:
value:
firstName: "Bill"
lastName: "Lumberg"
age: 41
/members:
type:
collection:
typeName: <<resourcePathName|!uppercamelcase>>
get:
````
|
1.0
|
DataType Examples are not parsed - In attempting to use dataTypes in place of schemas I have defined a type with an example. The example in the type is ignored by the console or mocking service unless the example is explicitly defined in the response's specific media type.
I raised this issue on the api-designer repo as well, and it was confirmed a bug this morning.
Example:
Type Definition:
```RAML
types:
Member:
properties:
firstName: string
lastName: string
age: integer
examples:
Bob:
value:
firstName: "Bob"
lastName: "Slidell"
age: 42
Bill:
value:
firstName: "Bill"
lastName: "Lumberg"
age: 41
````
If a response is resource type the example is not included
````RAML
resourceTypes:
collection:
usage: Use this resourceType to represent a collection of <<resourcePathName|!singularize>> items
description: A collection of <<resourcePathName>>
get:
description: |
Get all <<resourcePathName>>,
optionally filtered
responses:
200:
body:
type: <<typeName|!singularize>>[]
````
Resource:
````RAML
/members:
type:
collection:
typeName: <<resourcePathName|!uppercamelcase>>
get:
````
This will show the Object type and its properties but not the example. If the resource type explicitly defines the example it will be included in the console and returned by the Mock service.
````RAML
resourceTypes:
collection:
usage: Use this resourceType to represent a collection of <<resourcePathName|!singularize>> items
description: A collection of <<resourcePathName>>
get:
description: |
Get all <<resourcePathName>>,
optionally filtered
responses:
200:
body:
type: <<typeName|!singularize>>[]
application/json:
examples:
Bob:
value:
firstName: "Bob"
lastName: "Slidell"
age: 42
Bill:
value:
firstName: "Bill"
lastName: "Lumberg"
age: 41
````
The description of Examples in the 1.0 spec seems to indicate this would be a supported usage. Full RAML below:
````RAML
#%RAML 1.0
title: TestRAML
baseUri: http://someapi.com/api
mediaType: application/json
resourceTypes:
collection:
usage: Use this resourceType to represent a collection of <<resourcePathName|!singularize>> items
description: A collection of <<resourcePathName>>
get:
description: |
Get all <<resourcePathName>>,
optionally filtered
responses:
200:
body:
type: <<typeName|!singularize>>[]
types:
Member:
properties:
firstName: string
lastName: string
age: integer
examples:
Bob:
value:
firstName: "Bob"
lastName: "Slidell"
age: 42
Bill:
value:
firstName: "Bill"
lastName: "Lumberg"
age: 41
/members:
type:
collection:
typeName: <<resourcePathName|!uppercamelcase>>
get:
````
|
priority
|
datatype examples are not parsed in attempting to use datatypes in place of schemas i have defined a type with an example the example in the type is ignored by the console or mocking service unless the example is explicitly defined in the response s specific media type i raised this issue on the api designer repo as well and it was confirmed a bug this morning example type definition raml types member properties firstname string lastname string age integer examples bob value firstname bob lastname slidell age bill value firstname bill lastname lumberg age if a response is resource type the example is not included raml resourcetypes collection usage use this resourcetype to represent a collection of items description a collection of get description get all optionally filtered responses body type resource raml members type collection typename get this will show the object type and its properties but not the example if the resource type explicitly defines the example it will be included in the console and returned by the mock service raml resourcetypes collection usage use this resourcetype to represent a collection of items description a collection of get description get all optionally filtered responses body type application json examples bob value firstname bob lastname slidell age bill value firstname bill lastname lumberg age the description of examples in the spec seems to indicate this would be a supported usage full raml below raml raml title testraml baseuri mediatype application json resourcetypes collection usage use this resourcetype to represent a collection of items description a collection of get description get all optionally filtered responses body type types member properties firstname string lastname string age integer examples bob value firstname bob lastname slidell age bill value firstname bill lastname lumberg age members type collection typename get
| 1
|
80,453
| 3,561,747,325
|
IssuesEvent
|
2016-01-24 00:43:35
|
godotengine/godot
|
https://api.github.com/repos/godotengine/godot
|
closed
|
Subscenes no longer updates variables values in parent scenes
|
bug high priority topic:editor
|
Subscenes no longer updates their values in parent scenes. The easiest way to reproduce it is to use **exported** variable since for those it behaves like this every time.
That means if you change properties in subscene file the instanced ones will never be updated, even if you never touch these variables in parent scene.
Short video proof: https://youtu.be/A8Dbk8LaT2A
Tested in: 1bc91848e3cee91eccaf2150a74deaf1cd84be13
Bug is quite fresh I'm almost 100% sure it was not with us on the build from the beginning of December.
It's quite serious and takes away one of the reasons to use subscene, think it should be fixed for 2.0
Edit: **Sample in this comment: https://github.com/godotengine/godot/issues/3127#issuecomment-171634970**
|
1.0
|
Subscenes no longer updates variables values in parent scenes - Subscenes no longer updates their values in parent scenes. The easiest way to reproduce it is to use **exported** variable since for those it behaves like this every time.
That means if you change properties in subscene file the instanced ones will never be updated, even if you never touch these variables in parent scene.
Short video proof: https://youtu.be/A8Dbk8LaT2A
Tested in: 1bc91848e3cee91eccaf2150a74deaf1cd84be13
Bug is quite fresh I'm almost 100% sure it was not with us on the build from the beginning of December.
It's quite serious and takes away one of the reasons to use subscene, think it should be fixed for 2.0
Edit: **Sample in this comment: https://github.com/godotengine/godot/issues/3127#issuecomment-171634970**
|
priority
|
subscenes no longer updates variables values in parent scenes subscenes no longer updates their values in parent scenes the easiest way to reproduce it is to use exported variable since for those it behaves like this every time that means if you change properties in subscene file the instanced ones will never be updated even if you never touch these variables in parent scene short video proof tested in bug is quite fresh i m almost sure it was not with us on the build from the beginning of december it s quite serious and takes away one of the reasons to use subscene think it should be fixed for edit sample in this comment
| 1
|
348,319
| 10,440,880,427
|
IssuesEvent
|
2019-09-18 09:36:59
|
geosolutions-it/MapStore2
|
https://api.github.com/repos/geosolutions-it/MapStore2
|
closed
|
Switching CRS, circle annotations change size
|
Accepted CRS Selector Priority: High annotations bug
|
### Description
When switching CRS using the CRS selector, **circle annotations** (if present) change their size.

### In case of Bug (otherwise remove this paragraph)
*Browser Affected*
(use this site: https://www.whatsmybrowser.org/ for non expert users)
- [x] Internet Explorer
- [x] Chrome
- [x] Firefox
- [x] Safari
*Browser Version Affected*
Last
*Steps to reproduce*
- Open a map
- Add some layer
- Add a circle annotation
- switch CRS through the CRS selector
*Expected Result*
The circle annotation doesn't change size.
*Current Result*
The circle annotation changes size.
### Other useful information (optional):
|
1.0
|
Switching CRS, circle annotations change size - ### Description
When switching CRS using the CRS selector, **circle annotations** (if present) change their size.

### In case of Bug (otherwise remove this paragraph)
*Browser Affected*
(use this site: https://www.whatsmybrowser.org/ for non expert users)
- [x] Internet Explorer
- [x] Chrome
- [x] Firefox
- [x] Safari
*Browser Version Affected*
Last
*Steps to reproduce*
- Open a map
- Add some layer
- Add a circle annotation
- switch CRS through the CRS selector
*Expected Result*
The circle annotation doesn't change size.
*Current Result*
The circle annotation changes size.
### Other useful information (optional):
|
priority
|
switching crs circle annotations change size description when switching crs using the crs selector circle annotations if present change their size in case of bug otherwise remove this paragraph browser affected use this site for non expert users internet explorer chrome firefox safari browser version affected last steps to reproduce open a map add some layer add a circle annotation switch crs through the crs selector expected result the circle annotation doesn t change size current result the circle annotation changes size other useful information optional
| 1
|
554,923
| 16,442,517,241
|
IssuesEvent
|
2021-05-20 15:46:05
|
DSpace/dspace-angular
|
https://api.github.com/repos/DSpace/dspace-angular
|
closed
|
Edit of an eperson doesn't work
|
bug e/2 high priority
|
If you try to edit an eperson, for instance the one that come from the sample shibboleth integration to add them a Firstname and Lastname the system says that the changes are saved but the patch request sent to the backend is empty so no changes occur
|
1.0
|
Edit of an eperson doesn't work - If you try to edit an eperson, for instance the one that come from the sample shibboleth integration to add them a Firstname and Lastname the system says that the changes are saved but the patch request sent to the backend is empty so no changes occur
|
priority
|
edit of an eperson doesn t work if you try to edit an eperson for instance the one that come from the sample shibboleth integration to add them a firstname and lastname the system says that the changes are saved but the patch request sent to the backend is empty so no changes occur
| 1
|
196,376
| 6,927,381,303
|
IssuesEvent
|
2017-11-30 22:38:19
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
Extremly long loading times (ref greenleaf server )
|
High Priority Optimization
|
loading times can be over 3-10 minutes when loading times where allot faster before
|
1.0
|
Extremly long loading times (ref greenleaf server ) - loading times can be over 3-10 minutes when loading times where allot faster before
|
priority
|
extremly long loading times ref greenleaf server loading times can be over minutes when loading times where allot faster before
| 1
|
782,196
| 27,489,919,706
|
IssuesEvent
|
2023-03-04 13:45:18
|
bounswe/bounswe2023group2
|
https://api.github.com/repos/bounswe/bounswe2023group2
|
closed
|
Customize labels
|
type: enhancement state: assigned priority: high required: feedback
|
Labels should be customize since default labels are not clear enough for describing the issues:
- Please be careful while choosing the colors, there should not be duplicate color if possible
- Nor too many neither too few labels will not be helpful
Useful Links:
- [github label example](https://seantrane.com/posts/logical-colorful-github-labels-18230/)
- [simple one](https://medium.com/@dave_lunny/sane-github-labels-c5d2e6004b63)
- [tutorial](https://docs.github.com/en/issues/using-labels-and-milestones-to-track-work/managing-labels)
Due date: 07.03.2022
Revisor: @mervegrbz
|
1.0
|
Customize labels - Labels should be customize since default labels are not clear enough for describing the issues:
- Please be careful while choosing the colors, there should not be duplicate color if possible
- Nor too many neither too few labels will not be helpful
Useful Links:
- [github label example](https://seantrane.com/posts/logical-colorful-github-labels-18230/)
- [simple one](https://medium.com/@dave_lunny/sane-github-labels-c5d2e6004b63)
- [tutorial](https://docs.github.com/en/issues/using-labels-and-milestones-to-track-work/managing-labels)
Due date: 07.03.2022
Revisor: @mervegrbz
|
priority
|
customize labels labels should be customize since default labels are not clear enough for describing the issues please be careful while choosing the colors there should not be duplicate color if possible nor too many neither too few labels will not be helpful useful links due date revisor mervegrbz
| 1
|
211,542
| 7,202,019,804
|
IssuesEvent
|
2018-02-06 01:29:42
|
bmentges/aguas_rj_bigdata
|
https://api.github.com/repos/bmentges/aguas_rj_bigdata
|
closed
|
Crawler da Agencia Nacional de Águas
|
feature high priority
|
# Objetivo
Automatizar a ingestão de dados vindos da Agência Nacional de Águas para todos os reservatórios do Estado do Rio de Janeiro.
# Critérios de Aceite
- [x] Script executável que recebe via parâmetro a data de início e a data de fim para crawlear
- [x] Gera os arquivos .html corretamente para cada reservatório dentro do período selecionado
- [x] Se houver falha, não afeta o download anterior
Obs:
* Não precisa limpar downloads antigos (deve ser feito de outra forma)
# Onde estão os dados?
http://sar.ana.gov.br/MedicaoSin?dropDownListEstados=20&dropDownListReservatorios=19083&dataInicial=01%2F01%2F2016&dataFinal=03%2F02%2F2018&button=Buscar
# TODO
- [x] SPIKE (pesquisar): como crawlear os dados no formato que já baixamos antes (o exportar para excel deles, que na verdade baixa HTML) para aproveitar o CSV parser. Entrega: Definir a arquitetura do crawler no README
- [x] Permitir escolher o período para crawlear (para futuros deltas)
- [x] Atualizar o CSV parser para considerar diretórios de mais períodos
- [x] Teste integrado
- [x] Merge no master
|
1.0
|
Crawler da Agencia Nacional de Águas - # Objetivo
Automatizar a ingestão de dados vindos da Agência Nacional de Águas para todos os reservatórios do Estado do Rio de Janeiro.
# Critérios de Aceite
- [x] Script executável que recebe via parâmetro a data de início e a data de fim para crawlear
- [x] Gera os arquivos .html corretamente para cada reservatório dentro do período selecionado
- [x] Se houver falha, não afeta o download anterior
Obs:
* Não precisa limpar downloads antigos (deve ser feito de outra forma)
# Onde estão os dados?
http://sar.ana.gov.br/MedicaoSin?dropDownListEstados=20&dropDownListReservatorios=19083&dataInicial=01%2F01%2F2016&dataFinal=03%2F02%2F2018&button=Buscar
# TODO
- [x] SPIKE (pesquisar): como crawlear os dados no formato que já baixamos antes (o exportar para excel deles, que na verdade baixa HTML) para aproveitar o CSV parser. Entrega: Definir a arquitetura do crawler no README
- [x] Permitir escolher o período para crawlear (para futuros deltas)
- [x] Atualizar o CSV parser para considerar diretórios de mais períodos
- [x] Teste integrado
- [x] Merge no master
|
priority
|
crawler da agencia nacional de águas objetivo automatizar a ingestão de dados vindos da agência nacional de águas para todos os reservatórios do estado do rio de janeiro critérios de aceite script executável que recebe via parâmetro a data de início e a data de fim para crawlear gera os arquivos html corretamente para cada reservatório dentro do período selecionado se houver falha não afeta o download anterior obs não precisa limpar downloads antigos deve ser feito de outra forma onde estão os dados todo spike pesquisar como crawlear os dados no formato que já baixamos antes o exportar para excel deles que na verdade baixa html para aproveitar o csv parser entrega definir a arquitetura do crawler no readme permitir escolher o período para crawlear para futuros deltas atualizar o csv parser para considerar diretórios de mais períodos teste integrado merge no master
| 1
|
401,943
| 11,800,031,587
|
IssuesEvent
|
2020-03-18 16:50:46
|
Sage-Bionetworks/dccvalidator
|
https://api.github.com/repos/Sage-Bionetworks/dccvalidator
|
closed
|
Check that 'parent' in the manifest is a synID
|
high priority
|
The `parent` column in the manifest is required to be complete. However, we had a contributor pass this check with simply putting "syn" in every row. We need to add a check that every value in the column is a synID. It might be beneficial to check that there is only a single unique value, as well, but that does limit the data to being uploaded to a single folder every time.
|
1.0
|
Check that 'parent' in the manifest is a synID - The `parent` column in the manifest is required to be complete. However, we had a contributor pass this check with simply putting "syn" in every row. We need to add a check that every value in the column is a synID. It might be beneficial to check that there is only a single unique value, as well, but that does limit the data to being uploaded to a single folder every time.
|
priority
|
check that parent in the manifest is a synid the parent column in the manifest is required to be complete however we had a contributor pass this check with simply putting syn in every row we need to add a check that every value in the column is a synid it might be beneficial to check that there is only a single unique value as well but that does limit the data to being uploaded to a single folder every time
| 1
|
60,944
| 3,136,004,383
|
IssuesEvent
|
2015-09-10 17:49:13
|
GluuFederation/oxTrust
|
https://api.github.com/repos/GluuFederation/oxTrust
|
closed
|
'Import people' feature not working
|
bug High Priority
|
Gluu Server version: 2.3.3
Situation: Tried to import a xls file ( Users --> Import people --> Upload xls file ). But oxTrust is failing to accept the xls file with error `Import failed, there are similar UID(s) in input file`. The 'Username' of xls sheet has different Usernames for every users.

No indication of error in logs. Stack trace:
INFO | jvm 1 | 2015/08/24 11:35:27 | 2015-08-24 11:35:27,572 INFO [org.gluu.oxtrust.action.PersonImportAction] Found 4 persons in input Excel file
INFO | jvm 1 | 2015/08/24 11:35:27 | 2015-08-24 11:35:27,932 DEBUG [org.gluu.oxtrust.servlet.FaviconImageServlet] Starting organization favicon upload
|
1.0
|
'Import people' feature not working - Gluu Server version: 2.3.3
Situation: Tried to import a xls file ( Users --> Import people --> Upload xls file ). But oxTrust is failing to accept the xls file with error `Import failed, there are similar UID(s) in input file`. The 'Username' of xls sheet has different Usernames for every users.

No indication of error in logs. Stack trace:
INFO | jvm 1 | 2015/08/24 11:35:27 | 2015-08-24 11:35:27,572 INFO [org.gluu.oxtrust.action.PersonImportAction] Found 4 persons in input Excel file
INFO | jvm 1 | 2015/08/24 11:35:27 | 2015-08-24 11:35:27,932 DEBUG [org.gluu.oxtrust.servlet.FaviconImageServlet] Starting organization favicon upload
|
priority
|
import people feature not working gluu server version situation tried to import a xls file users import people upload xls file but oxtrust is failing to accept the xls file with error import failed there are similar uid s in input file the username of xls sheet has different usernames for every users no indication of error in logs stack trace info jvm info found persons in input excel file info jvm debug starting organization favicon upload
| 1
|
270,448
| 8,460,514,315
|
IssuesEvent
|
2018-10-22 19:04:00
|
CS2113-AY1819S1-F09-4/main
|
https://api.github.com/repos/CS2113-AY1819S1-F09-4/main
|
closed
|
Need "listc" / "listC" commands allow user to view all candidates / all companies + all job offers
|
priority.high type.epic
|
#89 refer to PR's description for further elaboration on what is implemented in these 2 commands. Done.
|
1.0
|
Need "listc" / "listC" commands allow user to view all candidates / all companies + all job offers - #89 refer to PR's description for further elaboration on what is implemented in these 2 commands. Done.
|
priority
|
need listc listc commands allow user to view all candidates all companies all job offers refer to pr s description for further elaboration on what is implemented in these commands done
| 1
|
823,103
| 30,928,110,487
|
IssuesEvent
|
2023-08-06 18:42:16
|
jrsteensen/OpenHornet
|
https://api.github.com/repos/jrsteensen/OpenHornet
|
closed
|
interconnect: Ant_Gnd connected to D39, not GND
|
Type: Bug/Obsolesce Category: ECAD Priority: High
|
### Summary:
The ANT_GND of the ANT SEL PANEL should be shown connected to GND_J5-17. Instead, it is connected to D39.
### Screenshots:
<img width="528" alt="image" src="https://github.com/jrsteensen/OpenHornet/assets/62520296/654be750-73eb-4515-bb38-1f3f4f8e66af">
## More Information
*Add an "X" in the square brackets to check the applicable checkboxs.*
### Category:
*Check one or more items.*
- [ ] MCAD (SolidWorks)
- [X] ECAD (PCB Design or other electrical hardware)
- [ ] Software - Sketch
- [ ] Software - DCS-BIOS
- [ ] Software - Library
### Type:
*Check one item.*
- [X] Bug
- [ ] Feature Enhancement
- [ ] Maintenance
- [ ] Question
- [ ] Documentation
### Applicable End Item:
*Check one item.*
- [ ] Top Level Assembly
- [ ] Lower Instrument Panel (LIP)
- [ ] Main Instrument Panel (MIP)
- [X] Left Console
- [ ] Right Console
- [ ] Seat
- [ ] Center Tub
- [ ] Flight Stick
- [ ] Throttle
- [ ] General Software
|
1.0
|
interconnect: Ant_Gnd connected to D39, not GND - ### Summary:
The ANT_GND of the ANT SEL PANEL should be shown connected to GND_J5-17. Instead, it is connected to D39.
### Screenshots:
<img width="528" alt="image" src="https://github.com/jrsteensen/OpenHornet/assets/62520296/654be750-73eb-4515-bb38-1f3f4f8e66af">
## More Information
*Add an "X" in the square brackets to check the applicable checkboxs.*
### Category:
*Check one or more items.*
- [ ] MCAD (SolidWorks)
- [X] ECAD (PCB Design or other electrical hardware)
- [ ] Software - Sketch
- [ ] Software - DCS-BIOS
- [ ] Software - Library
### Type:
*Check one item.*
- [X] Bug
- [ ] Feature Enhancement
- [ ] Maintenance
- [ ] Question
- [ ] Documentation
### Applicable End Item:
*Check one item.*
- [ ] Top Level Assembly
- [ ] Lower Instrument Panel (LIP)
- [ ] Main Instrument Panel (MIP)
- [X] Left Console
- [ ] Right Console
- [ ] Seat
- [ ] Center Tub
- [ ] Flight Stick
- [ ] Throttle
- [ ] General Software
|
priority
|
interconnect ant gnd connected to not gnd summary the ant gnd of the ant sel panel should be shown connected to gnd instead it is connected to screenshots img width alt image src more information add an x in the square brackets to check the applicable checkboxs category check one or more items mcad solidworks ecad pcb design or other electrical hardware software sketch software dcs bios software library type check one item bug feature enhancement maintenance question documentation applicable end item check one item top level assembly lower instrument panel lip main instrument panel mip left console right console seat center tub flight stick throttle general software
| 1
|
318,516
| 9,693,699,312
|
IssuesEvent
|
2019-05-24 16:47:45
|
inverse-inc/packetfence
|
https://api.github.com/repos/inverse-inc/packetfence
|
closed
|
Portal: Clicking multiple times on the signin submit is not handled properly
|
Priority: High Type: Bug
|
When doing sponsor or email registration and the server is a bit slow to respond (like when not using an SMTP relay), a user may be tempted to click again on the button generating multiple access codes and emails that will be sent.
The main issue is that only one of those emails contains a valid activation code so the user has to play a bit of russian roulette to get his activation completed.
This would be fixed by adding a piece of javascript in the signin form that disables the button and says "Processing..." or some similar text so that the user can't multi-submit that form
|
1.0
|
Portal: Clicking multiple times on the signin submit is not handled properly - When doing sponsor or email registration and the server is a bit slow to respond (like when not using an SMTP relay), a user may be tempted to click again on the button generating multiple access codes and emails that will be sent.
The main issue is that only one of those emails contains a valid activation code so the user has to play a bit of russian roulette to get his activation completed.
This would be fixed by adding a piece of javascript in the signin form that disables the button and says "Processing..." or some similar text so that the user can't multi-submit that form
|
priority
|
portal clicking multiple times on the signin submit is not handled properly when doing sponsor or email registration and the server is a bit slow to respond like when not using an smtp relay a user may be tempted to click again on the button generating multiple access codes and emails that will be sent the main issue is that only one of those emails contains a valid activation code so the user has to play a bit of russian roulette to get his activation completed this would be fixed by adding a piece of javascript in the signin form that disables the button and says processing or some similar text so that the user can t multi submit that form
| 1
|
770,275
| 27,035,484,683
|
IssuesEvent
|
2023-02-12 18:22:48
|
juno-fx/report
|
https://api.github.com/repos/juno-fx/report
|
closed
|
Unable to connect to instance 404 page not found
|
bug high priority
|
Launching an edge instance last night and now this morning results in me not being able to connect. It takes me to a page that just says 404 page not found.
|
1.0
|
Unable to connect to instance 404 page not found - Launching an edge instance last night and now this morning results in me not being able to connect. It takes me to a page that just says 404 page not found.
|
priority
|
unable to connect to instance page not found launching an edge instance last night and now this morning results in me not being able to connect it takes me to a page that just says page not found
| 1
|
305,470
| 9,369,828,949
|
IssuesEvent
|
2019-04-03 12:07:39
|
mojaloop/project
|
https://api.github.com/repos/mojaloop/project
|
closed
|
ML OSS Account Lookup Service (ALS) Design - 2
|
Priority: High Story
|
## **Goal**:
As a DFSP
I would like to have a design to _request a Participant detail based on a Party ID as per the Specification_
so that _a Party Identifier [currently MSISDN] can be used for looking up Participant details_
As a DFSP
I would like to have a design to _request a Party's details, as per the Specification_
so that _a Party Identifier [currently MSISDN] can be used for looking up Party details_
As a Hub Operator
I would like to have a design to _support looking up a Participant detail based on a Party ID as per the Specification_
so that _a Party Identifier [currently MSISDN] can be used for looking up Participant details_
As a Hub Operator
I would like to have a design to _support looking up a Party's details, as per the Specification_
so that _a Party Identifier [currently MSISDN] can be used for looking up Party details_
**Tasks**:
- [x] Refactor sequence diagrams for generic use and extensibility with the proposed design
- [x] DELETE /participant/{Type}/{ID}
- [x] Updates from DA Review (to the ALS designs)
- [x] Re-factored design to accommodate only a single switch for initial implementation
- [x] Updated DB Design & Assets
- [x] Updated Sequence Diagrams
- [x] Updated Arch Diagrams
- [x] Aligned swagger to Mojaloop specification
**Acceptance Criteria**:
- [x] The existing contributed ALS component is re-aligned to the OSS requirements, standards, etc determined by the ML DA
- [x] The ALS design caters for GET, POST and DEL Participants
- [x] The ALS design caters for GET Parties
**Pull Requests**:
- [ ] https://github.com/mojaloop/documentation/pull/15
**Follow-up**:
- N/A
**Dependencies**:
- [x] #655
**Accountability**:
- Owner: TBC
- QA/Review: DA
|
1.0
|
ML OSS Account Lookup Service (ALS) Design - 2 - ## **Goal**:
As a DFSP
I would like to have a design to _request a Participant detail based on a Party ID as per the Specification_
so that _a Party Identifier [currently MSISDN] can be used for looking up Participant details_
As a DFSP
I would like to have a design to _request a Party's details, as per the Specification_
so that _a Party Identifier [currently MSISDN] can be used for looking up Party details_
As a Hub Operator
I would like to have a design to _support looking up a Participant detail based on a Party ID as per the Specification_
so that _a Party Identifier [currently MSISDN] can be used for looking up Participant details_
As a Hub Operator
I would like to have a design to _support looking up a Party's details, as per the Specification_
so that _a Party Identifier [currently MSISDN] can be used for looking up Party details_
**Tasks**:
- [x] Refactor sequence diagrams for generic use and extensibility with the proposed design
- [x] DELETE /participant/{Type}/{ID}
- [x] Updates from DA Review (to the ALS designs)
- [x] Re-factored design to accommodate only a single switch for initial implementation
- [x] Updated DB Design & Assets
- [x] Updated Sequence Diagrams
- [x] Updated Arch Diagrams
- [x] Aligned swagger to Mojaloop specification
**Acceptance Criteria**:
- [x] The existing contributed ALS component is re-aligned to the OSS requirements, standards, etc determined by the ML DA
- [x] The ALS design caters for GET, POST and DEL Participants
- [x] The ALS design caters for GET Parties
**Pull Requests**:
- [ ] https://github.com/mojaloop/documentation/pull/15
**Follow-up**:
- N/A
**Dependencies**:
- [x] #655
**Accountability**:
- Owner: TBC
- QA/Review: DA
|
priority
|
ml oss account lookup service als design goal as a dfsp i would like to have a design to request a participant detail based on a party id as per the specification so that a party identifier can be used for looking up participant details as a dfsp i would like to have a design to request a party s details as per the specification so that a party identifier can be used for looking up party details as a hub operator i would like to have a design to support looking up a participant detail based on a party id as per the specification so that a party identifier can be used for looking up participant details as a hub operator i would like to have a design to support looking up a party s details as per the specification so that a party identifier can be used for looking up party details tasks refactor sequence diagrams for generic use and extensibility with the proposed design delete participant type id updates from da review to the als designs re factored design to accommodate only a single switch for initial implementation updated db design assets updated sequence diagrams updated arch diagrams aligned swagger to mojaloop specification acceptance criteria the existing contributed als component is re aligned to the oss requirements standards etc determined by the ml da the als design caters for get post and del participants the als design caters for get parties pull requests follow up n a dependencies accountability owner tbc qa review da
| 1
|
748,726
| 26,134,929,961
|
IssuesEvent
|
2022-12-29 10:49:08
|
nocodb/nocodb
|
https://api.github.com/repos/nocodb/nocodb
|
closed
|
🐛 Bug: Adding attachment is failing for editor role
|
🚀 Status: Ready for Next Release 🟠 Priority: High
|
### Please confirm if bug report does NOT exists already ?
- [X] I confirm there is no existing issue for this
### Steps to reproduce ?
- Editor role is unable to add attachment
### Desired Behavior
.
### Project Details
0.100.2
### Attachements
_No response_
|
1.0
|
🐛 Bug: Adding attachment is failing for editor role - ### Please confirm if bug report does NOT exists already ?
- [X] I confirm there is no existing issue for this
### Steps to reproduce ?
- Editor role is unable to add attachment
### Desired Behavior
.
### Project Details
0.100.2
### Attachements
_No response_
|
priority
|
🐛 bug adding attachment is failing for editor role please confirm if bug report does not exists already i confirm there is no existing issue for this steps to reproduce editor role is unable to add attachment desired behavior project details attachements no response
| 1
|
479,142
| 13,791,790,090
|
IssuesEvent
|
2020-10-09 12:42:38
|
AY2021S1-CS2113-T14-4/tp
|
https://api.github.com/repos/AY2021S1-CS2113-T14-4/tp
|
opened
|
Add UI class that enables printing of output messages after executing the commands given
|
priority.High type.Epic
|
Commands: adding user information, adding food item, listing food item and user details, calculating nutritional intake, deleting food items, clearing food list and exiting the program
|
1.0
|
Add UI class that enables printing of output messages after executing the commands given - Commands: adding user information, adding food item, listing food item and user details, calculating nutritional intake, deleting food items, clearing food list and exiting the program
|
priority
|
add ui class that enables printing of output messages after executing the commands given commands adding user information adding food item listing food item and user details calculating nutritional intake deleting food items clearing food list and exiting the program
| 1
|
197,185
| 6,953,267,655
|
IssuesEvent
|
2017-12-06 20:23:44
|
JacquesCarette/Drasil
|
https://api.github.com/repos/JacquesCarette/Drasil
|
closed
|
SSP - Codegen hangs indefinitely
|
bug High Priority
|
After finally fixing everything up so that it compiles and is runnable, I noticed that attempting to generate code for SSP will cause the execution to hang indefinitely. I assume there's some infinite recursion happening, but haven't been able to track down the error yet. I commented it out as of commit dd5e8fe8e046e07968712ed2fc5b907787cadac7.
|
1.0
|
SSP - Codegen hangs indefinitely - After finally fixing everything up so that it compiles and is runnable, I noticed that attempting to generate code for SSP will cause the execution to hang indefinitely. I assume there's some infinite recursion happening, but haven't been able to track down the error yet. I commented it out as of commit dd5e8fe8e046e07968712ed2fc5b907787cadac7.
|
priority
|
ssp codegen hangs indefinitely after finally fixing everything up so that it compiles and is runnable i noticed that attempting to generate code for ssp will cause the execution to hang indefinitely i assume there s some infinite recursion happening but haven t been able to track down the error yet i commented it out as of commit
| 1
|
701,128
| 24,087,259,935
|
IssuesEvent
|
2022-09-19 12:05:24
|
AgnostiqHQ/covalent
|
https://api.github.com/repos/AgnostiqHQ/covalent
|
closed
|
Build Covalent AMI and deploy via console and TF
|
devops priority / high Epic cloud Team East
|
- Build a Covalent AMI and deploy it to AWS
- Add a job in `release.yml` that builds and deploys the AMI for every stable release
- Put code in the terraform repo that will deploy Covalent on EC2 using the Covalent AMI, such that a user can clone the repo and `terraform apply` to deploy Covalent
|
1.0
|
Build Covalent AMI and deploy via console and TF - - Build a Covalent AMI and deploy it to AWS
- Add a job in `release.yml` that builds and deploys the AMI for every stable release
- Put code in the terraform repo that will deploy Covalent on EC2 using the Covalent AMI, such that a user can clone the repo and `terraform apply` to deploy Covalent
|
priority
|
build covalent ami and deploy via console and tf build a covalent ami and deploy it to aws add a job in release yml that builds and deploys the ami for every stable release put code in the terraform repo that will deploy covalent on using the covalent ami such that a user can clone the repo and terraform apply to deploy covalent
| 1
|
471,873
| 13,612,426,681
|
IssuesEvent
|
2020-09-23 10:16:33
|
AY2021S1-CS2103T-T13-1/tp
|
https://api.github.com/repos/AY2021S1-CS2103T-T13-1/tp
|
closed
|
[US] View locations with a high risk of infection
|
Developer Guide priority.High type.Story
|
As a healthcare official involved with setting up SafeEntry checkpoints, I want to identify locations with a high risk of infection so that I can know which places need these checkpoints the most.
|
1.0
|
[US] View locations with a high risk of infection - As a healthcare official involved with setting up SafeEntry checkpoints, I want to identify locations with a high risk of infection so that I can know which places need these checkpoints the most.
|
priority
|
view locations with a high risk of infection as a healthcare official involved with setting up safeentry checkpoints i want to identify locations with a high risk of infection so that i can know which places need these checkpoints the most
| 1
|
760,320
| 26,636,767,769
|
IssuesEvent
|
2023-01-24 22:47:58
|
luanyluz/portifolio
|
https://api.github.com/repos/luanyluz/portifolio
|
closed
|
Criar estrutura inicial do projeto.
|
Priority: High Weight: 2 Type: Feature
|
##Descrição
Selecionar um template pronto e adicionar os arquivos para o projeto.
|
1.0
|
Criar estrutura inicial do projeto. - ##Descrição
Selecionar um template pronto e adicionar os arquivos para o projeto.
|
priority
|
criar estrutura inicial do projeto descrição selecionar um template pronto e adicionar os arquivos para o projeto
| 1
|
133,478
| 5,204,236,309
|
IssuesEvent
|
2017-01-24 15:05:45
|
spring-projects/spring-boot
|
https://api.github.com/repos/spring-projects/spring-boot
|
opened
|
Introduce an enum for web environment on SpringApplication
|
for: team-call priority: high type: enhancement
|
The introduction of Web Flux support in 2.0 will mean that we have a second kind of web application. As described in #8017, this makes `setWebEnvironment(boolean)` on `SpringApplication` insufficient. I think we need an enum. In 1.5, it'd only have two values: `NONE` and `SERVLET`. In 2.0 we can add `REACTIVE` or `WEB_FLUX` or whatever we want to call it. We'd then deprecate `setWebEnvironment(boolean)` in favour of a new method that takes the new enum.
We already have a `WebEnvironment` enum in the testing support with a different meaning so I think we may need another name. I wonder if it could be `WebApplication` with a corresponding `setWebApplication(WebApplication)` method on `SpringApplication`. `WebApplication` ties in quite nicely with `@ConditionalOnWebApplication` and could also complement more specific variants of that condition in 2.0 such as `@ConditionalOnServletWebApplication` and `@ConditionalOnReactiveWebApplication`.
|
1.0
|
Introduce an enum for web environment on SpringApplication - The introduction of Web Flux support in 2.0 will mean that we have a second kind of web application. As described in #8017, this makes `setWebEnvironment(boolean)` on `SpringApplication` insufficient. I think we need an enum. In 1.5, it'd only have two values: `NONE` and `SERVLET`. In 2.0 we can add `REACTIVE` or `WEB_FLUX` or whatever we want to call it. We'd then deprecate `setWebEnvironment(boolean)` in favour of a new method that takes the new enum.
We already have a `WebEnvironment` enum in the testing support with a different meaning so I think we may need another name. I wonder if it could be `WebApplication` with a corresponding `setWebApplication(WebApplication)` method on `SpringApplication`. `WebApplication` ties in quite nicely with `@ConditionalOnWebApplication` and could also complement more specific variants of that condition in 2.0 such as `@ConditionalOnServletWebApplication` and `@ConditionalOnReactiveWebApplication`.
|
priority
|
introduce an enum for web environment on springapplication the introduction of web flux support in will mean that we have a second kind of web application as described in this makes setwebenvironment boolean on springapplication insufficient i think we need an enum in it d only have two values none and servlet in we can add reactive or web flux or whatever we want to call it we d then deprecate setwebenvironment boolean in favour of a new method that takes the new enum we already have a webenvironment enum in the testing support with a different meaning so i think we may need another name i wonder if it could be webapplication with a corresponding setwebapplication webapplication method on springapplication webapplication ties in quite nicely with conditionalonwebapplication and could also complement more specific variants of that condition in such as conditionalonservletwebapplication and conditionalonreactivewebapplication
| 1
|
556,144
| 16,475,897,244
|
IssuesEvent
|
2021-05-24 05:13:27
|
apache/echarts
|
https://api.github.com/repos/apache/echarts
|
closed
|
Safari canvas error: CanvasRenderingContext2D.save() has been called without a matching restore() too many times. Ignoring save().
|
bug en priority: high
|
### Version
5.1.1
### Reproduction link
[https://echarts.apache.org/examples/en/editor.html?c=dynamic-data2](https://echarts.apache.org/examples/en/editor.html?c=dynamic-data2)
### Steps to reproduce
1. Open `https://echarts.apache.org/examples/en/editor.html?c=dynamic-data2` in Safari (e.g. Version 14.1)
2. Set interval-time to a smaller value like 100 or 10 to increase the update-interval
3. Open console and wait
This also happens on iPad and iPhone.
### What is expected?
It updates the plot without throwing an error after time.
### What is actually happening?
After waiting for some minutes, the browser will throw the following error: `CanvasRenderingContext2D.save() has been called without a matching restore() too many times. Ignoring save().`

<!-- This issue is generated by echarts-issue-helper. DO NOT REMOVE -->
<!-- This issue is in English. DO NOT REMOVE -->
|
1.0
|
Safari canvas error: CanvasRenderingContext2D.save() has been called without a matching restore() too many times. Ignoring save(). - ### Version
5.1.1
### Reproduction link
[https://echarts.apache.org/examples/en/editor.html?c=dynamic-data2](https://echarts.apache.org/examples/en/editor.html?c=dynamic-data2)
### Steps to reproduce
1. Open `https://echarts.apache.org/examples/en/editor.html?c=dynamic-data2` in Safari (e.g. Version 14.1)
2. Set interval-time to a smaller value like 100 or 10 to increase the update-interval
3. Open console and wait
This also happens on iPad and iPhone.
### What is expected?
It updates the plot without throwing an error after time.
### What is actually happening?
After waiting for some minutes, the browser will throw the following error: `CanvasRenderingContext2D.save() has been called without a matching restore() too many times. Ignoring save().`

<!-- This issue is generated by echarts-issue-helper. DO NOT REMOVE -->
<!-- This issue is in English. DO NOT REMOVE -->
|
priority
|
safari canvas error save has been called without a matching restore too many times ignoring save version reproduction link steps to reproduce open in safari e g version set interval time to a smaller value like or to increase the update interval open console and wait this also happens on ipad and iphone what is expected it updates the plot without throwing an error after time what is actually happening after waiting for some minutes the browser will throw the following error save has been called without a matching restore too many times ignoring save
| 1
|
112,012
| 4,500,827,910
|
IssuesEvent
|
2016-09-01 07:02:51
|
odalic/odalic-ui
|
https://api.github.com/repos/odalic/odalic-ui
|
closed
|
Edges should be longer by default, so that one can read the label
|
bug graphics priority: High ui
|
When I run Odalic for the books.csv file, I got:

The length of arrows is really small, no one can read the label. Please specify longer edges by default (relates with #59 ). I would expect something like below (in terms of the length of the edges):

|
1.0
|
Edges should be longer by default, so that one can read the label - When I run Odalic for the books.csv file, I got:

The length of arrows is really small, no one can read the label. Please specify longer edges by default (relates with #59 ). I would expect something like below (in terms of the length of the edges):

|
priority
|
edges should be longer by default so that one can read the label when i run odalic for the books csv file i got the length of arrows is really small no one can read the label please specify longer edges by default relates with i would expect something like below in terms of the length of the edges
| 1
|
381,073
| 11,273,157,986
|
IssuesEvent
|
2020-01-14 16:04:35
|
alfaro96/scikit-lr
|
https://api.github.com/repos/alfaro96/scikit-lr
|
closed
|
Support top-k rankings for the Label Ranking problem
|
enhancement hard high priority
|
#### Description
Now that the theoretical concepts are well-known, we can implement the top-k rankings for the Label Ranking problem.
In particular, we just need to consider that the top-k classes are missed and ranked in the last positions of the ranking when completing it.
|
1.0
|
Support top-k rankings for the Label Ranking problem - #### Description
Now that the theoretical concepts are well-known, we can implement the top-k rankings for the Label Ranking problem.
In particular, we just need to consider that the top-k classes are missed and ranked in the last positions of the ranking when completing it.
|
priority
|
support top k rankings for the label ranking problem description now that the theoretical concepts are well known we can implement the top k rankings for the label ranking problem in particular we just need to consider that the top k classes are missed and ranked in the last positions of the ranking when completing it
| 1
|
764,037
| 26,782,563,854
|
IssuesEvent
|
2023-01-31 22:37:57
|
dmwm/WMCore
|
https://api.github.com/repos/dmwm/WMCore
|
closed
|
Update HTCondor python bindings for WMAgent to the 9 series
|
WMAgent High Priority Feature change dependencies Stakeholders QPrio: High
|
**Impact of the new feature**
WMAgent
**Is your feature request related to a problem? Please describe.**
Starting from May 2023 the 9 series of HTCondor python bindings are about to be deprecated. The current issue is intended to keep track of the migration process for WMAgent.
**Describe the solution you'd like**
Migrate to the latest version
**Describe alternatives you've considered**
It must be done
**Additional context**
None
|
1.0
|
Update HTCondor python bindings for WMAgent to the 9 series - **Impact of the new feature**
WMAgent
**Is your feature request related to a problem? Please describe.**
Starting from May 2023 the 9 series of HTCondor python bindings are about to be deprecated. The current issue is intended to keep track of the migration process for WMAgent.
**Describe the solution you'd like**
Migrate to the latest version
**Describe alternatives you've considered**
It must be done
**Additional context**
None
|
priority
|
update htcondor python bindings for wmagent to the series impact of the new feature wmagent is your feature request related to a problem please describe starting from may the series of htcondor python bindings are about to be deprecated the current issue is intended to keep track of the migration process for wmagent describe the solution you d like migrate to the latest version describe alternatives you ve considered it must be done additional context none
| 1
|
813,841
| 30,475,461,977
|
IssuesEvent
|
2023-07-17 16:11:00
|
hubmapconsortium/ccf-asct-reporter
|
https://api.github.com/repos/hubmapconsortium/ccf-asct-reporter
|
closed
|
BUG: Exit playground leads to "Select one or more ASCT+B Tables"
|
HIGHEST PRIORITY
|
When I was trying to use the Playground feature of the ASCT+B reporter, it would not accept an uploaded .csv without error.
I double checked the table, adjusted as needed, tried again, but it still did not work.
So when I then tried to "exit playground" instead of being dumped back into the Reporter, I ended up at the "Select one or more ASCT+B Tables" dialog instead of just being in the Reporter to try again using the playground.
If I clicked CANCEL on the diaglog, it just put me back in the error situation of the previous playground and I could NOT exit the playground.
Only way out was completely going back to start window of selecting either "go to visualization" or "go to playground"
|
1.0
|
BUG: Exit playground leads to "Select one or more ASCT+B Tables" - When I was trying to use the Playground feature of the ASCT+B reporter, it would not accept an uploaded .csv without error.
I double checked the table, adjusted as needed, tried again, but it still did not work.
So when I then tried to "exit playground" instead of being dumped back into the Reporter, I ended up at the "Select one or more ASCT+B Tables" dialog instead of just being in the Reporter to try again using the playground.
If I clicked CANCEL on the diaglog, it just put me back in the error situation of the previous playground and I could NOT exit the playground.
Only way out was completely going back to start window of selecting either "go to visualization" or "go to playground"
|
priority
|
bug exit playground leads to select one or more asct b tables when i was trying to use the playground feature of the asct b reporter it would not accept an uploaded csv without error i double checked the table adjusted as needed tried again but it still did not work so when i then tried to exit playground instead of being dumped back into the reporter i ended up at the select one or more asct b tables dialog instead of just being in the reporter to try again using the playground if i clicked cancel on the diaglog it just put me back in the error situation of the previous playground and i could not exit the playground only way out was completely going back to start window of selecting either go to visualization or go to playground
| 1
|
639,009
| 20,744,652,995
|
IssuesEvent
|
2022-03-14 21:21:01
|
microsoft/AdaptiveCards
|
https://api.github.com/repos/microsoft/AdaptiveCards
|
reopened
|
[Website] Blog-Community Call- Youtube Video: <iframe> and <frame> elements have not an accessible name.
|
Platform-JavaScript High Priority Area-Website Accessibility-Sev2-Scorecard
|
### Target Platforms
NodeJS
### SDK Version
https://adaptivecards.io/blog/2019/Community-Call-December/ 9/14/2021
### Application Name
Website
### Problem Description
[ADO 36108480](https://microsoft.visualstudio.com/OS/_workitems/edit/36108480)
Repro Steps:
1. Launch the above URL.
2. Run Accessibility Insight Tool of Web.
3. Observe the issue.
Actual Result:
`<iframe>` and `<frame>` elements have not an accessible name.
Expected Result:
Ensure `<iframe>` and `<frame>` elements have an accessible name.
### Screenshots
_No response_
### Card JSON
```JSON
No card
```
### Sample Code Language
_No response_
### Sample Code
_No response_
|
1.0
|
[Website] Blog-Community Call- Youtube Video: <iframe> and <frame> elements have not an accessible name. - ### Target Platforms
NodeJS
### SDK Version
https://adaptivecards.io/blog/2019/Community-Call-December/ 9/14/2021
### Application Name
Website
### Problem Description
[ADO 36108480](https://microsoft.visualstudio.com/OS/_workitems/edit/36108480)
Repro Steps:
1. Launch the above URL.
2. Run Accessibility Insight Tool of Web.
3. Observe the issue.
Actual Result:
`<iframe>` and `<frame>` elements have not an accessible name.
Expected Result:
Ensure `<iframe>` and `<frame>` elements have an accessible name.
### Screenshots
_No response_
### Card JSON
```JSON
No card
```
### Sample Code Language
_No response_
### Sample Code
_No response_
|
priority
|
blog community call youtube video and elements have not an accessible name target platforms nodejs sdk version application name website problem description repro steps launch the above url run accessibility insight tool of web observe the issue actual result and elements have not an accessible name expected result ensure and elements have an accessible name screenshots no response card json json no card sample code language no response sample code no response
| 1
|
589,997
| 17,768,456,710
|
IssuesEvent
|
2021-08-30 10:35:58
|
ita-social-projects/OoS-Backend
|
https://api.github.com/repos/ita-social-projects/OoS-Backend
|
opened
|
[Date field] Date-Month digits are replaced in the 'Дата народження керівника' field when it is corrected
|
bug priority:high sev:minor Type:Functional
|
**Environment:** Windows 10 Home, Google Chrome 92.0.4515.159.
**Reproducible:** always.
**Build found:** 30/08/2021 12:40
**Preconditions**
A user has already started registration as a Service Provider. The 'РЕЄСТРАЦІЯ НОВОГО ЗАКЛАДУ' page is opened.
**Steps to reproduce**
1. Fill the 'Дата народження керівника' field with digits of day and month fields < 12 (e.g. 01.09.XXXX).
2. Pay attention to the location of filled digits of date and month.
2. Change digits of a year.
3. Click anywhere on the page and pay attention to the location of the Date-Month digits after that.
**Actual result**
Date-Month digits are replaced in the 'Дата народження керівника' field when it is corrected.


**Expected result**
Digits of the Date field stay exactly as the user wants after changing.
**User story and test case links**
([#213 Story](https://github.com/ita-social-projects/OoS-Backend/issues/213))"
**Labels to be added**
"Bug", Priority ("pri: "), Severity ("severity:"), Type ("UI, "Functional"), "API" (for back-end bugs).
|
1.0
|
[Date field] Date-Month digits are replaced in the 'Дата народження керівника' field when it is corrected - **Environment:** Windows 10 Home, Google Chrome 92.0.4515.159.
**Reproducible:** always.
**Build found:** 30/08/2021 12:40
**Preconditions**
A user has already started registration as a Service Provider. The 'РЕЄСТРАЦІЯ НОВОГО ЗАКЛАДУ' page is opened.
**Steps to reproduce**
1. Fill the 'Дата народження керівника' field with digits of day and month fields < 12 (e.g. 01.09.XXXX).
2. Pay attention to the location of filled digits of date and month.
2. Change digits of a year.
3. Click anywhere on the page and pay attention to the location of the Date-Month digits after that.
**Actual result**
Date-Month digits are replaced in the 'Дата народження керівника' field when it is corrected.


**Expected result**
Digits of the Date field stay exactly as the user wants after changing.
**User story and test case links**
([#213 Story](https://github.com/ita-social-projects/OoS-Backend/issues/213))"
**Labels to be added**
"Bug", Priority ("pri: "), Severity ("severity:"), Type ("UI, "Functional"), "API" (for back-end bugs).
|
priority
|
date month digits are replaced in the дата народження керівника field when it is corrected environment windows home google chrome reproducible always build found preconditions a user has already started registration as a service provider the реєстрація нового закладу page is opened steps to reproduce fill the дата народження керівника field with digits of day and month fields e g xxxx pay attention to the location of filled digits of date and month change digits of a year click anywhere on the page and pay attention to the location of the date month digits after that actual result date month digits are replaced in the дата народження керівника field when it is corrected expected result digits of the date field stay exactly as the user wants after changing user story and test case links labels to be added bug priority pri severity severity type ui functional api for back end bugs
| 1
|
810,353
| 30,237,745,592
|
IssuesEvent
|
2023-07-06 11:28:03
|
recro/ats-issues
|
https://api.github.com/repos/recro/ats-issues
|
reopened
|
Interview Activity - selecting date and time for interview
|
priority/high area/application area/dashboard
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe the problem.
When moving a candidate application to the stage "interview scheduled" there is no option to select a date and time of when that interview is to take place.
### Describe the solution you'd like
- I'd like to be able to move an application to the "interview scheduled" stage and be able to select the date and time that interview is set to take place.
- I'd also like to be able to edit this in the event the interview changes/gets rescheduled.
- "interview scheduled" should be an activity that is captured in the activity table on the dashboard for each user based on the date the interview is scheduled for.
### Additional context
_No response_
|
1.0
|
Interview Activity - selecting date and time for interview - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe the problem.
When moving a candidate application to the stage "interview scheduled" there is no option to select a date and time of when that interview is to take place.
### Describe the solution you'd like
- I'd like to be able to move an application to the "interview scheduled" stage and be able to select the date and time that interview is set to take place.
- I'd also like to be able to edit this in the event the interview changes/gets rescheduled.
- "interview scheduled" should be an activity that is captured in the activity table on the dashboard for each user based on the date the interview is scheduled for.
### Additional context
_No response_
|
priority
|
interview activity selecting date and time for interview is there an existing issue for this i have searched the existing issues is your feature request related to a problem please describe the problem when moving a candidate application to the stage interview scheduled there is no option to select a date and time of when that interview is to take place describe the solution you d like i d like to be able to move an application to the interview scheduled stage and be able to select the date and time that interview is set to take place i d also like to be able to edit this in the event the interview changes gets rescheduled interview scheduled should be an activity that is captured in the activity table on the dashboard for each user based on the date the interview is scheduled for additional context no response
| 1
|
543,199
| 15,878,884,557
|
IssuesEvent
|
2021-04-09 11:40:55
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
Please make sure client-truststore.jks contains all the up to date trusted CAs
|
Complexity/Medium Priority/High Severity/Major task
|
Can we verify this in every product release ?
|
1.0
|
Please make sure client-truststore.jks contains all the up to date trusted CAs - Can we verify this in every product release ?
|
priority
|
please make sure client truststore jks contains all the up to date trusted cas can we verify this in every product release
| 1
|
535,176
| 15,683,684,521
|
IssuesEvent
|
2021-03-25 09:05:48
|
tantivy-search/tantivy
|
https://api.github.com/repos/tantivy-search/tantivy
|
closed
|
Encode the cardinal product of facet fields when it is small
|
enhancement high priority quickwit
|
Right now facets are encoded as a list of facet term ordinal in a multiple valued fast field.
This is neat... but if the number of facets is small, it is much more efficient to store a single facet field that contains
the "facet set". This is especially relevant for logs, where some of the facets values might be not independent at all (e.g.
log level and logging class).
Ideally we should detect this situation in the serializer and encode the facets in a different fashion in that case.
|
1.0
|
Encode the cardinal product of facet fields when it is small - Right now facets are encoded as a list of facet term ordinal in a multiple valued fast field.
This is neat... but if the number of facets is small, it is much more efficient to store a single facet field that contains
the "facet set". This is especially relevant for logs, where some of the facets values might be not independent at all (e.g.
log level and logging class).
Ideally we should detect this situation in the serializer and encode the facets in a different fashion in that case.
|
priority
|
encode the cardinal product of facet fields when it is small right now facets are encoded as a list of facet term ordinal in a multiple valued fast field this is neat but if the number of facets is small it is much more efficient to store a single facet field that contains the facet set this is especially relevant for logs where some of the facets values might be not independent at all e g log level and logging class ideally we should detect this situation in the serializer and encode the facets in a different fashion in that case
| 1
|
648,623
| 21,190,381,260
|
IssuesEvent
|
2022-04-08 16:42:37
|
InlinedLambdas/AstralFlow
|
https://api.github.com/repos/InlinedLambdas/AstralFlow
|
closed
|
[BUG] Cancelled events should be considered
|
bug good first issue priority: high
|
Obviously most events can be cancelled. But in listeners such as [BlockListener](https://github.com/InlinedLambdas/AstralFlow/blob/b2286cf152f9ab6cb431ef6f281fd19af12ff9cd/src/main/java/io/ib67/astralflow/internal/listener/BlockListener.java#L94) and the whole process, cancelled events are not handled correctly, which may lead to severe bugs.
|
1.0
|
[BUG] Cancelled events should be considered - Obviously most events can be cancelled. But in listeners such as [BlockListener](https://github.com/InlinedLambdas/AstralFlow/blob/b2286cf152f9ab6cb431ef6f281fd19af12ff9cd/src/main/java/io/ib67/astralflow/internal/listener/BlockListener.java#L94) and the whole process, cancelled events are not handled correctly, which may lead to severe bugs.
|
priority
|
cancelled events should be considered obviously most events can be cancelled but in listeners such as and the whole process cancelled events are not handled correctly which may lead to severe bugs
| 1
|
543,303
| 15,879,684,749
|
IssuesEvent
|
2021-04-09 12:47:38
|
wso2/product-apim
|
https://api.github.com/repos/wso2/product-apim
|
closed
|
Carbon-db oracle_rac script syntax error
|
API-M 4.0.0 Feature/Databases Priority/Highest Severity/Critical Type/Bug
|
### Description:
The oracle_rac.sql script to create Shared_DB doesn't work due to a syntax issue (closing bracket at wrong place) in line 37. It should be corrected as below.
CREATE TABLE REG_PATH(
REG_PATH_ID INTEGER NULL,
REG_PATH_VALUE VARCHAR2(2000) NOT NULL,
REG_PATH_PARENT_ID INTEGER,
REG_TENANT_ID INTEGER DEFAULT 0,
CONSTRAINT PK_PATH PRIMARY KEY(REG_PATH_ID, REG_TENANT_ID),
CONSTRAINT UNIQUE_REG_PATH_TENANT_ID UNIQUE (REG_PATH_VALUE,REG_TENANT_ID))
### Steps to reproduce:
### Affected Product Version:
apim-3.1.0
### Environment details (with versions):
- OS:
- Client:
- Env (Docker/K8s):
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
bug
#### Suggested Assignees:
<!--Only to be used by non-members-->
|
1.0
|
Carbon-db oracle_rac script syntax error - ### Description:
The oracle_rac.sql script to create Shared_DB doesn't work due to a syntax issue (closing bracket at wrong place) in line 37. It should be corrected as below.
CREATE TABLE REG_PATH(
REG_PATH_ID INTEGER NULL,
REG_PATH_VALUE VARCHAR2(2000) NOT NULL,
REG_PATH_PARENT_ID INTEGER,
REG_TENANT_ID INTEGER DEFAULT 0,
CONSTRAINT PK_PATH PRIMARY KEY(REG_PATH_ID, REG_TENANT_ID),
CONSTRAINT UNIQUE_REG_PATH_TENANT_ID UNIQUE (REG_PATH_VALUE,REG_TENANT_ID))
### Steps to reproduce:
### Affected Product Version:
apim-3.1.0
### Environment details (with versions):
- OS:
- Client:
- Env (Docker/K8s):
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
bug
#### Suggested Assignees:
<!--Only to be used by non-members-->
|
priority
|
carbon db oracle rac script syntax error description the oracle rac sql script to create shared db doesn t work due to a syntax issue closing bracket at wrong place in line it should be corrected as below create table reg path reg path id integer null reg path value not null reg path parent id integer reg tenant id integer default constraint pk path primary key reg path id reg tenant id constraint unique reg path tenant id unique reg path value reg tenant id steps to reproduce affected product version apim environment details with versions os client env docker optional fields related issues suggested labels bug suggested assignees
| 1
|
789,644
| 27,799,530,403
|
IssuesEvent
|
2023-03-17 14:53:54
|
wp-media/wp-rocket
|
https://api.github.com/repos/wp-media/wp-rocket
|
opened
|
3.13 / Refactor Delay JS tweaks
|
priority: high
|
**Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version => 3.13_alpha1
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
While exploratory testing, test execution and team discussion, We see that those points needs to be handled before the release
- [ ] Rollback version needs to be updated to 3.12.6.1
- [ ] Bump minimal WordPress version to 5.8
- [ ] Lists should be sorted alphabetically
- [ ] Checked exclusion still applied even after disable related plugin/theme
**To Reproduce**
Steps to reproduce the behavior:
For Checked exclusion still applied even after disable related plugin
1- check ninja form in DJS and save => jquery is excluded from DJS
2- deactivate ninja plugin
3- clear cache and check home page => jquery still excluded from DJS (if we re-enabled DJS , jquery won't be excluded)
**Expected behavior**
Exclusions of deactivated plugin/theme isnot applied
**Additional context**
Add any other context about the problem here.
Reference sheet https://docs.google.com/spreadsheets/d/1AhroWjZKXEpqtIWJdFIcBcvpeE961Gw-LM20wh8GxqA/edit#gid=0
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
|
1.0
|
3.13 / Refactor Delay JS tweaks - **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version => 3.13_alpha1
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
While exploratory testing, test execution and team discussion, We see that those points needs to be handled before the release
- [ ] Rollback version needs to be updated to 3.12.6.1
- [ ] Bump minimal WordPress version to 5.8
- [ ] Lists should be sorted alphabetically
- [ ] Checked exclusion still applied even after disable related plugin/theme
**To Reproduce**
Steps to reproduce the behavior:
For Checked exclusion still applied even after disable related plugin
1- check ninja form in DJS and save => jquery is excluded from DJS
2- deactivate ninja plugin
3- clear cache and check home page => jquery still excluded from DJS (if we re-enabled DJS , jquery won't be excluded)
**Expected behavior**
Exclusions of deactivated plugin/theme isnot applied
**Additional context**
Add any other context about the problem here.
Reference sheet https://docs.google.com/spreadsheets/d/1AhroWjZKXEpqtIWJdFIcBcvpeE961Gw-LM20wh8GxqA/edit#gid=0
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
|
priority
|
refactor delay js tweaks before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version used the search feature to ensure that the bug hasn’t been reported before describe the bug while exploratory testing test execution and team discussion we see that those points needs to be handled before the release rollback version needs to be updated to bump minimal wordpress version to lists should be sorted alphabetically checked exclusion still applied even after disable related plugin theme to reproduce steps to reproduce the behavior for checked exclusion still applied even after disable related plugin check ninja form in djs and save jquery is excluded from djs deactivate ninja plugin clear cache and check home page jquery still excluded from djs if we re enabled djs jquery won t be excluded expected behavior exclusions of deactivated plugin theme isnot applied additional context add any other context about the problem here reference sheet backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort
| 1
|
140,917
| 5,426,028,304
|
IssuesEvent
|
2017-03-03 08:47:41
|
UnknownShadow200/ClassicalSharp
|
https://api.github.com/repos/UnknownShadow200/ClassicalSharp
|
opened
|
AdvLighting leaks light from corners
|
bug high priority
|
for example from a bird's eye view
Stone | air
wood | stone
The wood will still get some 'light' from the air, even though it's blocked by both stone blocks.
|
1.0
|
AdvLighting leaks light from corners - for example from a bird's eye view
Stone | air
wood | stone
The wood will still get some 'light' from the air, even though it's blocked by both stone blocks.
|
priority
|
advlighting leaks light from corners for example from a bird s eye view stone air wood stone the wood will still get some light from the air even though it s blocked by both stone blocks
| 1
|
245,174
| 7,882,197,160
|
IssuesEvent
|
2018-06-26 21:41:24
|
pelias/pelias
|
https://api.github.com/repos/pelias/pelias
|
closed
|
Search endpoint does not respect `sources` flag for admin queries
|
High Priority bug
|
Performing a search query for a known admin area specifying the `sources=geonames` parameter returns Who's on First results anyways:
http://pelias.github.io/compare/#/v1/search%3Fsources=gn&text=montana

While its possible that we will in the future stop importing Geonames (especially admin areas, as they will have concordances to Who's on First), we need to ensure that API behavior is clear in the meantime.
Ideally, we would still be able to return Geonames results, since they're still imported for now. Worst case, we should return a warning or error saying that we are no longer returning Geonames admin results. However, we should definitely not return sources other than those specified under any circumstances.
|
1.0
|
Search endpoint does not respect `sources` flag for admin queries - Performing a search query for a known admin area specifying the `sources=geonames` parameter returns Who's on First results anyways:
http://pelias.github.io/compare/#/v1/search%3Fsources=gn&text=montana

While its possible that we will in the future stop importing Geonames (especially admin areas, as they will have concordances to Who's on First), we need to ensure that API behavior is clear in the meantime.
Ideally, we would still be able to return Geonames results, since they're still imported for now. Worst case, we should return a warning or error saying that we are no longer returning Geonames admin results. However, we should definitely not return sources other than those specified under any circumstances.
|
priority
|
search endpoint does not respect sources flag for admin queries performing a search query for a known admin area specifying the sources geonames parameter returns who s on first results anyways while its possible that we will in the future stop importing geonames especially admin areas as they will have concordances to who s on first we need to ensure that api behavior is clear in the meantime ideally we would still be able to return geonames results since they re still imported for now worst case we should return a warning or error saying that we are no longer returning geonames admin results however we should definitely not return sources other than those specified under any circumstances
| 1
|
646,952
| 21,082,803,698
|
IssuesEvent
|
2022-04-03 06:29:46
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
TLS certificate of download.pytorch.org is expired
|
high priority triage review
|
The TLS certificate that is used, for example, at https://download.pytorch.org/whl/torch_stable.html is expired as of right now.
cc @ezyang @gchanan @zou3519
|
1.0
|
TLS certificate of download.pytorch.org is expired - The TLS certificate that is used, for example, at https://download.pytorch.org/whl/torch_stable.html is expired as of right now.
cc @ezyang @gchanan @zou3519
|
priority
|
tls certificate of download pytorch org is expired the tls certificate that is used for example at is expired as of right now cc ezyang gchanan
| 1
|
248,966
| 7,947,716,193
|
IssuesEvent
|
2018-07-11 04:37:25
|
magda-io/magda
|
https://api.github.com/repos/magda-io/magda
|
closed
|
Eliminate DES-CBC3-SHA Weak Cypher
|
priority: high
|
### Problem description
Apparently we're accepting the DES-CBC3-SHA cypher on search.gov.au, which is vulnerable to the FREAK attack. We need to figure out how to remove it... most likely we should just change search.data.gov.au over to the Google CDN, then make sure that's not allowing this cypher.
```
Supported Server Cipher(s):
Preferred TLSv1.2 128 bits ECDHE-RSA-AES128-GCM-SHA256 Curve P-256 DHE 256
Accepted TLSv1.2 128 bits ECDHE-RSA-AES128-SHA256 Curve P-256 DHE 256
Accepted TLSv1.2 128 bits ECDHE-RSA-AES128-SHA Curve P-256 DHE 256
Accepted TLSv1.2 256 bits ECDHE-RSA-AES256-GCM-SHA384 Curve P-256 DHE 256
Accepted TLSv1.2 256 bits ECDHE-RSA-AES256-SHA384
Accepted TLSv1.2 256 bits ECDHE-RSA-AES256-SHA
Accepted TLSv1.2 128 bits AES128-GCM-SHA256
Accepted TLSv1.2 256 bits AES256-GCM-SHA384
Accepted TLSv1.2 128 bits AES128-SHA256
Accepted TLSv1.2 256 bits AES256-SHA
Accepted TLSv1.2 128 bits AES128-SHA
*Accepted TLSv1.2 112 bits DES-CBC3-SHA*
Preferred
TLSv1.1 128 bits ECDHE-RSA-AES128-SHA
Accepted TLSv1.1 256 bits ECDHE-RSA-AES256-SHA
Accepted TLSv1.1 256 bits AES256-SHA
Accepted TLSv1.1 128 bits AES128-SHA
*Accepted TLSv1.1 112 bits DES-CBC3-SHA*
Preferred TLSv1.0 128 bits ECDHE-RSA-AES128-SHA
Accepted TLSv1.0 256 bits ECDHE-RSA-AES256-SHA
Accepted TLSv1.0 256 bits AES256-SHA
Accepted TLSv1.0 128 bits AES128-SHA
```
|
1.0
|
Eliminate DES-CBC3-SHA Weak Cypher - ### Problem description
Apparently we're accepting the DES-CBC3-SHA cypher on search.gov.au, which is vulnerable to the FREAK attack. We need to figure out how to remove it... most likely we should just change search.data.gov.au over to the Google CDN, then make sure that's not allowing this cypher.
```
Supported Server Cipher(s):
Preferred TLSv1.2 128 bits ECDHE-RSA-AES128-GCM-SHA256 Curve P-256 DHE 256
Accepted TLSv1.2 128 bits ECDHE-RSA-AES128-SHA256 Curve P-256 DHE 256
Accepted TLSv1.2 128 bits ECDHE-RSA-AES128-SHA Curve P-256 DHE 256
Accepted TLSv1.2 256 bits ECDHE-RSA-AES256-GCM-SHA384 Curve P-256 DHE 256
Accepted TLSv1.2 256 bits ECDHE-RSA-AES256-SHA384
Accepted TLSv1.2 256 bits ECDHE-RSA-AES256-SHA
Accepted TLSv1.2 128 bits AES128-GCM-SHA256
Accepted TLSv1.2 256 bits AES256-GCM-SHA384
Accepted TLSv1.2 128 bits AES128-SHA256
Accepted TLSv1.2 256 bits AES256-SHA
Accepted TLSv1.2 128 bits AES128-SHA
*Accepted TLSv1.2 112 bits DES-CBC3-SHA*
Preferred
TLSv1.1 128 bits ECDHE-RSA-AES128-SHA
Accepted TLSv1.1 256 bits ECDHE-RSA-AES256-SHA
Accepted TLSv1.1 256 bits AES256-SHA
Accepted TLSv1.1 128 bits AES128-SHA
*Accepted TLSv1.1 112 bits DES-CBC3-SHA*
Preferred TLSv1.0 128 bits ECDHE-RSA-AES128-SHA
Accepted TLSv1.0 256 bits ECDHE-RSA-AES256-SHA
Accepted TLSv1.0 256 bits AES256-SHA
Accepted TLSv1.0 128 bits AES128-SHA
```
|
priority
|
eliminate des sha weak cypher problem description apparently we re accepting the des sha cypher on search gov au which is vulnerable to the freak attack we need to figure out how to remove it most likely we should just change search data gov au over to the google cdn then make sure that s not allowing this cypher supported server cipher s preferred bits ecdhe rsa gcm curve p dhe accepted bits ecdhe rsa curve p dhe accepted bits ecdhe rsa sha curve p dhe accepted bits ecdhe rsa gcm curve p dhe accepted bits ecdhe rsa accepted bits ecdhe rsa sha accepted bits gcm accepted bits gcm accepted bits accepted bits sha accepted bits sha accepted bits des sha preferred bits ecdhe rsa sha accepted bits ecdhe rsa sha accepted bits sha accepted bits sha accepted bits des sha preferred bits ecdhe rsa sha accepted bits ecdhe rsa sha accepted bits sha accepted bits sha
| 1
|
398,668
| 11,742,155,346
|
IssuesEvent
|
2020-03-11 23:47:59
|
BoiseState/CS481-X-S20-capstone-bsu-transit
|
https://api.github.com/repos/BoiseState/CS481-X-S20-capstone-bsu-transit
|
reopened
|
Create CRUD Operations for Employees
|
Priority-High Story
|
As a developer, I want to create `Employee` CRUD endpoints in our API, so that we can deliver and manage `Employee` resources from our front end user interface.
**Acceptance Criteria:**
- [ ] Given that an `/employees` endpoint exists, when the `/employees` endpoint is requested with method `GET`, then an array of JSON `Employee` objects is received by the requester.
- [ ] Given that an `/employees/{id}` endpoint exists, when the `/employees/{id}` endpoint is requested with a unique `id` and request method type `GET`, then a JSON `Employee` object is received by the requester.
- [ ] Given that an `/employees` endpoint exists, when the `/employees` endpoint is requested with method `POST` and a valid JSON `Employee` object resides in the request body, a new `Employee` object is created and stored in the database.
- [ ] Given that an `/employees/{id}` endpoint exists, when the `/employees/{id}` endpoint is requested with a unique `id` and request method type `DELETE`, then the `Employee` record matching the unique `id` will be removed from the database
- [ ] Given that an `/employees/{id}` endpoint exists, when the `/employees/{id}` endpoint is requested with a unique `id`, a valid JSON `Employee` object resides in the request body, and request method type `PUT`, then the `Employee` record matching the unique `id` will be updated with the data residing in the request body
- [ ] Given that an `/employees` endpoint exists, when all expected functionality is tested and completed, proper exception responses must be implemented for incorrect endpoint usage:
- POST `/employees` : 400 Bad Request
- GET `/employees` & `/employees/{id}` : 404 Not found if resource doesn't exist or 403 Forbidden if user is not allowed access
- DELETE `/employees/{id}` : 404 Not found if resource doesn't exist or 403 Forbidden if user is not allowed access
- PUT `/employees/{id}` : 404 Not found if resource doesn't exist or 403 Forbidden if user is not allowed access
|
1.0
|
Create CRUD Operations for Employees - As a developer, I want to create `Employee` CRUD endpoints in our API, so that we can deliver and manage `Employee` resources from our front end user interface.
**Acceptance Criteria:**
- [ ] Given that an `/employees` endpoint exists, when the `/employees` endpoint is requested with method `GET`, then an array of JSON `Employee` objects is received by the requester.
- [ ] Given that an `/employees/{id}` endpoint exists, when the `/employees/{id}` endpoint is requested with a unique `id` and request method type `GET`, then a JSON `Employee` object is received by the requester.
- [ ] Given that an `/employees` endpoint exists, when the `/employees` endpoint is requested with method `POST` and a valid JSON `Employee` object resides in the request body, a new `Employee` object is created and stored in the database.
- [ ] Given that an `/employees/{id}` endpoint exists, when the `/employees/{id}` endpoint is requested with a unique `id` and request method type `DELETE`, then the `Employee` record matching the unique `id` will be removed from the database
- [ ] Given that an `/employees/{id}` endpoint exists, when the `/employees/{id}` endpoint is requested with a unique `id`, a valid JSON `Employee` object resides in the request body, and request method type `PUT`, then the `Employee` record matching the unique `id` will be updated with the data residing in the request body
- [ ] Given that an `/employees` endpoint exists, when all expected functionality is tested and completed, proper exception responses must be implemented for incorrect endpoint usage:
- POST `/employees` : 400 Bad Request
- GET `/employees` & `/employees/{id}` : 404 Not found if resource doesn't exist or 403 Forbidden if user is not allowed access
- DELETE `/employees/{id}` : 404 Not found if resource doesn't exist or 403 Forbidden if user is not allowed access
- PUT `/employees/{id}` : 404 Not found if resource doesn't exist or 403 Forbidden if user is not allowed access
|
priority
|
create crud operations for employees as a developer i want to create employee crud endpoints in our api so that we can deliver and manage employee resources from our front end user interface acceptance criteria given that an employees endpoint exists when the employees endpoint is requested with method get then an array of json employee objects is received by the requester given that an employees id endpoint exists when the employees id endpoint is requested with a unique id and request method type get then a json employee object is received by the requester given that an employees endpoint exists when the employees endpoint is requested with method post and a valid json employee object resides in the request body a new employee object is created and stored in the database given that an employees id endpoint exists when the employees id endpoint is requested with a unique id and request method type delete then the employee record matching the unique id will be removed from the database given that an employees id endpoint exists when the employees id endpoint is requested with a unique id a valid json employee object resides in the request body and request method type put then the employee record matching the unique id will be updated with the data residing in the request body given that an employees endpoint exists when all expected functionality is tested and completed proper exception responses must be implemented for incorrect endpoint usage post employees bad request get employees employees id not found if resource doesn t exist or forbidden if user is not allowed access delete employees id not found if resource doesn t exist or forbidden if user is not allowed access put employees id not found if resource doesn t exist or forbidden if user is not allowed access
| 1
|
816,039
| 30,585,506,963
|
IssuesEvent
|
2023-07-21 13:04:40
|
Qiskit/qiskit-terra
|
https://api.github.com/repos/Qiskit/qiskit-terra
|
closed
|
Full support for control flow in `transpile`
|
type: feature request priority: high type: epic mod: transpiler
|
### What should we add?
Over the course of the epic:
- #8630
we added support for control-flow to the transpiler at optimisation levels 0 and 1. For full support, we need to bring the missing passes up-to-speed as well. This epic tracks the other components to achieve that.
For the purposes of this issue, we particularly need `SabreSwap` upgrading. For optimisation level 2, we need the optimisation passes `CommutationAnalysis` and `CommutativeCancellation`, and optimisation level 3 (and some less common transpiler options) make up the rest. We do _not_ need to add support for control-flow to all of the scheduling passes right now, because Qiskit control-flow is only supported by IBM hardware at the moment, and scheduling this is handled by the IBM provider.
When these are updated, we should also make sure that the corresponding gated checks in the preset pass-manager constructors are corrected.
Sub-epics:
- [x] #9431
- [x] #9432
Tracked issues:
- [x] #9419
- [x] #9421
- [x] #9422
- [x] #9423
- [x] #9425
- [x] #9426
Low priority related issues - these should not be prioritised, and are not required for this epic:
- [ ] #9429
- [ ] #9430
|
1.0
|
Full support for control flow in `transpile` - ### What should we add?
Over the course of the epic:
- #8630
we added support for control-flow to the transpiler at optimisation levels 0 and 1. For full support, we need to bring the missing passes up-to-speed as well. This epic tracks the other components to achieve that.
For the purposes of this issue, we particularly need `SabreSwap` upgrading. For optimisation level 2, we need the optimisation passes `CommutationAnalysis` and `CommutativeCancellation`, and optimisation level 3 (and some less common transpiler options) make up the rest. We do _not_ need to add support for control-flow to all of the scheduling passes right now, because Qiskit control-flow is only supported by IBM hardware at the moment, and scheduling this is handled by the IBM provider.
When these are updated, we should also make sure that the corresponding gated checks in the preset pass-manager constructors are corrected.
Sub-epics:
- [x] #9431
- [x] #9432
Tracked issues:
- [x] #9419
- [x] #9421
- [x] #9422
- [x] #9423
- [x] #9425
- [x] #9426
Low priority related issues - these should not be prioritised, and are not required for this epic:
- [ ] #9429
- [ ] #9430
|
priority
|
full support for control flow in transpile what should we add over the course of the epic we added support for control flow to the transpiler at optimisation levels and for full support we need to bring the missing passes up to speed as well this epic tracks the other components to achieve that for the purposes of this issue we particularly need sabreswap upgrading for optimisation level we need the optimisation passes commutationanalysis and commutativecancellation and optimisation level and some less common transpiler options make up the rest we do not need to add support for control flow to all of the scheduling passes right now because qiskit control flow is only supported by ibm hardware at the moment and scheduling this is handled by the ibm provider when these are updated we should also make sure that the corresponding gated checks in the preset pass manager constructors are corrected sub epics tracked issues low priority related issues these should not be prioritised and are not required for this epic
| 1
|
219,742
| 7,345,532,063
|
IssuesEvent
|
2018-03-07 17:42:21
|
EvictionLab/eviction-maps
|
https://api.github.com/repos/EvictionLab/eviction-maps
|
opened
|
Disable evictors tab
|
enhancement high priority
|
Since this won't be included in the initial launch, we need to re-add the disabled style
|
1.0
|
Disable evictors tab - Since this won't be included in the initial launch, we need to re-add the disabled style
|
priority
|
disable evictors tab since this won t be included in the initial launch we need to re add the disabled style
| 1
|
326,831
| 9,961,592,328
|
IssuesEvent
|
2019-07-07 06:31:02
|
dhis2/maintenance-app
|
https://api.github.com/repos/dhis2/maintenance-app
|
closed
|
Configurable columns for list
|
enhancement priority:high stale wontfix
|
Make the columns to display in object lists configurable. Currently the app displays name, public access and last updated properties.
It should be possible to select among the following properties.
## Properties
- Name (selected by default)
- Short name (where applicable)
- Code
- UID
- Public access (selected by default)
- Last updated (selected by default)
- Created
- Owner (user)
|
1.0
|
Configurable columns for list - Make the columns to display in object lists configurable. Currently the app displays name, public access and last updated properties.
It should be possible to select among the following properties.
## Properties
- Name (selected by default)
- Short name (where applicable)
- Code
- UID
- Public access (selected by default)
- Last updated (selected by default)
- Created
- Owner (user)
|
priority
|
configurable columns for list make the columns to display in object lists configurable currently the app displays name public access and last updated properties it should be possible to select among the following properties properties name selected by default short name where applicable code uid public access selected by default last updated selected by default created owner user
| 1
|
364,390
| 10,763,655,874
|
IssuesEvent
|
2019-11-01 05:06:25
|
alexieyizhe/intern.plus
|
https://api.github.com/repos/alexieyizhe/intern.plus
|
closed
|
Set up testing framework
|
effort: medium priority: high
|
Cypress has been set up for end to end tests.
Still to do:
- [x] Run on CI
- [x] Set up ability to mock graphql server data (use `apollo-mock-client` pkg)
- [x] Write mocks for all data
- [x] Companies details
- [x] Jobs details
- [x] Reviews details
- [x] resolvers for [companies|jobs|reviews]List
- [x] Ensure mock data from Faker is consistent between reloads using seed (for tests)
- [x] Add Percy snapshots
- [ ] ~Set up framework to write unit tests (see #70 for options)~
- [ ] ~Write basic unit test~
|
1.0
|
Set up testing framework - Cypress has been set up for end to end tests.
Still to do:
- [x] Run on CI
- [x] Set up ability to mock graphql server data (use `apollo-mock-client` pkg)
- [x] Write mocks for all data
- [x] Companies details
- [x] Jobs details
- [x] Reviews details
- [x] resolvers for [companies|jobs|reviews]List
- [x] Ensure mock data from Faker is consistent between reloads using seed (for tests)
- [x] Add Percy snapshots
- [ ] ~Set up framework to write unit tests (see #70 for options)~
- [ ] ~Write basic unit test~
|
priority
|
set up testing framework cypress has been set up for end to end tests still to do run on ci set up ability to mock graphql server data use apollo mock client pkg write mocks for all data companies details jobs details reviews details resolvers for list ensure mock data from faker is consistent between reloads using seed for tests add percy snapshots set up framework to write unit tests see for options write basic unit test
| 1
|
169,109
| 6,395,048,286
|
IssuesEvent
|
2017-08-04 12:10:17
|
UGXaero/UGXrealms
|
https://api.github.com/repos/UGXaero/UGXrealms
|
closed
|
Server crash!
|
bug high priority server crash
|
```
2017-08-03 22:36:13: ERROR[Main]: ServerError: AsyncErr: ServerThread::run Lua: Runtime error from mod 'default' in callback luaentity_Step(): .../.minetest/worlds/UGXrealms/worldmods/maptools/tools.lua:65: attempt to index local 'puncher' (a nil value)
2017-08-03 22:36:13: ERROR[Main]: Stack Traceback
2017-08-03 22:36:13: ERROR[Main]: ===============
2017-08-03 22:36:13: ERROR[Main]: (2) Lua local 'callback' at file '/home/minetestservers/.minetest/worlds/UGXrealms/worldmods/maptools/tools.lua:65'
2017-08-03 22:36:13: ERROR[Main]: Local variables:
2017-08-03 22:36:13: ERROR[Main]: pos = table: 0x413025c0 {y:-13.7, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: node = table: 0x413026c8 {param1:0, name:default:silver_sand, param2:0}
2017-08-03 22:36:13: ERROR[Main]: puncher = nil
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = table: 0x41303660 {y:-14, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = nil
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = number: 0
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = number: 11
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = table: 0x413048c0 {v:11, p:table: 0x41303660}
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = string: "attempt to index local 'puncher' (a nil value)"
2017-08-03 22:36:13: ERROR[Main]: (3) Lua function '(anonymous)' at file '/usr/local/share/minetest/builtin/game/falling.lua:103' (best guess)
2017-08-03 22:36:13: ERROR[Main]: Local variables:
2017-08-03 22:36:13: ERROR[Main]: self = table: 0x40ffb7e8 {object:userdata: 0x40ffb7d8, meta:table: 0x40ffbe78, node:table: 0x40ffb9a8}
2017-08-03 22:36:13: ERROR[Main]: dtime = number: 0.1
2017-08-03 22:36:13: ERROR[Main]: acceleration = table: 0x413022c0 {y:-10, x:0, z:0}
2017-08-03 22:36:13: ERROR[Main]: pos = table: 0x41302468 {y:-14, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: bcp = table: 0x41302490 {y:-14.7, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: bcn = table: 0x41302598 {param1:0, name:default:stone, param2:0}
2017-08-03 22:36:13: ERROR[Main]: bcd = table: 0x41238a98 {tiles:table: 0x41238408, legacy_mineral:true, groups:table: 0x41238ac0, drop:default:cobble, sounds:table: 0x4188ce10, mod_origin:default, type:node, name:default:stone, description:Stone}
2017-08-03 22:36:13: ERROR[Main]: np = table: 0x413025c0 {y:-13.7, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: n2 = table: 0x413026c8 {param1:0, name:default:silver_sand, param2:0}
2017-08-03 22:36:13: ERROR[Main]: nd = table: 0x40906240 {tiles:table: 0x40906268, sounds:table: 0x40906330, mod_origin:default, groups:table: 0x409062a0, type:node, name:default:silver_sand, description:Silver Sand}
2017-08-03 22:36:13: ERROR[Main]: (for generator) = C function: next
2017-08-03 22:36:13: ERROR[Main]: (for state) = table: 0x40d78bc0 {1:function: 0x402b68f8, 2:function: 0x40fe06b8, 3:function: 0x40ff2c28, 4:function: 0x4071c288}
2017-08-03 22:36:13: ERROR[Main]: (for control) = number: nan
2017-08-03 22:36:13: ERROR[Main]: _ = number: 2
2017-08-03 22:36:13: ERROR[Main]: callback = Lua function '(anonymous)' (defined at line 64 of chunk .../.minetest/worlds/UGXrealms/worldmods/maptools/tools.lua)
```
:large_orange_diamond:
|
1.0
|
Server crash! - ```
2017-08-03 22:36:13: ERROR[Main]: ServerError: AsyncErr: ServerThread::run Lua: Runtime error from mod 'default' in callback luaentity_Step(): .../.minetest/worlds/UGXrealms/worldmods/maptools/tools.lua:65: attempt to index local 'puncher' (a nil value)
2017-08-03 22:36:13: ERROR[Main]: Stack Traceback
2017-08-03 22:36:13: ERROR[Main]: ===============
2017-08-03 22:36:13: ERROR[Main]: (2) Lua local 'callback' at file '/home/minetestservers/.minetest/worlds/UGXrealms/worldmods/maptools/tools.lua:65'
2017-08-03 22:36:13: ERROR[Main]: Local variables:
2017-08-03 22:36:13: ERROR[Main]: pos = table: 0x413025c0 {y:-13.7, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: node = table: 0x413026c8 {param1:0, name:default:silver_sand, param2:0}
2017-08-03 22:36:13: ERROR[Main]: puncher = nil
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = table: 0x41303660 {y:-14, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = nil
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = number: 0
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = number: 11
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = table: 0x413048c0 {v:11, p:table: 0x41303660}
2017-08-03 22:36:13: ERROR[Main]: (*temporary) = string: "attempt to index local 'puncher' (a nil value)"
2017-08-03 22:36:13: ERROR[Main]: (3) Lua function '(anonymous)' at file '/usr/local/share/minetest/builtin/game/falling.lua:103' (best guess)
2017-08-03 22:36:13: ERROR[Main]: Local variables:
2017-08-03 22:36:13: ERROR[Main]: self = table: 0x40ffb7e8 {object:userdata: 0x40ffb7d8, meta:table: 0x40ffbe78, node:table: 0x40ffb9a8}
2017-08-03 22:36:13: ERROR[Main]: dtime = number: 0.1
2017-08-03 22:36:13: ERROR[Main]: acceleration = table: 0x413022c0 {y:-10, x:0, z:0}
2017-08-03 22:36:13: ERROR[Main]: pos = table: 0x41302468 {y:-14, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: bcp = table: 0x41302490 {y:-14.7, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: bcn = table: 0x41302598 {param1:0, name:default:stone, param2:0}
2017-08-03 22:36:13: ERROR[Main]: bcd = table: 0x41238a98 {tiles:table: 0x41238408, legacy_mineral:true, groups:table: 0x41238ac0, drop:default:cobble, sounds:table: 0x4188ce10, mod_origin:default, type:node, name:default:stone, description:Stone}
2017-08-03 22:36:13: ERROR[Main]: np = table: 0x413025c0 {y:-13.7, x:16392, z:20490}
2017-08-03 22:36:13: ERROR[Main]: n2 = table: 0x413026c8 {param1:0, name:default:silver_sand, param2:0}
2017-08-03 22:36:13: ERROR[Main]: nd = table: 0x40906240 {tiles:table: 0x40906268, sounds:table: 0x40906330, mod_origin:default, groups:table: 0x409062a0, type:node, name:default:silver_sand, description:Silver Sand}
2017-08-03 22:36:13: ERROR[Main]: (for generator) = C function: next
2017-08-03 22:36:13: ERROR[Main]: (for state) = table: 0x40d78bc0 {1:function: 0x402b68f8, 2:function: 0x40fe06b8, 3:function: 0x40ff2c28, 4:function: 0x4071c288}
2017-08-03 22:36:13: ERROR[Main]: (for control) = number: nan
2017-08-03 22:36:13: ERROR[Main]: _ = number: 2
2017-08-03 22:36:13: ERROR[Main]: callback = Lua function '(anonymous)' (defined at line 64 of chunk .../.minetest/worlds/UGXrealms/worldmods/maptools/tools.lua)
```
:large_orange_diamond:
|
priority
|
server crash error servererror asyncerr serverthread run lua runtime error from mod default in callback luaentity step minetest worlds ugxrealms worldmods maptools tools lua attempt to index local puncher a nil value error stack traceback error error lua local callback at file home minetestservers minetest worlds ugxrealms worldmods maptools tools lua error local variables error pos table y x z error node table name default silver sand error puncher nil error temporary table y x z error temporary nil error temporary number error temporary number error temporary table v p table error temporary string attempt to index local puncher a nil value error lua function anonymous at file usr local share minetest builtin game falling lua best guess error local variables error self table object userdata meta table node table error dtime number error acceleration table y x z error pos table y x z error bcp table y x z error bcn table name default stone error bcd table tiles table legacy mineral true groups table drop default cobble sounds table mod origin default type node name default stone description stone error np table y x z error table name default silver sand error nd table tiles table sounds table mod origin default groups table type node name default silver sand description silver sand error for generator c function next error for state table function function function function error for control number nan error number error callback lua function anonymous defined at line of chunk minetest worlds ugxrealms worldmods maptools tools lua large orange diamond
| 1
|
448,096
| 12,943,574,983
|
IssuesEvent
|
2020-07-18 07:22:28
|
BuckleScript/bucklescript
|
https://api.github.com/repos/BuckleScript/bucklescript
|
closed
|
bsc.exe: unknown option '-c' error in `.re` files with bs-platform 8.1.1
|
PRIORITY:HIGH
|
I'm getting the following error at the top of every .re file in a project once I started using bs-platform 8.1.1 (calling it with NPX, `npx bsb -make-world -w`):
```
<project_path>/node_modules/bs-platform/darwin/bsc.exe: unknown option '-c'.
Usage: bsc <options> <files>
Options are:
-bs-super-errors Better error message combined with other tools
-unboxed-types unannotated unboxable types will be unboxed
-bs-re-out Print compiler output in Reason syntax
-bs-jsx Set jsx version
-bs-refmt Set customized refmt path
-bs-gentype Pass gentype command
-bs-suffix Set suffix to .bs.js
-bs-no-implicit-include Don't include current dir implicitly
-bs-read-cmi (internal) Assume mli always exist
-bs-D Define conditional variable e.g, -D DEBUG=true
-bs-unsafe-empty-array Allow [||] to be polymorphic
-nostdlib Don't use stdlib
-bs-internal-check Built in check corrupted data
-bs-list-conditionals List existing conditional variables
-bs-binary-ast Generate binary .mli_ast and ml_ast
-bs-simple-binary-ast Generate binary .mliast_simple and mlast_simple
-bs-syntax-only only check syntax
-bs-eval (experimental) Set the string to be evaluated in OCaml syntax
-e (experimental) Set the string to be evaluated in ReasonML syntax
-bs-cmi-only Stop after generating cmi file
-bs-cmi Not using cached cmi, always generate cmi
-bs-cmj Not using cached cmj, always generate cmj
-as-ppx As ppx for editor integration
-bs-g debug mode
-bs-sort-imports Sort the imports by lexical order so the output will be more stable (default false)
-bs-no-sort-imports No sort (see -bs-sort-imports)
-bs-package-name set package name, useful when you want to produce npm packages
-bs-ns set package map, not only set package name but also use it as a namespace
-bs-no-version-header Don't print version header
-bs-package-output set npm-output-path: [opt_module]:path, for example: 'lib/cjs', 'amdjs:lib/amdjs', 'es6:lib/es6'
-bs-no-builtin-ppx disable built-in ppx (internal use)
-bs-cross-module-opt enable cross module inlining(experimental), default(false)
-bs-no-cross-module-opt disable cross module inlining(experimental)
-bs-diagnose More verbose output
-bs-no-check-div-by-zero unsafe mode, don't check div by zero and mod by zero
-bs-noassertfalse no code for assert false
-noassert Do not compile assertion checks
-bs-loc dont display location with -dtypedtree, -dparsetree
-impl <file> Compile <file> as a .ml file
-intf <file> Compile <file> as a .mli file
-dtypedtree debug typedtree
-dparsetree debug parsetree
-drawlambda debug raw lambda
-dsource print source
-fmt (internal) format as Res syntax
-where Print location of standard library and exit
-ppx <command> Pipe abstract syntax trees through preprocessor <command>
-open <module> Opens the module <module> before typing
-verbose Print calls to external commands
-keep-locs Keep locations in .cmi files
-no-keep-locs Do not keep locations in .cmi files
-nopervasives (undocumented)
-v Print compiler version and location of standard library and exit
-version Print version and exit
-I <dir> Add <dir> to the list of include directories
-pp <command> Pipe sources through preprocessor <command>
-absname Show absolute filenames in error messages
-bs-no-bin-annot disable binary annotations (by default on)
-i Print inferred interface
-nolabels Ignore non-optional labels in types
-no-alias-deps Do not record dependencies for module aliases
-o <file> Set output file name to <file>
-principal Check principality of type inference
-short-paths Shorten paths in types
-unsafe Do not compile bounds checking on array and string access
-w <list> Enable or disable warnings according to <list>:
+<spec> enable warnings in <spec>
-<spec> disable warnings in <spec>
@<spec> enable warnings in <spec> and treat them as errors
<spec> can be:
<num> a single warning number
<num1>..<num2> a range of consecutive warning numbers
<letter> a predefined set
default setting is +a-4-9-20-40-41-42-50-61-102
-warn-error <list> Enable or disable error status for warnings according
to <list>. See option -w for the syntax of <list>.
Default setting is -a+5+6+101
-warn-help Show description of warning numbers
-color {auto|always|never} Enable or disable colors in compiler messages
The following settings are supported:
auto use heuristics to enable colors only if supported
always enable colors
never disable colors
The default setting is 'auto', and the current heuristic
checks that the TERM environment variable exists and is
not empty or "dumb", and that isatty(stderr) holds.
-help Display this list of options
--help Display this list of options
```
|
1.0
|
bsc.exe: unknown option '-c' error in `.re` files with bs-platform 8.1.1 - I'm getting the following error at the top of every .re file in a project once I started using bs-platform 8.1.1 (calling it with NPX, `npx bsb -make-world -w`):
```
<project_path>/node_modules/bs-platform/darwin/bsc.exe: unknown option '-c'.
Usage: bsc <options> <files>
Options are:
-bs-super-errors Better error message combined with other tools
-unboxed-types unannotated unboxable types will be unboxed
-bs-re-out Print compiler output in Reason syntax
-bs-jsx Set jsx version
-bs-refmt Set customized refmt path
-bs-gentype Pass gentype command
-bs-suffix Set suffix to .bs.js
-bs-no-implicit-include Don't include current dir implicitly
-bs-read-cmi (internal) Assume mli always exist
-bs-D Define conditional variable e.g, -D DEBUG=true
-bs-unsafe-empty-array Allow [||] to be polymorphic
-nostdlib Don't use stdlib
-bs-internal-check Built in check corrupted data
-bs-list-conditionals List existing conditional variables
-bs-binary-ast Generate binary .mli_ast and ml_ast
-bs-simple-binary-ast Generate binary .mliast_simple and mlast_simple
-bs-syntax-only only check syntax
-bs-eval (experimental) Set the string to be evaluated in OCaml syntax
-e (experimental) Set the string to be evaluated in ReasonML syntax
-bs-cmi-only Stop after generating cmi file
-bs-cmi Not using cached cmi, always generate cmi
-bs-cmj Not using cached cmj, always generate cmj
-as-ppx As ppx for editor integration
-bs-g debug mode
-bs-sort-imports Sort the imports by lexical order so the output will be more stable (default false)
-bs-no-sort-imports No sort (see -bs-sort-imports)
-bs-package-name set package name, useful when you want to produce npm packages
-bs-ns set package map, not only set package name but also use it as a namespace
-bs-no-version-header Don't print version header
-bs-package-output set npm-output-path: [opt_module]:path, for example: 'lib/cjs', 'amdjs:lib/amdjs', 'es6:lib/es6'
-bs-no-builtin-ppx disable built-in ppx (internal use)
-bs-cross-module-opt enable cross module inlining(experimental), default(false)
-bs-no-cross-module-opt disable cross module inlining(experimental)
-bs-diagnose More verbose output
-bs-no-check-div-by-zero unsafe mode, don't check div by zero and mod by zero
-bs-noassertfalse no code for assert false
-noassert Do not compile assertion checks
-bs-loc dont display location with -dtypedtree, -dparsetree
-impl <file> Compile <file> as a .ml file
-intf <file> Compile <file> as a .mli file
-dtypedtree debug typedtree
-dparsetree debug parsetree
-drawlambda debug raw lambda
-dsource print source
-fmt (internal) format as Res syntax
-where Print location of standard library and exit
-ppx <command> Pipe abstract syntax trees through preprocessor <command>
-open <module> Opens the module <module> before typing
-verbose Print calls to external commands
-keep-locs Keep locations in .cmi files
-no-keep-locs Do not keep locations in .cmi files
-nopervasives (undocumented)
-v Print compiler version and location of standard library and exit
-version Print version and exit
-I <dir> Add <dir> to the list of include directories
-pp <command> Pipe sources through preprocessor <command>
-absname Show absolute filenames in error messages
-bs-no-bin-annot disable binary annotations (by default on)
-i Print inferred interface
-nolabels Ignore non-optional labels in types
-no-alias-deps Do not record dependencies for module aliases
-o <file> Set output file name to <file>
-principal Check principality of type inference
-short-paths Shorten paths in types
-unsafe Do not compile bounds checking on array and string access
-w <list> Enable or disable warnings according to <list>:
+<spec> enable warnings in <spec>
-<spec> disable warnings in <spec>
@<spec> enable warnings in <spec> and treat them as errors
<spec> can be:
<num> a single warning number
<num1>..<num2> a range of consecutive warning numbers
<letter> a predefined set
default setting is +a-4-9-20-40-41-42-50-61-102
-warn-error <list> Enable or disable error status for warnings according
to <list>. See option -w for the syntax of <list>.
Default setting is -a+5+6+101
-warn-help Show description of warning numbers
-color {auto|always|never} Enable or disable colors in compiler messages
The following settings are supported:
auto use heuristics to enable colors only if supported
always enable colors
never disable colors
The default setting is 'auto', and the current heuristic
checks that the TERM environment variable exists and is
not empty or "dumb", and that isatty(stderr) holds.
-help Display this list of options
--help Display this list of options
```
|
priority
|
bsc exe unknown option c error in re files with bs platform i m getting the following error at the top of every re file in a project once i started using bs platform calling it with npx npx bsb make world w node modules bs platform darwin bsc exe unknown option c usage bsc options are bs super errors better error message combined with other tools unboxed types unannotated unboxable types will be unboxed bs re out print compiler output in reason syntax bs jsx set jsx version bs refmt set customized refmt path bs gentype pass gentype command bs suffix set suffix to bs js bs no implicit include don t include current dir implicitly bs read cmi internal assume mli always exist bs d define conditional variable e g d debug true bs unsafe empty array allow to be polymorphic nostdlib don t use stdlib bs internal check built in check corrupted data bs list conditionals list existing conditional variables bs binary ast generate binary mli ast and ml ast bs simple binary ast generate binary mliast simple and mlast simple bs syntax only only check syntax bs eval experimental set the string to be evaluated in ocaml syntax e experimental set the string to be evaluated in reasonml syntax bs cmi only stop after generating cmi file bs cmi not using cached cmi always generate cmi bs cmj not using cached cmj always generate cmj as ppx as ppx for editor integration bs g debug mode bs sort imports sort the imports by lexical order so the output will be more stable default false bs no sort imports no sort see bs sort imports bs package name set package name useful when you want to produce npm packages bs ns set package map not only set package name but also use it as a namespace bs no version header don t print version header bs package output set npm output path path for example lib cjs amdjs lib amdjs lib bs no builtin ppx disable built in ppx internal use bs cross module opt enable cross module inlining experimental default false bs no cross module opt disable cross module inlining experimental bs diagnose more verbose output bs no check div by zero unsafe mode don t check div by zero and mod by zero bs noassertfalse no code for assert false noassert do not compile assertion checks bs loc dont display location with dtypedtree dparsetree impl compile as a ml file intf compile as a mli file dtypedtree debug typedtree dparsetree debug parsetree drawlambda debug raw lambda dsource print source fmt internal format as res syntax where print location of standard library and exit ppx pipe abstract syntax trees through preprocessor open opens the module before typing verbose print calls to external commands keep locs keep locations in cmi files no keep locs do not keep locations in cmi files nopervasives undocumented v print compiler version and location of standard library and exit version print version and exit i add to the list of include directories pp pipe sources through preprocessor absname show absolute filenames in error messages bs no bin annot disable binary annotations by default on i print inferred interface nolabels ignore non optional labels in types no alias deps do not record dependencies for module aliases o set output file name to principal check principality of type inference short paths shorten paths in types unsafe do not compile bounds checking on array and string access w enable or disable warnings according to enable warnings in disable warnings in enable warnings in and treat them as errors can be a single warning number a range of consecutive warning numbers a predefined set default setting is a warn error enable or disable error status for warnings according to see option w for the syntax of default setting is a warn help show description of warning numbers color auto always never enable or disable colors in compiler messages the following settings are supported auto use heuristics to enable colors only if supported always enable colors never disable colors the default setting is auto and the current heuristic checks that the term environment variable exists and is not empty or dumb and that isatty stderr holds help display this list of options help display this list of options
| 1
|
475,887
| 13,727,470,631
|
IssuesEvent
|
2020-10-04 06:37:13
|
AY2021S1-CS2103T-F12-3/tp
|
https://api.github.com/repos/AY2021S1-CS2103T-F12-3/tp
|
opened
|
As a user, I can calculate the total grade based on results with different percentages
|
priority.High type.Story
|
so that I can see what grade I am currently at based on continual assessments done.
|
1.0
|
As a user, I can calculate the total grade based on results with different percentages - so that I can see what grade I am currently at based on continual assessments done.
|
priority
|
as a user i can calculate the total grade based on results with different percentages so that i can see what grade i am currently at based on continual assessments done
| 1
|
311,786
| 9,539,193,993
|
IssuesEvent
|
2019-04-30 16:20:12
|
dice-group/Squirrel
|
https://api.github.com/repos/dice-group/Squirrel
|
opened
|
Add the option to persist or not the queue
|
component:frontier priority:high type:enhancement
|
In the frontier, it would be nice to have an option to persist or not the queue on disk after the frontier terminator.
Suggestion: create an environment variable for it and drop or not the collections when the queue is being closed.
|
1.0
|
Add the option to persist or not the queue - In the frontier, it would be nice to have an option to persist or not the queue on disk after the frontier terminator.
Suggestion: create an environment variable for it and drop or not the collections when the queue is being closed.
|
priority
|
add the option to persist or not the queue in the frontier it would be nice to have an option to persist or not the queue on disk after the frontier terminator suggestion create an environment variable for it and drop or not the collections when the queue is being closed
| 1
|
373,177
| 11,034,877,495
|
IssuesEvent
|
2019-12-07 09:20:40
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
Build with MKLDNN broken
|
high priority module: mkldnn topic: dependency bug triaged
|
## 🐛 Bug
Source build fails because of MKL-DNN's use of `-Werror` and there being a warning thrown in the code
## To Reproduce
```
export CMAKE_PREFIX_PATH=${CONDA_PREFIX:-"$(dirname $(which conda))/../"}
REL_WITH_DEB_INFO=1 NO_CAFFE2_OPS=1 NO_CUDA=1 USE_FBGEMM=1 time python setup.py develop
```
```
[2/1073] Building CXX object third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o
FAILED: third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o
/usr/lib64/ccache/c++ -DMKLDNN_DLL -DMKLDNN_DLL_EXPORTS -DMKLDNN_THR=MKLDNN_THR_OMP -DONNX_ML=1 -DONNX_NAMESPACE=onnx_torch -DTH_BLAS_MKL -DUSE_C11_ATOMICS=1 -DUSE_MKL -D__STDC_CONSTANT_MACROS -D__STDC_LIMIT_MACROS -I../third_party/protobuf/src -I../cmake/../third_party/benchmark/include -I../third_party/onnx -Ithird_party/onnx -I../third_party/foxi -Ithird_party/foxi -I../third_party/ideep/mkl-dnn/include -Ithird_party/ideep/mkl-dnn/include -I../third_party/ideep/mkl-dnn/src -I../third_party/ideep/mkl-dnn/src/common -I../third_party/ideep/mkl-dnn/src/cpu -I../third_party/ideep/mkl-dnn/src/cpu/xbyak -I../third_party/ideep/mkl-dnn/src/../include -isystem third_party/gloo -isystem ../cmake/../third_party/gloo -isystem ../cmake/../third_party/googletest/googlemock/include -isystem ../cmake/../third_party/googletest/googletest/include -isystem /home/jamesreed/miniconda3/include -isystem ../third_party/gemmlowp -isystem ../third_party/neon2sse -isystem ../third_party -isystem ../cmake/../third_party/eigen -isystem /home/jamesreed/miniconda3/include/python3.7m -isystem /home/jamesreed/miniconda3/lib/python3.7/site-packages/numpy/core/include -isystem ../torch/include -Wno-deprecated -fvisibility-inlines-hidden -fopenmp -std=c++11 -fvisibility-inlines-hidden -Wall -Wno-unknown-pragmas -Werror -fvisibility=internal -march=native -mtune=native -fPIC -Wformat -Wformat-security -fstack-protector-strong -fopenmp -Wmissing-field-initializers -Wno-strict-overflow -O2 -g -DNDEBUG -fPIC -DCAFFE2_USE_GLOO -DHAVE_GCC_GET_CPUID -DUSE_AVX -DUSE_AVX2 -Wno-maybe-uninitialized -Wno-strict-overflow -Wno-error=strict-overflow -Wno-tautological-compare -std=gnu++11 -MD -MT third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o -MF third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o.d -o third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o -c ../third_party/ideep/mkl-dnn/src/common/utils.cpp
../third_party/ideep/mkl-dnn/src/common/utils.cpp: In function ‘int mkldnn::impl::mkldnn_getenv(char*, const char*, int)’:
../third_party/ideep/mkl-dnn/src/common/utils.cpp:50:24: error: ‘char* strncpy(char*, const char*, size_t)’ output truncated before terminating nul copying as many bytes from a string as its length [-Werror=stringop-truncation]
strncpy(value, buffer, value_length);
~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../third_party/ideep/mkl-dnn/src/common/utils.cpp:46:34: note: length computed here
value_length = strlen(buffer);
~~~~~~^~~~~~~~
cc1plus: all warnings being treated as errors
```
Build works fine with `USE_MKLDNN=0`.
## Expected behavior
Builds successfully
## Environment
Collecting environment information...
PyTorch version: N/A
Is debug build: N/A
CUDA used to build PyTorch: N/A
OS: Fedora release 29 (Twenty Nine)
GCC version: (GCC) 8.2.1 20180801 (Red Hat 8.2.1-2)
CMake version: version 3.14.0
Python version: 3.7
Is CUDA available: N/A
CUDA runtime version: 10.1.168
GPU models and configuration: GPU 0: GeForce RTX 2080
Nvidia driver version: 430.26
cuDNN version: Could not collect
Versions of relevant libraries:
[pip] numpy==1.16.4
[conda] blas 1.0 mkl
[conda] mkl 2019.4 243
[conda] mkl-include 2019.4 243
[conda] mkl_fft 1.0.12 py37ha843d7b_0
[conda] mkl_random 1.0.2 py37hd81dba3_0
## Additional context
|
1.0
|
Build with MKLDNN broken - ## 🐛 Bug
Source build fails because of MKL-DNN's use of `-Werror` and there being a warning thrown in the code
## To Reproduce
```
export CMAKE_PREFIX_PATH=${CONDA_PREFIX:-"$(dirname $(which conda))/../"}
REL_WITH_DEB_INFO=1 NO_CAFFE2_OPS=1 NO_CUDA=1 USE_FBGEMM=1 time python setup.py develop
```
```
[2/1073] Building CXX object third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o
FAILED: third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o
/usr/lib64/ccache/c++ -DMKLDNN_DLL -DMKLDNN_DLL_EXPORTS -DMKLDNN_THR=MKLDNN_THR_OMP -DONNX_ML=1 -DONNX_NAMESPACE=onnx_torch -DTH_BLAS_MKL -DUSE_C11_ATOMICS=1 -DUSE_MKL -D__STDC_CONSTANT_MACROS -D__STDC_LIMIT_MACROS -I../third_party/protobuf/src -I../cmake/../third_party/benchmark/include -I../third_party/onnx -Ithird_party/onnx -I../third_party/foxi -Ithird_party/foxi -I../third_party/ideep/mkl-dnn/include -Ithird_party/ideep/mkl-dnn/include -I../third_party/ideep/mkl-dnn/src -I../third_party/ideep/mkl-dnn/src/common -I../third_party/ideep/mkl-dnn/src/cpu -I../third_party/ideep/mkl-dnn/src/cpu/xbyak -I../third_party/ideep/mkl-dnn/src/../include -isystem third_party/gloo -isystem ../cmake/../third_party/gloo -isystem ../cmake/../third_party/googletest/googlemock/include -isystem ../cmake/../third_party/googletest/googletest/include -isystem /home/jamesreed/miniconda3/include -isystem ../third_party/gemmlowp -isystem ../third_party/neon2sse -isystem ../third_party -isystem ../cmake/../third_party/eigen -isystem /home/jamesreed/miniconda3/include/python3.7m -isystem /home/jamesreed/miniconda3/lib/python3.7/site-packages/numpy/core/include -isystem ../torch/include -Wno-deprecated -fvisibility-inlines-hidden -fopenmp -std=c++11 -fvisibility-inlines-hidden -Wall -Wno-unknown-pragmas -Werror -fvisibility=internal -march=native -mtune=native -fPIC -Wformat -Wformat-security -fstack-protector-strong -fopenmp -Wmissing-field-initializers -Wno-strict-overflow -O2 -g -DNDEBUG -fPIC -DCAFFE2_USE_GLOO -DHAVE_GCC_GET_CPUID -DUSE_AVX -DUSE_AVX2 -Wno-maybe-uninitialized -Wno-strict-overflow -Wno-error=strict-overflow -Wno-tautological-compare -std=gnu++11 -MD -MT third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o -MF third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o.d -o third_party/ideep/mkl-dnn/src/CMakeFiles/mkldnn.dir/common/utils.cpp.o -c ../third_party/ideep/mkl-dnn/src/common/utils.cpp
../third_party/ideep/mkl-dnn/src/common/utils.cpp: In function ‘int mkldnn::impl::mkldnn_getenv(char*, const char*, int)’:
../third_party/ideep/mkl-dnn/src/common/utils.cpp:50:24: error: ‘char* strncpy(char*, const char*, size_t)’ output truncated before terminating nul copying as many bytes from a string as its length [-Werror=stringop-truncation]
strncpy(value, buffer, value_length);
~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../third_party/ideep/mkl-dnn/src/common/utils.cpp:46:34: note: length computed here
value_length = strlen(buffer);
~~~~~~^~~~~~~~
cc1plus: all warnings being treated as errors
```
Build works fine with `USE_MKLDNN=0`.
## Expected behavior
Builds successfully
## Environment
Collecting environment information...
PyTorch version: N/A
Is debug build: N/A
CUDA used to build PyTorch: N/A
OS: Fedora release 29 (Twenty Nine)
GCC version: (GCC) 8.2.1 20180801 (Red Hat 8.2.1-2)
CMake version: version 3.14.0
Python version: 3.7
Is CUDA available: N/A
CUDA runtime version: 10.1.168
GPU models and configuration: GPU 0: GeForce RTX 2080
Nvidia driver version: 430.26
cuDNN version: Could not collect
Versions of relevant libraries:
[pip] numpy==1.16.4
[conda] blas 1.0 mkl
[conda] mkl 2019.4 243
[conda] mkl-include 2019.4 243
[conda] mkl_fft 1.0.12 py37ha843d7b_0
[conda] mkl_random 1.0.2 py37hd81dba3_0
## Additional context
|
priority
|
build with mkldnn broken 🐛 bug source build fails because of mkl dnn s use of werror and there being a warning thrown in the code to reproduce export cmake prefix path conda prefix dirname which conda rel with deb info no ops no cuda use fbgemm time python setup py develop building cxx object third party ideep mkl dnn src cmakefiles mkldnn dir common utils cpp o failed third party ideep mkl dnn src cmakefiles mkldnn dir common utils cpp o usr ccache c dmkldnn dll dmkldnn dll exports dmkldnn thr mkldnn thr omp donnx ml donnx namespace onnx torch dth blas mkl duse atomics duse mkl d stdc constant macros d stdc limit macros i third party protobuf src i cmake third party benchmark include i third party onnx ithird party onnx i third party foxi ithird party foxi i third party ideep mkl dnn include ithird party ideep mkl dnn include i third party ideep mkl dnn src i third party ideep mkl dnn src common i third party ideep mkl dnn src cpu i third party ideep mkl dnn src cpu xbyak i third party ideep mkl dnn src include isystem third party gloo isystem cmake third party gloo isystem cmake third party googletest googlemock include isystem cmake third party googletest googletest include isystem home jamesreed include isystem third party gemmlowp isystem third party isystem third party isystem cmake third party eigen isystem home jamesreed include isystem home jamesreed lib site packages numpy core include isystem torch include wno deprecated fvisibility inlines hidden fopenmp std c fvisibility inlines hidden wall wno unknown pragmas werror fvisibility internal march native mtune native fpic wformat wformat security fstack protector strong fopenmp wmissing field initializers wno strict overflow g dndebug fpic use gloo dhave gcc get cpuid duse avx duse wno maybe uninitialized wno strict overflow wno error strict overflow wno tautological compare std gnu md mt third party ideep mkl dnn src cmakefiles mkldnn dir common utils cpp o mf third party ideep mkl dnn src cmakefiles mkldnn dir common utils cpp o d o third party ideep mkl dnn src cmakefiles mkldnn dir common utils cpp o c third party ideep mkl dnn src common utils cpp third party ideep mkl dnn src common utils cpp in function ‘int mkldnn impl mkldnn getenv char const char int ’ third party ideep mkl dnn src common utils cpp error ‘char strncpy char const char size t ’ output truncated before terminating nul copying as many bytes from a string as its length strncpy value buffer value length third party ideep mkl dnn src common utils cpp note length computed here value length strlen buffer all warnings being treated as errors build works fine with use mkldnn expected behavior builds successfully environment collecting environment information pytorch version n a is debug build n a cuda used to build pytorch n a os fedora release twenty nine gcc version gcc red hat cmake version version python version is cuda available n a cuda runtime version gpu models and configuration gpu geforce rtx nvidia driver version cudnn version could not collect versions of relevant libraries numpy blas mkl mkl mkl include mkl fft mkl random additional context
| 1
|
701,277
| 24,093,804,395
|
IssuesEvent
|
2022-09-19 16:52:24
|
MaibornWolff/codecharta
|
https://api.github.com/repos/MaibornWolff/codecharta
|
closed
|
Delta Mode not working for codecharta.cc.json
|
bug pr-visualization priority:high
|
# Bug
## Expected Behavior
GIVEN
- current codecharta.cc.json
- old map: Unzip and upload: [codecharta_2022-04-28_08-53.cc.zip](https://github.com/MaibornWolff/codecharta/files/8581208/codecharta_2022-04-28_08-53.cc.zip)
WHEN
comparing them in the delta mode
THEN
I can see the actual deltas.
## Actual Behavior
the two maps are shown side by side and no deltas are shown
## Screenshots

## Specifications
- is released in [online-demo](https://maibornwolff.github.io/codecharta/visualization/app/index.html?file=codecharta.cc.json&file=codecharta_analysis.cc.json): yes
- CodeCharta Version: 1.95.2
- OS: all
- Browser: all
|
1.0
|
Delta Mode not working for codecharta.cc.json - # Bug
## Expected Behavior
GIVEN
- current codecharta.cc.json
- old map: Unzip and upload: [codecharta_2022-04-28_08-53.cc.zip](https://github.com/MaibornWolff/codecharta/files/8581208/codecharta_2022-04-28_08-53.cc.zip)
WHEN
comparing them in the delta mode
THEN
I can see the actual deltas.
## Actual Behavior
the two maps are shown side by side and no deltas are shown
## Screenshots

## Specifications
- is released in [online-demo](https://maibornwolff.github.io/codecharta/visualization/app/index.html?file=codecharta.cc.json&file=codecharta_analysis.cc.json): yes
- CodeCharta Version: 1.95.2
- OS: all
- Browser: all
|
priority
|
delta mode not working for codecharta cc json bug expected behavior given current codecharta cc json old map unzip and upload when comparing them in the delta mode then i can see the actual deltas actual behavior the two maps are shown side by side and no deltas are shown screenshots specifications is released in yes codecharta version os all browser all
| 1
|
647,985
| 21,161,543,014
|
IssuesEvent
|
2022-04-07 09:48:15
|
AY2122S2-CS2103-F09-2/tp
|
https://api.github.com/repos/AY2122S2-CS2103-F09-2/tp
|
closed
|
ExpandedCard eventlist GUI change bug
|
priority.High type.Bug
|
when calling `listevents` from the expandedPersonCard and switching back to expandedPersonCard by clicking on the friends tab, all the events are displayed in the expandedPersonCard.
|
1.0
|
ExpandedCard eventlist GUI change bug - when calling `listevents` from the expandedPersonCard and switching back to expandedPersonCard by clicking on the friends tab, all the events are displayed in the expandedPersonCard.
|
priority
|
expandedcard eventlist gui change bug when calling listevents from the expandedpersoncard and switching back to expandedpersoncard by clicking on the friends tab all the events are displayed in the expandedpersoncard
| 1
|
487,291
| 14,021,736,661
|
IssuesEvent
|
2020-10-29 21:49:26
|
open-contracting/extension-explorer
|
https://api.github.com/repos/open-contracting/extension-explorer
|
closed
|
Restore daily automatic deployment
|
high priority
|
We used to do this with Travis cron jobs. There's a similar feature for GitHub Workflows:
https://help.github.com/en/actions/reference/workflow-syntax-for-github-actions#onschedule
|
1.0
|
Restore daily automatic deployment - We used to do this with Travis cron jobs. There's a similar feature for GitHub Workflows:
https://help.github.com/en/actions/reference/workflow-syntax-for-github-actions#onschedule
|
priority
|
restore daily automatic deployment we used to do this with travis cron jobs there s a similar feature for github workflows
| 1
|
55,255
| 3,072,552,364
|
IssuesEvent
|
2015-08-19 17:30:30
|
ubc/acj-versus
|
https://api.github.com/repos/ubc/acj-versus
|
closed
|
Add/Edit Course/Question Screens: Explain criteria in terms of the student UI
|
enhancement high priority instructor request
|
Instructors have no way of accessing the student UI to understand the context criteria are shown in and could use a more visual explanation.
Add a preview window for instructors that will show how the criteria will appear when students go to evaluate. This should appear on both the course and question forms (as criteria are used in both places).
|
1.0
|
Add/Edit Course/Question Screens: Explain criteria in terms of the student UI - Instructors have no way of accessing the student UI to understand the context criteria are shown in and could use a more visual explanation.
Add a preview window for instructors that will show how the criteria will appear when students go to evaluate. This should appear on both the course and question forms (as criteria are used in both places).
|
priority
|
add edit course question screens explain criteria in terms of the student ui instructors have no way of accessing the student ui to understand the context criteria are shown in and could use a more visual explanation add a preview window for instructors that will show how the criteria will appear when students go to evaluate this should appear on both the course and question forms as criteria are used in both places
| 1
|
354,233
| 10,564,258,760
|
IssuesEvent
|
2019-10-05 00:25:17
|
bounswe/bounswe2019group8
|
https://api.github.com/repos/bounswe/bounswe2019group8
|
closed
|
Review the backend project
|
Backend Effort: Medium Group work Planning Priority: High Status: Available
|
**Actions:**
1. Review the initialized version of backend project.
2. Comment on the necessity of a potential tech stack change.
**Deadline:** 7.10.2019 - 21.00
|
1.0
|
Review the backend project - **Actions:**
1. Review the initialized version of backend project.
2. Comment on the necessity of a potential tech stack change.
**Deadline:** 7.10.2019 - 21.00
|
priority
|
review the backend project actions review the initialized version of backend project comment on the necessity of a potential tech stack change deadline
| 1
|
408,636
| 11,950,175,849
|
IssuesEvent
|
2020-04-03 14:47:59
|
AY1920S2-CS2103T-W12-4/main
|
https://api.github.com/repos/AY1920S2-CS2103T-W12-4/main
|
closed
|
[PE-D] UG provided command does not work in app
|
priority.High severity.High status.Ongoing type.Bug
|

To replicate mistake, copy and paste the command given in the UG onto the app for new recipe
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: YuuTon9/ped#1
|
1.0
|
[PE-D] UG provided command does not work in app - 
To replicate mistake, copy and paste the command given in the UG onto the app for new recipe
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: YuuTon9/ped#1
|
priority
|
ug provided command does not work in app to replicate mistake copy and paste the command given in the ug onto the app for new recipe labels severity low type documentationbug original ped
| 1
|
766,354
| 26,880,479,019
|
IssuesEvent
|
2023-02-05 15:22:53
|
LikeLion-VJS10/TAING10
|
https://api.github.com/repos/LikeLion-VJS10/TAING10
|
closed
|
유효성, 성능최적화 중간기록
|
🦴 Html&Css High Priority
|
2/5(일) 오전 10시회의시, 최종 커밋 내역으로 유효성, 성능최적화 기록 필요.
[벨리데이션]
https://www.notion.so/fb8c5ddade0d4d218eb453aaefc5a844#e57f7692a9ac4a368d90b2450b0b1be8
[성능최적화]
https://www.notion.so/fb8c5ddade0d4d218eb453aaefc5a844#33528aca32b0402ea695d845b07b2886
화면캡쳐로 기록 남긴 후, (개선시)개선후의 기록과 비교해본다.
|
1.0
|
유효성, 성능최적화 중간기록 - 2/5(일) 오전 10시회의시, 최종 커밋 내역으로 유효성, 성능최적화 기록 필요.
[벨리데이션]
https://www.notion.so/fb8c5ddade0d4d218eb453aaefc5a844#e57f7692a9ac4a368d90b2450b0b1be8
[성능최적화]
https://www.notion.so/fb8c5ddade0d4d218eb453aaefc5a844#33528aca32b0402ea695d845b07b2886
화면캡쳐로 기록 남긴 후, (개선시)개선후의 기록과 비교해본다.
|
priority
|
유효성 성능최적화 중간기록 일 오전 최종 커밋 내역으로 유효성 성능최적화 기록 필요 화면캡쳐로 기록 남긴 후 개선시 개선후의 기록과 비교해본다
| 1
|
52,473
| 3,023,578,751
|
IssuesEvent
|
2015-08-01 17:02:37
|
Baystation12/Baystation12
|
https://api.github.com/repos/Baystation12/Baystation12
|
closed
|
[Dev-Freeze] AI can still free-roam its eye while removed from its chassis.
|
bug needs review oversight priority: high
|
**How to reproduce**
* Transfer as an AI from a chassis into an inteliCard.
* Use the movement keys.
* Note how your vision isn't locked to your current location.
|
1.0
|
[Dev-Freeze] AI can still free-roam its eye while removed from its chassis. - **How to reproduce**
* Transfer as an AI from a chassis into an inteliCard.
* Use the movement keys.
* Note how your vision isn't locked to your current location.
|
priority
|
ai can still free roam its eye while removed from its chassis how to reproduce transfer as an ai from a chassis into an intelicard use the movement keys note how your vision isn t locked to your current location
| 1
|
442,975
| 12,753,964,472
|
IssuesEvent
|
2020-06-28 02:04:25
|
projectacrn/acrn-hypervisor
|
https://api.github.com/repos/projectacrn/acrn-hypervisor
|
closed
|
linux-iot-lts2018-sos-4.19.78-95.x86_64
|
priority: P2-High type: bug
|
Login Prompt appeared successfully but keyboard not responding for input entry

|
1.0
|
linux-iot-lts2018-sos-4.19.78-95.x86_64 - Login Prompt appeared successfully but keyboard not responding for input entry

|
priority
|
linux iot sos login prompt appeared successfully but keyboard not responding for input entry
| 1
|
118,668
| 4,751,664,389
|
IssuesEvent
|
2016-10-23 01:09:01
|
Hetal728/MCGalaxy
|
https://api.github.com/repos/Hetal728/MCGalaxy
|
closed
|
Very small chance of levels being corrupted
|
bug high priority
|
This happens in LvlFile.Save due to the block data changing (by another thread such as physics)
between the call to updating the crc32, and the call to actually writing the blocks.
|
1.0
|
Very small chance of levels being corrupted - This happens in LvlFile.Save due to the block data changing (by another thread such as physics)
between the call to updating the crc32, and the call to actually writing the blocks.
|
priority
|
very small chance of levels being corrupted this happens in lvlfile save due to the block data changing by another thread such as physics between the call to updating the and the call to actually writing the blocks
| 1
|
428,073
| 12,402,444,031
|
IssuesEvent
|
2020-05-21 12:00:35
|
inverse-inc/packetfence
|
https://api.github.com/repos/inverse-inc/packetfence
|
closed
|
Impossible to set multiple access level in administration rule
|
Priority: High Type: Bug
|
**Describe the bug**
In the previous admin we used to be able to set multiple access level rules. The new admin needs to support this as well.
**To Reproduce**
1. Go in an LDAP/AD source
2. Create an administration rule
3. Attempt to set the "Access level"
4. Try to put multiple values in that field
**Expected behavior**
In a single "Access level" action, we should be able to set multiple access levels
|
1.0
|
Impossible to set multiple access level in administration rule - **Describe the bug**
In the previous admin we used to be able to set multiple access level rules. The new admin needs to support this as well.
**To Reproduce**
1. Go in an LDAP/AD source
2. Create an administration rule
3. Attempt to set the "Access level"
4. Try to put multiple values in that field
**Expected behavior**
In a single "Access level" action, we should be able to set multiple access levels
|
priority
|
impossible to set multiple access level in administration rule describe the bug in the previous admin we used to be able to set multiple access level rules the new admin needs to support this as well to reproduce go in an ldap ad source create an administration rule attempt to set the access level try to put multiple values in that field expected behavior in a single access level action we should be able to set multiple access levels
| 1
|
336,925
| 10,207,690,617
|
IssuesEvent
|
2019-08-14 08:20:39
|
zdnscloud/singlecloud
|
https://api.github.com/repos/zdnscloud/singlecloud
|
opened
|
missing "deleting storage node" function in front-end
|
bug master priority: High
|
1. add a ceph storage with two node storage,
2. try to remove one of them,
the following occurs:
no deleting function found in the UI.

|
1.0
|
missing "deleting storage node" function in front-end - 1. add a ceph storage with two node storage,
2. try to remove one of them,
the following occurs:
no deleting function found in the UI.

|
priority
|
missing deleting storage node function in front end add a ceph storage with two node storage try to remove one of them the following occurs no deleting function found in the ui
| 1
|
106,774
| 4,283,660,035
|
IssuesEvent
|
2016-07-15 14:14:10
|
RobotLocomotion/drake
|
https://api.github.com/repos/RobotLocomotion/drake
|
closed
|
Add a Mac MATLAB JDK 8 Nightly Build
|
priority: high team: kitware type: continuous integration
|
This is a common enough configuration that we should support it, even though the system and MATLAB JDK versions will be in conflict. See #2772.
|
1.0
|
Add a Mac MATLAB JDK 8 Nightly Build - This is a common enough configuration that we should support it, even though the system and MATLAB JDK versions will be in conflict. See #2772.
|
priority
|
add a mac matlab jdk nightly build this is a common enough configuration that we should support it even though the system and matlab jdk versions will be in conflict see
| 1
|
678,062
| 23,185,817,013
|
IssuesEvent
|
2022-08-01 08:17:18
|
geosolutions-it/MapStore2
|
https://api.github.com/repos/geosolutions-it/MapStore2
|
closed
|
Map widget sync issue
|
bug Priority: High Accepted C040-COMUNE_GE-2022-SUPPORT
|
## Description
<!-- Add here a few sentences describing the bug. -->
There is a regression that currently affects the master branch (DEV instance [here](https://dev-mapstore.geosolutionsgroup.com/mapstore/#/)). if you connect a map widget to another map widget, the viewport is sync hronized only when zooming and not when panning the guide map
## How to reproduce
<!-- A list of steps to reproduce the bug -->
- Create a new dashboar
- Create a map widget
- Create a second map widget a d connect it to the first one
*Expected Result*
<!-- Describe here the expected result -->
As it is currently working in QA (2022.01.xx)

*Current Result*
<!-- Describe here the current behavior -->
How it is working in DEV

- [x] Not browser related
<details><summary> <b>Browser info</b> </summary>
<!-- If browser related, please compile the following table -->
<!-- If your browser is not in the list please add a new row to the table with the version -->
(use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users)
| Browser Affected | Version |
|---|---|
|Internet Explorer| |
|Edge| |
|Chrome| |
|Firefox| |
|Safari| |
</details>
## Other useful information
<!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
|
1.0
|
Map widget sync issue - ## Description
<!-- Add here a few sentences describing the bug. -->
There is a regression that currently affects the master branch (DEV instance [here](https://dev-mapstore.geosolutionsgroup.com/mapstore/#/)). if you connect a map widget to another map widget, the viewport is sync hronized only when zooming and not when panning the guide map
## How to reproduce
<!-- A list of steps to reproduce the bug -->
- Create a new dashboar
- Create a map widget
- Create a second map widget a d connect it to the first one
*Expected Result*
<!-- Describe here the expected result -->
As it is currently working in QA (2022.01.xx)

*Current Result*
<!-- Describe here the current behavior -->
How it is working in DEV

- [x] Not browser related
<details><summary> <b>Browser info</b> </summary>
<!-- If browser related, please compile the following table -->
<!-- If your browser is not in the list please add a new row to the table with the version -->
(use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users)
| Browser Affected | Version |
|---|---|
|Internet Explorer| |
|Edge| |
|Chrome| |
|Firefox| |
|Safari| |
</details>
## Other useful information
<!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
|
priority
|
map widget sync issue description there is a regression that currently affects the master branch dev instance if you connect a map widget to another map widget the viewport is sync hronized only when zooming and not when panning the guide map how to reproduce create a new dashboar create a map widget create a second map widget a d connect it to the first one expected result as it is currently working in qa xx current result how it is working in dev not browser related browser info use this site a href for non expert users browser affected version internet explorer edge chrome firefox safari other useful information
| 1
|
468,437
| 13,482,874,759
|
IssuesEvent
|
2020-09-11 02:34:07
|
Bruce773/lyric-keeper
|
https://api.github.com/repos/Bruce773/lyric-keeper
|
closed
|
Add contact form to FAQ and Contact Us page
|
New Feature TIMEFRAME: High-priority
|
- Create contact form component using Formik, Yup, and Formspree
- Add contact form component to a section on the FAQ page
- Create a `Contact Us` page which is accessible from the nav or `/contact-us` –– page should include new contact form component in the body
|
1.0
|
Add contact form to FAQ and Contact Us page - - Create contact form component using Formik, Yup, and Formspree
- Add contact form component to a section on the FAQ page
- Create a `Contact Us` page which is accessible from the nav or `/contact-us` –– page should include new contact form component in the body
|
priority
|
add contact form to faq and contact us page create contact form component using formik yup and formspree add contact form component to a section on the faq page create a contact us page which is accessible from the nav or contact us –– page should include new contact form component in the body
| 1
|
4,455
| 2,552,225,294
|
IssuesEvent
|
2015-02-02 15:55:05
|
YetiForceCompany/YetiForceCRM
|
https://api.github.com/repos/YetiForceCompany/YetiForceCRM
|
closed
|
bug in activities
|
high priority bug
|
When you try to add activity from contact summary view it does not add the contact name to the event and when you try to add it in the full form it gives you error: Record you are trying to access is not found. 104 accounts.
Tested it on test.yetiform.com
|
1.0
|
bug in activities - When you try to add activity from contact summary view it does not add the contact name to the event and when you try to add it in the full form it gives you error: Record you are trying to access is not found. 104 accounts.
Tested it on test.yetiform.com
|
priority
|
bug in activities when you try to add activity from contact summary view it does not add the contact name to the event and when you try to add it in the full form it gives you error record you are trying to access is not found accounts tested it on test yetiform com
| 1
|
492,369
| 14,201,161,435
|
IssuesEvent
|
2020-11-16 07:10:27
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
youtrack.jetbrains.com - desktop site instead of mobile site
|
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal
|
<!-- @browser: Firefox Mobile 83.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:83.0) Gecko/83.0 Firefox/83.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/61862 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://youtrack.jetbrains.com/issue/IDEA-212147?_ga=2.55445602.874277953.1605467280-822587761.1605467280
**Browser / Version**: Firefox Mobile 83.0
**Operating System**: Android
**Tested Another Browser**: Yes Internet Explorer
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/11/12f848ca-0895-4e27-9130-f137ab9ce5f3.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201108174701</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/11/7179cb6f-5653-4c9e-af8f-e08e4791ef3b)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
youtrack.jetbrains.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 83.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:83.0) Gecko/83.0 Firefox/83.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/61862 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://youtrack.jetbrains.com/issue/IDEA-212147?_ga=2.55445602.874277953.1605467280-822587761.1605467280
**Browser / Version**: Firefox Mobile 83.0
**Operating System**: Android
**Tested Another Browser**: Yes Internet Explorer
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/11/12f848ca-0895-4e27-9130-f137ab9ce5f3.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201108174701</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/11/7179cb6f-5653-4c9e-af8f-e08e4791ef3b)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
priority
|
youtrack jetbrains com desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes internet explorer problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 1
|
132,735
| 5,192,050,372
|
IssuesEvent
|
2017-01-22 03:27:20
|
bethlakshmi/GBE2
|
https://api.github.com/repos/bethlakshmi/GBE2
|
closed
|
Act Scheduling - only include shows from the 'active' conference
|
8 point Estimated High Priority Merged
|
The Act Scheduling page:
http://localhost:8282/scheduler/acts/<show choice>
Is both slow and ugly. I thought we had this ticket already - but I just can't find it.
Both the slowness and the visual breakage (it flies off the end of the screen layout) can be fixed by making the list of shows in the drop down be only the few shows that are applicable for the given conference. Right now, the Show drop down shows EVERY scheduler.event in the system. That's every scheduled time for every thing. Ever.
Some of those things have super-long names, so the drop down is very wide.
And marshalling all that from the database is way slow. And it slows on a per act basis. The more acts you have in the show, the slower it is.
-------------------
Recommended fix:
In scheduler/views.py in the function schedule_acts change either the view or the ActScheduleForm to take *only* the shows that are in the same conference as *this* show. That means that we constrict the drop down to the reasonable list of what these acts should actually be cast into.
and we speed up the page...
and it looks nicer...
Assigning to @burlexpo so he can set priority. Burlexpo - when you do that, you can switch it to unassigned, if you want it before the expo, please mark high and "estimation".
|
1.0
|
Act Scheduling - only include shows from the 'active' conference - The Act Scheduling page:
http://localhost:8282/scheduler/acts/<show choice>
Is both slow and ugly. I thought we had this ticket already - but I just can't find it.
Both the slowness and the visual breakage (it flies off the end of the screen layout) can be fixed by making the list of shows in the drop down be only the few shows that are applicable for the given conference. Right now, the Show drop down shows EVERY scheduler.event in the system. That's every scheduled time for every thing. Ever.
Some of those things have super-long names, so the drop down is very wide.
And marshalling all that from the database is way slow. And it slows on a per act basis. The more acts you have in the show, the slower it is.
-------------------
Recommended fix:
In scheduler/views.py in the function schedule_acts change either the view or the ActScheduleForm to take *only* the shows that are in the same conference as *this* show. That means that we constrict the drop down to the reasonable list of what these acts should actually be cast into.
and we speed up the page...
and it looks nicer...
Assigning to @burlexpo so he can set priority. Burlexpo - when you do that, you can switch it to unassigned, if you want it before the expo, please mark high and "estimation".
|
priority
|
act scheduling only include shows from the active conference the act scheduling page choice is both slow and ugly i thought we had this ticket already but i just can t find it both the slowness and the visual breakage it flies off the end of the screen layout can be fixed by making the list of shows in the drop down be only the few shows that are applicable for the given conference right now the show drop down shows every scheduler event in the system that s every scheduled time for every thing ever some of those things have super long names so the drop down is very wide and marshalling all that from the database is way slow and it slows on a per act basis the more acts you have in the show the slower it is recommended fix in scheduler views py in the function schedule acts change either the view or the actscheduleform to take only the shows that are in the same conference as this show that means that we constrict the drop down to the reasonable list of what these acts should actually be cast into and we speed up the page and it looks nicer assigning to burlexpo so he can set priority burlexpo when you do that you can switch it to unassigned if you want it before the expo please mark high and estimation
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.