Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
87,075
3,736,717,427
IssuesEvent
2016-03-08 16:48:35
Fermat-ORG/fermat-org
https://api.github.com/repos/Fermat-ORG/fermat-org
closed
Update developer list
Priority: HIGH question server
Caso de estudio: Miguelcldn es un nuevo developer que quiere entrar en Fermat, nunca tuvo una ficha asignada. - Entra en la página - Inicia sesión - Crea un componente - Pone "miguelcldn" en el nombre del developer - Se crea la ficha y se pone en su posición, pero... ¿aparecerá Miguelcldn en la lista de developers al presionar F5?
1.0
Update developer list - Caso de estudio: Miguelcldn es un nuevo developer que quiere entrar en Fermat, nunca tuvo una ficha asignada. - Entra en la página - Inicia sesión - Crea un componente - Pone "miguelcldn" en el nombre del developer - Se crea la ficha y se pone en su posición, pero... ¿aparecerá Miguelcldn en la lista de developers al presionar F5?
priority
update developer list caso de estudio miguelcldn es un nuevo developer que quiere entrar en fermat nunca tuvo una ficha asignada entra en la página inicia sesión crea un componente pone miguelcldn en el nombre del developer se crea la ficha y se pone en su posición pero ¿aparecerá miguelcldn en la lista de developers al presionar
1
403,934
11,849,278,852
IssuesEvent
2020-03-24 15:01:02
Monika-After-Story/MonikaModDev
https://api.github.com/repos/Monika-After-Story/MonikaModDev
closed
In-game explanation of gifting mechanic
enhancement high priority
A quick search of code shows no instance of "characters folder" or "characters/" in any dialogue relating to gifting, which implies there is no in-game explanation of the gifting mechanic. This is probably why new players don't understand or miss out on things as it was initially introduced in a 9-22 event as an external file. We should have some in-game, repeatable explanation of this mechanic (story-events, with a say prompt like "How do I gift you things?"). This is especially important given consumables. This should be also noted in intro, like how we explain hotkeys and games. Additionally, I'd say an update script to queue the topic is unnecessary, a new prompt will draw attention in unseen anyway. @multimokia I am assigning this to you but you can triage as you see fit. Just get this in before next release.
1.0
In-game explanation of gifting mechanic - A quick search of code shows no instance of "characters folder" or "characters/" in any dialogue relating to gifting, which implies there is no in-game explanation of the gifting mechanic. This is probably why new players don't understand or miss out on things as it was initially introduced in a 9-22 event as an external file. We should have some in-game, repeatable explanation of this mechanic (story-events, with a say prompt like "How do I gift you things?"). This is especially important given consumables. This should be also noted in intro, like how we explain hotkeys and games. Additionally, I'd say an update script to queue the topic is unnecessary, a new prompt will draw attention in unseen anyway. @multimokia I am assigning this to you but you can triage as you see fit. Just get this in before next release.
priority
in game explanation of gifting mechanic a quick search of code shows no instance of characters folder or characters in any dialogue relating to gifting which implies there is no in game explanation of the gifting mechanic this is probably why new players don t understand or miss out on things as it was initially introduced in a event as an external file we should have some in game repeatable explanation of this mechanic story events with a say prompt like how do i gift you things this is especially important given consumables this should be also noted in intro like how we explain hotkeys and games additionally i d say an update script to queue the topic is unnecessary a new prompt will draw attention in unseen anyway multimokia i am assigning this to you but you can triage as you see fit just get this in before next release
1
321,773
9,808,860,529
IssuesEvent
2019-06-12 16:31:52
wso2/product-is
https://api.github.com/repos/wso2/product-is
opened
IS : Custom claims lost when updating user profile
Complexity/High Component/OAuth Priority/High
Moved from: https://wso2.org/jira/browse/IDENTITY-7335 Hi WSO2 team, WSO2 Identity Server offers some extension points for instance ability to add custom claims. This is normally donc by writing a custom claim handler as described in one wso2 post : http://pushpalankajaya.blogspot.com/2014/07/adding-custom-claims-to-saml-response.html When using WSO2 OAuth2.0 Playground sample and OAuth2.0 Authorization code grant type, custom claim handler is invoked normally when resource owner is authenticated. Then at the end of process you get an access token enabling to retrieve, for instance, user info with customs claim ("computed before when custom claim handler is invoked). If you update some information of user profile and request again user info from user info endpoint with access_token, all the +custom + claims are lost in provided response. Actually, when you update user profile, information that were previously cached are emptied so when user info endpoint is called with access_token, user claims (and their corresponding values) are computed from userstore perspective only so that why custom claims are lost. Would it be possible in that scenario of retrieving claims from userstore that custom claim handler would be also called to ensure that customs claims are available in answer? Regards, Franck
1.0
IS : Custom claims lost when updating user profile - Moved from: https://wso2.org/jira/browse/IDENTITY-7335 Hi WSO2 team, WSO2 Identity Server offers some extension points for instance ability to add custom claims. This is normally donc by writing a custom claim handler as described in one wso2 post : http://pushpalankajaya.blogspot.com/2014/07/adding-custom-claims-to-saml-response.html When using WSO2 OAuth2.0 Playground sample and OAuth2.0 Authorization code grant type, custom claim handler is invoked normally when resource owner is authenticated. Then at the end of process you get an access token enabling to retrieve, for instance, user info with customs claim ("computed before when custom claim handler is invoked). If you update some information of user profile and request again user info from user info endpoint with access_token, all the +custom + claims are lost in provided response. Actually, when you update user profile, information that were previously cached are emptied so when user info endpoint is called with access_token, user claims (and their corresponding values) are computed from userstore perspective only so that why custom claims are lost. Would it be possible in that scenario of retrieving claims from userstore that custom claim handler would be also called to ensure that customs claims are available in answer? Regards, Franck
priority
is custom claims lost when updating user profile moved from hi team identity server offers some extension points for instance ability to add custom claims this is normally donc by writing a custom claim handler as described in one post when using playground sample and authorization code grant type custom claim handler is invoked normally when resource owner is authenticated then at the end of process you get an access token enabling to retrieve for instance user info with customs claim computed before when custom claim handler is invoked if you update some information of user profile and request again user info from user info endpoint with access token all the custom claims are lost in provided response actually when you update user profile information that were previously cached are emptied so when user info endpoint is called with access token user claims and their corresponding values are computed from userstore perspective only so that why custom claims are lost would it be possible in that scenario of retrieving claims from userstore that custom claim handler would be also called to ensure that customs claims are available in answer regards franck
1
567,375
16,857,149,375
IssuesEvent
2021-06-21 08:19:34
FEDMix/eshmun
https://api.github.com/repos/FEDMix/eshmun
closed
Orthogonal view to view DICOM Images
Brachytherapy FEDmix High Priority Modir User Story
## Story As a Clinician I want to view DICOM images of a selected patient So that I can see a preview of the choice I am trying to make ## Proposed work Within the application it should be possible to view DICOM images. There should be 3 views that are orthogonal to each other (x, y and z directions). The user should be able to pan and zoom the views. ## Acceptance Criteria - [x] 3 Orthogonal Views - [x] Panning & Zooming ## Designs ![screen-design](https://user-images.githubusercontent.com/16776108/112948196-0dc2d300-9138-11eb-8724-be0381313b92.png)
1.0
Orthogonal view to view DICOM Images - ## Story As a Clinician I want to view DICOM images of a selected patient So that I can see a preview of the choice I am trying to make ## Proposed work Within the application it should be possible to view DICOM images. There should be 3 views that are orthogonal to each other (x, y and z directions). The user should be able to pan and zoom the views. ## Acceptance Criteria - [x] 3 Orthogonal Views - [x] Panning & Zooming ## Designs ![screen-design](https://user-images.githubusercontent.com/16776108/112948196-0dc2d300-9138-11eb-8724-be0381313b92.png)
priority
orthogonal view to view dicom images story as a clinician i want to view dicom images of a selected patient so that i can see a preview of the choice i am trying to make proposed work within the application it should be possible to view dicom images there should be views that are orthogonal to each other x y and z directions the user should be able to pan and zoom the views acceptance criteria orthogonal views panning zooming designs
1
692,450
23,735,308,479
IssuesEvent
2022-08-31 07:33:57
vignetteapp/sekai
https://api.github.com/repos/vignetteapp/sekai
closed
Implement a basic rendering system
enhancement priority:high
As our rendering engine is primarily for 3D, we'd need a basic rendering system that shows 3D models with lighting and shadows.
1.0
Implement a basic rendering system - As our rendering engine is primarily for 3D, we'd need a basic rendering system that shows 3D models with lighting and shadows.
priority
implement a basic rendering system as our rendering engine is primarily for we d need a basic rendering system that shows models with lighting and shadows
1
404,490
11,858,245,224
IssuesEvent
2020-03-25 11:04:21
AY1920S2-CS2103T-F10-2/main
https://api.github.com/repos/AY1920S2-CS2103T-F10-2/main
closed
Implement Reminder feature
priority.High type.Epic
Update the default list view of the Internship Diary to be shown according to urgency of the application, in terms of the application deadline followed by interview(s) date(s).
1.0
Implement Reminder feature - Update the default list view of the Internship Diary to be shown according to urgency of the application, in terms of the application deadline followed by interview(s) date(s).
priority
implement reminder feature update the default list view of the internship diary to be shown according to urgency of the application in terms of the application deadline followed by interview s date s
1
229,631
7,582,167,904
IssuesEvent
2018-04-25 02:29:17
Myoats/preprod
https://api.github.com/repos/Myoats/preprod
opened
The Nav on small on medium devices on the should just be the hamburger
Highest Priority
It should look like this ![image](https://user-images.githubusercontent.com/37152320/39223236-d30a7102-480e-11e8-99de-88a08665f547.png)
1.0
The Nav on small on medium devices on the should just be the hamburger - It should look like this ![image](https://user-images.githubusercontent.com/37152320/39223236-d30a7102-480e-11e8-99de-88a08665f547.png)
priority
the nav on small on medium devices on the should just be the hamburger it should look like this
1
359,071
10,659,685,809
IssuesEvent
2019-10-18 08:17:45
AY1920S1-CS2113T-W13-2/main
https://api.github.com/repos/AY1920S1-CS2113T-W13-2/main
opened
Create Review function for Quiz
priority.High type.Task
Allows user to review their answers and the correct answer at the end of a quiz session (CLI).
1.0
Create Review function for Quiz - Allows user to review their answers and the correct answer at the end of a quiz session (CLI).
priority
create review function for quiz allows user to review their answers and the correct answer at the end of a quiz session cli
1
86,978
3,735,649,127
IssuesEvent
2016-03-08 13:08:54
asterics/AsTeRICS
https://api.github.com/repos/asterics/AsTeRICS
closed
JNativeHook error message when starting two instances of AsTeRICS
high priority
When starting two instances of AsTeRICS the jnativehook service bundle cannot be activated and an error message is thrown. The debug console shows the following root exception: Caused by: java.lang.RuntimeException: C:\Users\mad\AppData\Local\Temp\JNativeHook-1.2.RC2.dll (Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird) at org.jnativehook.GlobalScreen.loadNativeLibrary(GlobalScreen.java:609) at org.jnativehook.GlobalScreen.<init>(GlobalScreen.java:86) at org.jnativehook.GlobalScreen.<clinit>(GlobalScreen.java:67) ... 28 more Caused by: java.io.FileNotFoundException: C:\Users\mad\AppData\Local\Temp\JNativeHook-1.2.RC2.dll (Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird) at java.io.FileOutputStream.open(Native Method) at java.io.FileOutputStream.<init>(FileOutputStream.java:221) at java.io.FileOutputStream.<init>(FileOutputStream.java:171) at org.jnativehook.GlobalScreen.loadNativeLibrary(GlobalScreen.java:570) ... 30 more Obviously the native dll file is busy and cannot be accessed two times. Possible solutions: upgrade to newer version of jnativehook, maybe it's already fixed? workaround: Manually delete .dll file after startup of bundle.
1.0
JNativeHook error message when starting two instances of AsTeRICS - When starting two instances of AsTeRICS the jnativehook service bundle cannot be activated and an error message is thrown. The debug console shows the following root exception: Caused by: java.lang.RuntimeException: C:\Users\mad\AppData\Local\Temp\JNativeHook-1.2.RC2.dll (Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird) at org.jnativehook.GlobalScreen.loadNativeLibrary(GlobalScreen.java:609) at org.jnativehook.GlobalScreen.<init>(GlobalScreen.java:86) at org.jnativehook.GlobalScreen.<clinit>(GlobalScreen.java:67) ... 28 more Caused by: java.io.FileNotFoundException: C:\Users\mad\AppData\Local\Temp\JNativeHook-1.2.RC2.dll (Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird) at java.io.FileOutputStream.open(Native Method) at java.io.FileOutputStream.<init>(FileOutputStream.java:221) at java.io.FileOutputStream.<init>(FileOutputStream.java:171) at org.jnativehook.GlobalScreen.loadNativeLibrary(GlobalScreen.java:570) ... 30 more Obviously the native dll file is busy and cannot be accessed two times. Possible solutions: upgrade to newer version of jnativehook, maybe it's already fixed? workaround: Manually delete .dll file after startup of bundle.
priority
jnativehook error message when starting two instances of asterics when starting two instances of asterics the jnativehook service bundle cannot be activated and an error message is thrown the debug console shows the following root exception caused by java lang runtimeexception c users mad appdata local temp jnativehook dll der prozess kann nicht auf die datei zugreifen da sie von einem anderen prozess verwendet wird at org jnativehook globalscreen loadnativelibrary globalscreen java at org jnativehook globalscreen globalscreen java at org jnativehook globalscreen globalscreen java more caused by java io filenotfoundexception c users mad appdata local temp jnativehook dll der prozess kann nicht auf die datei zugreifen da sie von einem anderen prozess verwendet wird at java io fileoutputstream open native method at java io fileoutputstream fileoutputstream java at java io fileoutputstream fileoutputstream java at org jnativehook globalscreen loadnativelibrary globalscreen java more obviously the native dll file is busy and cannot be accessed two times possible solutions upgrade to newer version of jnativehook maybe it s already fixed workaround manually delete dll file after startup of bundle
1
725,264
24,956,556,151
IssuesEvent
2022-11-01 12:18:11
AY2223S1-CS2113-T18-1b/tp
https://api.github.com/repos/AY2223S1-CS2113-T18-1b/tp
closed
[PE-D][Tester C] sort should come configurable ordering to make sense - i.e. asc or desc. Otherwise, fix current defaults as some orderings do not make sense
priority.High
For example, the current sort for review is ascending, which is unusual as most people sort shows by descending rating <!--session: 1666946737319-de45f0f8-5014-4481-873f-b0f1008853e3--><!--Version: Web v3.4.4--> ------------- Labels: `type.FeatureFlaw` `severity.Low` original: winston-lim/ped#7
1.0
[PE-D][Tester C] sort should come configurable ordering to make sense - i.e. asc or desc. Otherwise, fix current defaults as some orderings do not make sense - For example, the current sort for review is ascending, which is unusual as most people sort shows by descending rating <!--session: 1666946737319-de45f0f8-5014-4481-873f-b0f1008853e3--><!--Version: Web v3.4.4--> ------------- Labels: `type.FeatureFlaw` `severity.Low` original: winston-lim/ped#7
priority
sort should come configurable ordering to make sense i e asc or desc otherwise fix current defaults as some orderings do not make sense for example the current sort for review is ascending which is unusual as most people sort shows by descending rating labels type featureflaw severity low original winston lim ped
1
351,271
10,514,749,872
IssuesEvent
2019-09-28 03:17:59
xournalpp/xournalpp
https://api.github.com/repos/xournalpp/xournalpp
closed
Crash on opening files.
Crash PR available bug confirmed difficulty:easy priority: high
**Affects versions :** - OS: [Linux Mint 19.2 Cinnamon] - Version of Xournal++ [current master] **Describe the bug** Pressing the open button crashes the software **To Reproduce** compile with clang (in DEBUG Mode), open xournal++ try to open a file **Expected behavior** Do not crash. fixed it already in a local branch, all `std::strings` in `Settings` are returned by value thus they only live after the line if not assigned. There was a .c_str() call on a temporary which is destructed immediately. Replaced all returns of a `std::string` by value in `Settings` with a `string const&`. PR following.
1.0
Crash on opening files. - **Affects versions :** - OS: [Linux Mint 19.2 Cinnamon] - Version of Xournal++ [current master] **Describe the bug** Pressing the open button crashes the software **To Reproduce** compile with clang (in DEBUG Mode), open xournal++ try to open a file **Expected behavior** Do not crash. fixed it already in a local branch, all `std::strings` in `Settings` are returned by value thus they only live after the line if not assigned. There was a .c_str() call on a temporary which is destructed immediately. Replaced all returns of a `std::string` by value in `Settings` with a `string const&`. PR following.
priority
crash on opening files affects versions os version of xournal describe the bug pressing the open button crashes the software to reproduce compile with clang in debug mode open xournal try to open a file expected behavior do not crash fixed it already in a local branch all std strings in settings are returned by value thus they only live after the line if not assigned there was a c str call on a temporary which is destructed immediately replaced all returns of a std string by value in settings with a string const pr following
1
204,129
7,084,095,754
IssuesEvent
2018-01-11 04:32:06
EFForg/privacybadger
https://api.github.com/repos/EFForg/privacybadger
reopened
Error reports for www.youtube.com
broken site high priority
We are getting lots of error reports for `www.youtube.com`; it's our most-reported-on domain for the latest Privacy Badger (version 2017.7.24). Here is some of what our users say: >video won't load >Youtube video shows: 'Request blocked by extension'. Fixed by switching on s0.2mdn.net >video said an extension blocked requests to the server >Videos laden anfangs nicht - Meldung "Anfragen an den Server wurden durch Erweiterung blockiert" >Ads don't appear instead it says an extension is blocking the ad content. Ad blocks still appear just without the actual ad. >preroll Advertisement on Youtube "blocked by an extension".
1.0
Error reports for www.youtube.com - We are getting lots of error reports for `www.youtube.com`; it's our most-reported-on domain for the latest Privacy Badger (version 2017.7.24). Here is some of what our users say: >video won't load >Youtube video shows: 'Request blocked by extension'. Fixed by switching on s0.2mdn.net >video said an extension blocked requests to the server >Videos laden anfangs nicht - Meldung "Anfragen an den Server wurden durch Erweiterung blockiert" >Ads don't appear instead it says an extension is blocking the ad content. Ad blocks still appear just without the actual ad. >preroll Advertisement on Youtube "blocked by an extension".
priority
error reports for we are getting lots of error reports for it s our most reported on domain for the latest privacy badger version here is some of what our users say video won t load youtube video shows request blocked by extension fixed by switching on net video said an extension blocked requests to the server videos laden anfangs nicht meldung anfragen an den server wurden durch erweiterung blockiert ads don t appear instead it says an extension is blocking the ad content ad blocks still appear just without the actual ad preroll advertisement on youtube blocked by an extension
1
810,681
30,254,112,094
IssuesEvent
2023-07-07 00:00:59
hwgilbert16/scholarsome
https://api.github.com/repos/hwgilbert16/scholarsome
closed
Exclude HTML tags from written response answers
high priority
Written response quiz questions check for exact strings. Cards created using the new rich text editor include `<p>` tags, which are checked for. These need to be removed before creating a quiz, or we need to filter the set beforehand to exclude these from being quizzed on.
1.0
Exclude HTML tags from written response answers - Written response quiz questions check for exact strings. Cards created using the new rich text editor include `<p>` tags, which are checked for. These need to be removed before creating a quiz, or we need to filter the set beforehand to exclude these from being quizzed on.
priority
exclude html tags from written response answers written response quiz questions check for exact strings cards created using the new rich text editor include tags which are checked for these need to be removed before creating a quiz or we need to filter the set beforehand to exclude these from being quizzed on
1
727,942
25,060,630,893
IssuesEvent
2022-11-07 00:56:45
AY2223S1-CS2103T-W15-4/tp
https://api.github.com/repos/AY2223S1-CS2103T-W15-4/tp
closed
[PE-D][Tester A] Mark INDEX marked 2 students
type.Bug priority.High severity.High
1. add n/DavId Li 2. add n/Davlid Lin 3. sort t/d 4. mark 1 will pass mastery check for both 5. unmark similarly removes pass for both <i><video controls><source src="https://raw.githubusercontent.com/KSHan29/ped/main/files/024d6dca-4e30-48ff-a3c0-15dbe2a8bf03.mov" type="video/mp4">Your browser does not support the video tag.</video><br>video:https://raw.githubusercontent.com/KSHan29/ped/main/files/024d6dca-4e30-48ff-a3c0-15dbe2a8bf03.mov</i> <!--session: 1666944123338-1a1c6713-a846-47ba-a447-d3dd287bc156--> <!--Version: Web v3.4.4--> ------------- Labels: `severity.Medium` `type.FunctionalityBug` original: KSHan29/ped#1
1.0
[PE-D][Tester A] Mark INDEX marked 2 students - 1. add n/DavId Li 2. add n/Davlid Lin 3. sort t/d 4. mark 1 will pass mastery check for both 5. unmark similarly removes pass for both <i><video controls><source src="https://raw.githubusercontent.com/KSHan29/ped/main/files/024d6dca-4e30-48ff-a3c0-15dbe2a8bf03.mov" type="video/mp4">Your browser does not support the video tag.</video><br>video:https://raw.githubusercontent.com/KSHan29/ped/main/files/024d6dca-4e30-48ff-a3c0-15dbe2a8bf03.mov</i> <!--session: 1666944123338-1a1c6713-a846-47ba-a447-d3dd287bc156--> <!--Version: Web v3.4.4--> ------------- Labels: `severity.Medium` `type.FunctionalityBug` original: KSHan29/ped#1
priority
mark index marked students add n david li add n davlid lin sort t d mark will pass mastery check for both unmark similarly removes pass for both your browser does not support the video tag video labels severity medium type functionalitybug original ped
1
402,408
11,809,307,977
IssuesEvent
2020-03-19 14:46:01
wso2/docker-is
https://api.github.com/repos/wso2/docker-is
closed
Use WSO2 product pack downloadable links to binaries available at GitHub release pages
Priority/Highest Type/Task
**Description:** Currently, WSO2 product binaries available at JFrog Bintray are used for the Docker image builds. It has been suggested to use WSO2 product pack downloadable links to binaries available at GitHub product release pages, as the default. **Affected Product Version:** Docker resources for WSO2 IAM v5.10.0 and beyond **Sub Tasks:** - [x] Integrate to Alpine based Docker resources - [x] Integrate to CentOS based Docker resources - [x] Integrate to Ubuntu based Docker resources - [x] Code/Peer review
1.0
Use WSO2 product pack downloadable links to binaries available at GitHub release pages - **Description:** Currently, WSO2 product binaries available at JFrog Bintray are used for the Docker image builds. It has been suggested to use WSO2 product pack downloadable links to binaries available at GitHub product release pages, as the default. **Affected Product Version:** Docker resources for WSO2 IAM v5.10.0 and beyond **Sub Tasks:** - [x] Integrate to Alpine based Docker resources - [x] Integrate to CentOS based Docker resources - [x] Integrate to Ubuntu based Docker resources - [x] Code/Peer review
priority
use product pack downloadable links to binaries available at github release pages description currently product binaries available at jfrog bintray are used for the docker image builds it has been suggested to use product pack downloadable links to binaries available at github product release pages as the default affected product version docker resources for iam and beyond sub tasks integrate to alpine based docker resources integrate to centos based docker resources integrate to ubuntu based docker resources code peer review
1
205,670
7,104,589,060
IssuesEvent
2018-01-16 10:29:33
AnSyn/ansyn
https://api.github.com/repos/AnSyn/ansyn
closed
Bug- shadow mouse- cursor on inactive screen
Bug Priority: High Severity: Medium
**Current behavior** When hovering an inactive screen with the cursor- the user can see both the shadow mouse cross and the cursor in the same screen. **Expected behavior** the shadow mouse cross should not appear, so the user wont be confused. **Minimal reproduction of the problem with instructions** open more then one screen activate shadow mouse hover the inactive screen
1.0
Bug- shadow mouse- cursor on inactive screen - **Current behavior** When hovering an inactive screen with the cursor- the user can see both the shadow mouse cross and the cursor in the same screen. **Expected behavior** the shadow mouse cross should not appear, so the user wont be confused. **Minimal reproduction of the problem with instructions** open more then one screen activate shadow mouse hover the inactive screen
priority
bug shadow mouse cursor on inactive screen current behavior when hovering an inactive screen with the cursor the user can see both the shadow mouse cross and the cursor in the same screen expected behavior the shadow mouse cross should not appear so the user wont be confused minimal reproduction of the problem with instructions open more then one screen activate shadow mouse hover the inactive screen
1
30,980
2,730,528,396
IssuesEvent
2015-04-16 15:20:01
Esri/briefing-book
https://api.github.com/repos/Esri/briefing-book
closed
Comment the codebase
bug develop21drop High Priority
Aside from the config.js, the majority of the code base isn't commented. Please provide quality commenting within the code base.
1.0
Comment the codebase - Aside from the config.js, the majority of the code base isn't commented. Please provide quality commenting within the code base.
priority
comment the codebase aside from the config js the majority of the code base isn t commented please provide quality commenting within the code base
1
481,243
13,882,518,034
IssuesEvent
2020-10-18 07:23:11
gileli121/WindowTop
https://api.github.com/repos/gileli121/WindowTop
closed
Modern toolbar UI is not working after saving window configuration on window with Shrink mode
bug fixed high-priority
A user reported that the new toolbar UI is not showing even when enabling the modern toolbar feature. I asked the user for the WindowTop.settings file, and he sent it. I found that the configuration that saved in this file was reproduced the issue. While debugging the issue, I found that the configuration key that reproduces it is: ``` [WindowsSettings] =ArrowXPos=0.45 ``` It should not be like that. it should be: ``` [WindowsSettings] some_process.exe=ArrowXPos=0.45 ``` I reviewed the code, and according to the code review, this configuration issue is caused in the following scenario: 1. Disable the modern toolbar 2. Shrink some window 3. Resize the shrink box so the arrow button will show when you move to the top area... 4. Drag the arrow button to some another position 5. Unshrink the window or exit WindowTop At this point, this configuration issue will reproduce. Because this configuration issue affects only the modern toolbar UI, you will need to enable the modern toolbar UI feature to reproduce it. Seems that the user did this scenario in older version or he did it before enabaling the modern toolbar feature.
1.0
Modern toolbar UI is not working after saving window configuration on window with Shrink mode - A user reported that the new toolbar UI is not showing even when enabling the modern toolbar feature. I asked the user for the WindowTop.settings file, and he sent it. I found that the configuration that saved in this file was reproduced the issue. While debugging the issue, I found that the configuration key that reproduces it is: ``` [WindowsSettings] =ArrowXPos=0.45 ``` It should not be like that. it should be: ``` [WindowsSettings] some_process.exe=ArrowXPos=0.45 ``` I reviewed the code, and according to the code review, this configuration issue is caused in the following scenario: 1. Disable the modern toolbar 2. Shrink some window 3. Resize the shrink box so the arrow button will show when you move to the top area... 4. Drag the arrow button to some another position 5. Unshrink the window or exit WindowTop At this point, this configuration issue will reproduce. Because this configuration issue affects only the modern toolbar UI, you will need to enable the modern toolbar UI feature to reproduce it. Seems that the user did this scenario in older version or he did it before enabaling the modern toolbar feature.
priority
modern toolbar ui is not working after saving window configuration on window with shrink mode a user reported that the new toolbar ui is not showing even when enabling the modern toolbar feature i asked the user for the windowtop settings file and he sent it i found that the configuration that saved in this file was reproduced the issue while debugging the issue i found that the configuration key that reproduces it is arrowxpos it should not be like that it should be some process exe arrowxpos i reviewed the code and according to the code review this configuration issue is caused in the following scenario disable the modern toolbar shrink some window resize the shrink box so the arrow button will show when you move to the top area drag the arrow button to some another position unshrink the window or exit windowtop at this point this configuration issue will reproduce because this configuration issue affects only the modern toolbar ui you will need to enable the modern toolbar ui feature to reproduce it seems that the user did this scenario in older version or he did it before enabaling the modern toolbar feature
1
804,372
29,485,251,542
IssuesEvent
2023-06-02 09:14:22
Avaiga/taipy-core
https://api.github.com/repos/Avaiga/taipy-core
closed
Event: fire an event when scenario name is modified
🟧 Priority: High 📈 Improvement
**Description** An event should be fired when scenario name is modified.
1.0
Event: fire an event when scenario name is modified - **Description** An event should be fired when scenario name is modified.
priority
event fire an event when scenario name is modified description an event should be fired when scenario name is modified
1
708,856
24,357,537,377
IssuesEvent
2022-10-03 08:49:19
NethermindEth/nethermind
https://api.github.com/repos/NethermindEth/nethermind
closed
[Peers] When Peers number drop to 0, nethermind have problems to recover
high priority stability
**Describe the bug** After some network issues peers number drops to 0. For about 20 minutes it was logging "Waiting for Peers" and nothing interesting happened. After restart of nethermind app, peers came back properly. **To Reproduce** Steps to reproduce the behavior: Disconnect the network for a moment and check if peers dropped to 0 and then recover network connection. **Expected behavior** After reconnecting we should be able to recover and find peers back. **Additional context** Add any other context about the problem here.
1.0
[Peers] When Peers number drop to 0, nethermind have problems to recover - **Describe the bug** After some network issues peers number drops to 0. For about 20 minutes it was logging "Waiting for Peers" and nothing interesting happened. After restart of nethermind app, peers came back properly. **To Reproduce** Steps to reproduce the behavior: Disconnect the network for a moment and check if peers dropped to 0 and then recover network connection. **Expected behavior** After reconnecting we should be able to recover and find peers back. **Additional context** Add any other context about the problem here.
priority
when peers number drop to nethermind have problems to recover describe the bug after some network issues peers number drops to for about minutes it was logging waiting for peers and nothing interesting happened after restart of nethermind app peers came back properly to reproduce steps to reproduce the behavior disconnect the network for a moment and check if peers dropped to and then recover network connection expected behavior after reconnecting we should be able to recover and find peers back additional context add any other context about the problem here
1
321,249
9,795,819,627
IssuesEvent
2019-06-11 05:36:34
apache/skywalking
https://api.github.com/repos/apache/skywalking
opened
Time series ElasticSearch implementation bug of record type
OAP-backend bug high priority
As the following screenshot shows, the segment(also alarm and other record types) don't create index by day, but in hour wrongly. ![image](https://user-images.githubusercontent.com/5441976/59246051-b5ae9a00-8c4d-11e9-9f11-829f86ef0373.png) This needs to be fixed and make sure the TTL works for it(not working today, possible because of wrong table name).
1.0
Time series ElasticSearch implementation bug of record type - As the following screenshot shows, the segment(also alarm and other record types) don't create index by day, but in hour wrongly. ![image](https://user-images.githubusercontent.com/5441976/59246051-b5ae9a00-8c4d-11e9-9f11-829f86ef0373.png) This needs to be fixed and make sure the TTL works for it(not working today, possible because of wrong table name).
priority
time series elasticsearch implementation bug of record type as the following screenshot shows the segment also alarm and other record types don t create index by day but in hour wrongly this needs to be fixed and make sure the ttl works for it not working today possible because of wrong table name
1
409,446
11,962,589,959
IssuesEvent
2020-04-05 13:01:11
traffic-control-fyp-aub/ns3-gym
https://api.github.com/repos/traffic-control-fyp-aub/ns3-gym
closed
System Avg Speed [ Performance ]
High Priority
Study the effect of different vehicle control percentages (incorporate human driver control) on the system wide average speed and whether it increases or decreases.
1.0
System Avg Speed [ Performance ] - Study the effect of different vehicle control percentages (incorporate human driver control) on the system wide average speed and whether it increases or decreases.
priority
system avg speed study the effect of different vehicle control percentages incorporate human driver control on the system wide average speed and whether it increases or decreases
1
440,283
12,697,315,600
IssuesEvent
2020-06-22 11:36:19
chiyadev/MudaeFarm
https://api.github.com/repos/chiyadev/MudaeFarm
closed
Bot breaks and spams chat
bug high priority
At some point the farm falls apart and fails to parse anything Mudaebot sends, even normal rolls. It seems insane to me that MudaeFarm would try every ~30 seconds indefinitely and create hundreds of messages. Unsanitized logs below, timestamp is 2020-06-19T21:59:41.2842398-05:00 [log_2020-06-19 06.22.24Z-20200619.txt](https://github.com/chiyadev/MudaeFarm/files/4807768/log_2020-06-19.06.22.24Z-20200619.txt) EDIT: I just realized these logs aren't verbose. Given it's been happening every night, I'll have some soon
1.0
Bot breaks and spams chat - At some point the farm falls apart and fails to parse anything Mudaebot sends, even normal rolls. It seems insane to me that MudaeFarm would try every ~30 seconds indefinitely and create hundreds of messages. Unsanitized logs below, timestamp is 2020-06-19T21:59:41.2842398-05:00 [log_2020-06-19 06.22.24Z-20200619.txt](https://github.com/chiyadev/MudaeFarm/files/4807768/log_2020-06-19.06.22.24Z-20200619.txt) EDIT: I just realized these logs aren't verbose. Given it's been happening every night, I'll have some soon
priority
bot breaks and spams chat at some point the farm falls apart and fails to parse anything mudaebot sends even normal rolls it seems insane to me that mudaefarm would try every seconds indefinitely and create hundreds of messages unsanitized logs below timestamp is edit i just realized these logs aren t verbose given it s been happening every night i ll have some soon
1
678,068
23,186,038,094
IssuesEvent
2022-08-01 08:28:08
netdata/netdata-cloud
https://api.github.com/repos/netdata/netdata-cloud
closed
Enhance the Alerts drawer / modal in the Active alerts and Alerts Configuration Page
priority/high cloud-frontend cloud-backend alerts-team Q2 GOAL feature request
### Problem The current Alerts drawer / modal does not display the details of node instances where this alert has been raised and how it has been aggregated at the node level. ### Description The alert drawer is currently being used in the Active Alerts tab and will need to be used also in the "Alerts Configuration" / "Manage Alerts" tab. The drawer / model needs to be enhanced to provide more clear information on the alert itself: - The Chart relevant to the alert (already available) - Details of the alert event as a vertical bar with value (already available for Warning and Critical) also for the Clear state. - Table with Node instances where the alert was raised (with criticality) - Alert configurations in a more easily readable and understandable way for each node instance - CTA to view the alert configuration at the node level (on a dedicated page?) This can possibly be extended to also show historical logs relevant to the specific alert within the drawer. cc: @ktsaou @amalkov @vinnygats @car12o @YaroslavDev @novykh @jacekkolasa ### Importance must have ### Value proposition 1. The current alert drawer does not convey a clear description of the alert 2. The user will need to see details at the node instance level 3. Possible future extension to also show the historical logs relevant to the alert ... ### Proposed implementation _No response_
1.0
Enhance the Alerts drawer / modal in the Active alerts and Alerts Configuration Page - ### Problem The current Alerts drawer / modal does not display the details of node instances where this alert has been raised and how it has been aggregated at the node level. ### Description The alert drawer is currently being used in the Active Alerts tab and will need to be used also in the "Alerts Configuration" / "Manage Alerts" tab. The drawer / model needs to be enhanced to provide more clear information on the alert itself: - The Chart relevant to the alert (already available) - Details of the alert event as a vertical bar with value (already available for Warning and Critical) also for the Clear state. - Table with Node instances where the alert was raised (with criticality) - Alert configurations in a more easily readable and understandable way for each node instance - CTA to view the alert configuration at the node level (on a dedicated page?) This can possibly be extended to also show historical logs relevant to the specific alert within the drawer. cc: @ktsaou @amalkov @vinnygats @car12o @YaroslavDev @novykh @jacekkolasa ### Importance must have ### Value proposition 1. The current alert drawer does not convey a clear description of the alert 2. The user will need to see details at the node instance level 3. Possible future extension to also show the historical logs relevant to the alert ... ### Proposed implementation _No response_
priority
enhance the alerts drawer modal in the active alerts and alerts configuration page problem the current alerts drawer modal does not display the details of node instances where this alert has been raised and how it has been aggregated at the node level description the alert drawer is currently being used in the active alerts tab and will need to be used also in the alerts configuration manage alerts tab the drawer model needs to be enhanced to provide more clear information on the alert itself the chart relevant to the alert already available details of the alert event as a vertical bar with value already available for warning and critical also for the clear state table with node instances where the alert was raised with criticality alert configurations in a more easily readable and understandable way for each node instance cta to view the alert configuration at the node level on a dedicated page this can possibly be extended to also show historical logs relevant to the specific alert within the drawer cc ktsaou amalkov vinnygats yaroslavdev novykh jacekkolasa importance must have value proposition the current alert drawer does not convey a clear description of the alert the user will need to see details at the node instance level possible future extension to also show the historical logs relevant to the alert proposed implementation no response
1
528,619
15,370,896,137
IssuesEvent
2021-03-02 09:22:16
enthought/enable
https://api.github.com/repos/enthought/enable
closed
draw_rect behaves weirdly with QPainter backend
priority: high type: bug
|agg|QPainter| |---|---| |<img width="256" alt="kiva agg draw_rect 2x" src="https://user-images.githubusercontent.com/1926457/109619792-1623f000-7b31-11eb-90e4-b5590aba5d9c.png">|<img width="256" alt="qpainter draw_rect 2x" src="https://user-images.githubusercontent.com/1926457/109619815-18864a00-7b31-11eb-900d-cae1da311d98.png">| The above images have been generated using the new `enable.gcbench` on windows with PyQt5.
1.0
draw_rect behaves weirdly with QPainter backend - |agg|QPainter| |---|---| |<img width="256" alt="kiva agg draw_rect 2x" src="https://user-images.githubusercontent.com/1926457/109619792-1623f000-7b31-11eb-90e4-b5590aba5d9c.png">|<img width="256" alt="qpainter draw_rect 2x" src="https://user-images.githubusercontent.com/1926457/109619815-18864a00-7b31-11eb-900d-cae1da311d98.png">| The above images have been generated using the new `enable.gcbench` on windows with PyQt5.
priority
draw rect behaves weirdly with qpainter backend agg qpainter img width alt kiva agg draw rect src width alt qpainter draw rect src the above images have been generated using the new enable gcbench on windows with
1
413,788
12,092,155,253
IssuesEvent
2020-04-19 14:35:40
lorenzwalthert/precommit
https://api.github.com/repos/lorenzwalthert/precommit
closed
Allow to choose installation environment
Complexity: Medium Priority: High Status: WIP Type: Enhancement
As mentioned in [#113 ](https://github.com/lorenzwalthert/precommit/issues/113#issuecomment-603808455). Recently I've tried to run `keras::install_keras()` inside the Docker image with already installed precommit. Looks like both keras and tensorflow are now [installed by default to `r-reticulate`](https://github.com/rstudio/keras/issues/1014). This results in the following error: >Collecting package metadata (current_repodata.json): ...working... done Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve. Solving environment: ...working... failed with repodata from current_repodata.json, will retry with next repodata source. Collecting package metadata (repodata.json): ...working... done Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve. Examining setuptools: 7%|▋ | 2/30 [00:00<00:00, 2603.54it/s] Comparing specs that have this dependency: 0%| | 0/15 [00:00<?, ?it/s] Finding conflict paths: 0%| | 0/2 [00:00<?, ?it/s] Finding shortest conflict path for setuptools: 0%| | 0/2 [00:00<?, ?it/s] Finding shortest conflict path for setuptools: 50%|█████ | 1/2 [00:01<00:01, 1.99s/it] Finding shortest conflict path for setuptools: 100%|██████████| 2/2 [00:01<00:00, 1.01it/s] ...truncated...
1.0
Allow to choose installation environment - As mentioned in [#113 ](https://github.com/lorenzwalthert/precommit/issues/113#issuecomment-603808455). Recently I've tried to run `keras::install_keras()` inside the Docker image with already installed precommit. Looks like both keras and tensorflow are now [installed by default to `r-reticulate`](https://github.com/rstudio/keras/issues/1014). This results in the following error: >Collecting package metadata (current_repodata.json): ...working... done Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve. Solving environment: ...working... failed with repodata from current_repodata.json, will retry with next repodata source. Collecting package metadata (repodata.json): ...working... done Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve. Examining setuptools: 7%|▋ | 2/30 [00:00<00:00, 2603.54it/s] Comparing specs that have this dependency: 0%| | 0/15 [00:00<?, ?it/s] Finding conflict paths: 0%| | 0/2 [00:00<?, ?it/s] Finding shortest conflict path for setuptools: 0%| | 0/2 [00:00<?, ?it/s] Finding shortest conflict path for setuptools: 50%|█████ | 1/2 [00:01<00:01, 1.99s/it] Finding shortest conflict path for setuptools: 100%|██████████| 2/2 [00:01<00:00, 1.01it/s] ...truncated...
priority
allow to choose installation environment as mentioned in recently i ve tried to run keras install keras inside the docker image with already installed precommit looks like both keras and tensorflow are now this results in the following error collecting package metadata current repodata json working done solving environment working failed with initial frozen solve retrying with flexible solve solving environment working failed with repodata from current repodata json will retry with next repodata source collecting package metadata repodata json working done solving environment working failed with initial frozen solve retrying with flexible solve examining setuptools ▋ comparing specs that have this dependency finding conflict paths finding shortest conflict path for setuptools finding shortest conflict path for setuptools █████ finding shortest conflict path for setuptools ██████████ truncated
1
813,717
30,469,039,169
IssuesEvent
2023-07-17 12:29:53
tinkoff-ai/etna
https://api.github.com/repos/tinkoff-ai/etna
opened
LimitTransform
enhancement priority/high
### 🚀 Feature Request Create a transform that limits values of some feature between the borders. ### Proposal Create `LimitTransform`. Parameters: - `in_column`: column to make transformation on; - `lower_bound`: lower bound for the value of the column; -infty by default; - `upper_bound`: upper bound for the value of the column; +infty by default; - If there is value out of limit the exception should be raised. - NaNs should be ignored. Reference: [Ensure time series forecasts stay within limits](https://datasciencestunt.com/time-series-forecasting-within-limits/). To discuss: - Should this transform work for non-target column? - Should this transform have `inplace` parameter for working in non-inplace mode? ### Test cases What should be checked: - Working on target / non-target column - Working with set/unset lower/upper values - Exception on out-of-limit value - Full pipeline that predicts some arbitrary values can be used Don't forget to add inference tests into `tests/test_models/test_inference/`. ### Additional context _No response_
1.0
LimitTransform - ### 🚀 Feature Request Create a transform that limits values of some feature between the borders. ### Proposal Create `LimitTransform`. Parameters: - `in_column`: column to make transformation on; - `lower_bound`: lower bound for the value of the column; -infty by default; - `upper_bound`: upper bound for the value of the column; +infty by default; - If there is value out of limit the exception should be raised. - NaNs should be ignored. Reference: [Ensure time series forecasts stay within limits](https://datasciencestunt.com/time-series-forecasting-within-limits/). To discuss: - Should this transform work for non-target column? - Should this transform have `inplace` parameter for working in non-inplace mode? ### Test cases What should be checked: - Working on target / non-target column - Working with set/unset lower/upper values - Exception on out-of-limit value - Full pipeline that predicts some arbitrary values can be used Don't forget to add inference tests into `tests/test_models/test_inference/`. ### Additional context _No response_
priority
limittransform 🚀 feature request create a transform that limits values of some feature between the borders proposal create limittransform parameters in column column to make transformation on lower bound lower bound for the value of the column infty by default upper bound upper bound for the value of the column infty by default if there is value out of limit the exception should be raised nans should be ignored reference to discuss should this transform work for non target column should this transform have inplace parameter for working in non inplace mode test cases what should be checked working on target non target column working with set unset lower upper values exception on out of limit value full pipeline that predicts some arbitrary values can be used don t forget to add inference tests into tests test models test inference additional context no response
1
480,829
13,876,853,552
IssuesEvent
2020-10-17 01:02:37
webiny/webiny-js
https://api.github.com/repos/webiny/webiny-js
closed
Headless CMS - Access Token unable to access listContentModels
bug cms no-issue-activity priority: high
This is: - Bug ## Expected Behavior Using Access Token with /read API should allow me to use the `listConentModels` field. ## Actual Behavior Throws `Not Authorized` error. ![image](https://user-images.githubusercontent.com/3920893/85414395-57a8d800-b56c-11ea-9868-93f266f57339.png)
1.0
Headless CMS - Access Token unable to access listContentModels - This is: - Bug ## Expected Behavior Using Access Token with /read API should allow me to use the `listConentModels` field. ## Actual Behavior Throws `Not Authorized` error. ![image](https://user-images.githubusercontent.com/3920893/85414395-57a8d800-b56c-11ea-9868-93f266f57339.png)
priority
headless cms access token unable to access listcontentmodels this is bug expected behavior using access token with read api should allow me to use the listconentmodels field actual behavior throws not authorized error
1
420,768
12,243,764,121
IssuesEvent
2020-05-05 09:50:43
pokt-network/pocket-core
https://api.github.com/repos/pokt-network/pocket-core
opened
Remove HTTP Round Trip for Queries
enhancement high priority optimization performance
**Is your feature request related to a problem? Please describe.** Remove HTTP calls for queries, change to a programmatic interface. Remove all QueryWithData calls and replace them with a direct programmatic interface. **This is high priority for relays/dispatches only**
1.0
Remove HTTP Round Trip for Queries - **Is your feature request related to a problem? Please describe.** Remove HTTP calls for queries, change to a programmatic interface. Remove all QueryWithData calls and replace them with a direct programmatic interface. **This is high priority for relays/dispatches only**
priority
remove http round trip for queries is your feature request related to a problem please describe remove http calls for queries change to a programmatic interface remove all querywithdata calls and replace them with a direct programmatic interface this is high priority for relays dispatches only
1
443,215
12,761,475,610
IssuesEvent
2020-06-29 11:32:29
getkirby/kirby
https://api.github.com/repos/getkirby/kirby
closed
CURLOPT_SSL_VERIFYPEER disabled in Remote::fetch()
priority: high 🔥 type: bug 🐛
Hello, I just took a look at `Remote::fetch()` due to some forum question and I noticed that `CURLOPT_SSL_VERIFYPEER` is set to `FALSE` by default. Is there a reason why this security feature is disabled by default? I would have expected SSL cert validation to be enabled by default or to find some information why this is disabled but I found nothing about this. See this line: https://github.com/getkirby/kirby/blob/6bd14bc8099d6c9dbe67ef0582b8c0a2d4c0888c/src/Http/Remote.php#L152 I'm sorry if this should have been posted in the forum or if it's an obvious misunderstanding on my side. Best Regards,
1.0
CURLOPT_SSL_VERIFYPEER disabled in Remote::fetch() - Hello, I just took a look at `Remote::fetch()` due to some forum question and I noticed that `CURLOPT_SSL_VERIFYPEER` is set to `FALSE` by default. Is there a reason why this security feature is disabled by default? I would have expected SSL cert validation to be enabled by default or to find some information why this is disabled but I found nothing about this. See this line: https://github.com/getkirby/kirby/blob/6bd14bc8099d6c9dbe67ef0582b8c0a2d4c0888c/src/Http/Remote.php#L152 I'm sorry if this should have been posted in the forum or if it's an obvious misunderstanding on my side. Best Regards,
priority
curlopt ssl verifypeer disabled in remote fetch hello i just took a look at remote fetch due to some forum question and i noticed that curlopt ssl verifypeer is set to false by default is there a reason why this security feature is disabled by default i would have expected ssl cert validation to be enabled by default or to find some information why this is disabled but i found nothing about this see this line i m sorry if this should have been posted in the forum or if it s an obvious misunderstanding on my side best regards
1
684,578
23,422,835,725
IssuesEvent
2022-08-14 00:28:54
ChaosInitiative/Portal-2-Community-Edition
https://api.github.com/repos/ChaosInitiative/Portal-2-Community-Edition
closed
Bug: Multiplayer players spawn at info_player_start instead of info_coop_spawn
Type: bug Component: gameplay Priority 2: High Status: resolved Focus: Co-Op Size 5: Tiny
### Describe the bug Self-explanatory... This is in the most recent nightly build. Despite the issue being super obvious, I didn't hear anyone talk about it yet so I just wanted to make an issue for it. Happens regardless of player team. ### Issue Map All maps. ### To Reproduce 1. Set ConVar **mp_dev_wait_for_other_player** to 0 2. Start a multiplayer game through **map mp_coop_2paints_1bridge** 3. Player spawns at info_player_start ### Operating System Tested on Windows 10
1.0
Bug: Multiplayer players spawn at info_player_start instead of info_coop_spawn - ### Describe the bug Self-explanatory... This is in the most recent nightly build. Despite the issue being super obvious, I didn't hear anyone talk about it yet so I just wanted to make an issue for it. Happens regardless of player team. ### Issue Map All maps. ### To Reproduce 1. Set ConVar **mp_dev_wait_for_other_player** to 0 2. Start a multiplayer game through **map mp_coop_2paints_1bridge** 3. Player spawns at info_player_start ### Operating System Tested on Windows 10
priority
bug multiplayer players spawn at info player start instead of info coop spawn describe the bug self explanatory this is in the most recent nightly build despite the issue being super obvious i didn t hear anyone talk about it yet so i just wanted to make an issue for it happens regardless of player team issue map all maps to reproduce set convar mp dev wait for other player to start a multiplayer game through map mp coop player spawns at info player start operating system tested on windows
1
312,831
9,553,708,880
IssuesEvent
2019-05-02 20:00:17
eJourn-al/eJournal
https://api.github.com/repos/eJourn-al/eJournal
closed
Non-required file upload throws an error when left empty
Priority: High Status: Review Needed Type: Bug Workload: Low
**Describe the bug** The back end always denies requests that contain `None` for a file upload field. **To Reproduce** Steps to reproduce the behavior: 1. Make a template with a non-required file upload field 2. Try to post it as a student 3. Observe that the server returns a bad request stating `One of your files was not correctly uploaded, please try gain.` **Expected behavior** Non-required fields to be non-required.
1.0
Non-required file upload throws an error when left empty - **Describe the bug** The back end always denies requests that contain `None` for a file upload field. **To Reproduce** Steps to reproduce the behavior: 1. Make a template with a non-required file upload field 2. Try to post it as a student 3. Observe that the server returns a bad request stating `One of your files was not correctly uploaded, please try gain.` **Expected behavior** Non-required fields to be non-required.
priority
non required file upload throws an error when left empty describe the bug the back end always denies requests that contain none for a file upload field to reproduce steps to reproduce the behavior make a template with a non required file upload field try to post it as a student observe that the server returns a bad request stating one of your files was not correctly uploaded please try gain expected behavior non required fields to be non required
1
668,301
22,577,064,654
IssuesEvent
2022-06-28 08:19:23
freesewing/freesewing
https://api.github.com/repos/freesewing/freesewing
closed
Ursula is not drafting in 2.21.0
:bug: bug :rotating_light: high priority :package: ursula
With either inputed measures or the "standard measurses* Ursula is not drafting in 2.21.0. I'm not 100 but I suspect this is the splitting packages folders.
1.0
Ursula is not drafting in 2.21.0 - With either inputed measures or the "standard measurses* Ursula is not drafting in 2.21.0. I'm not 100 but I suspect this is the splitting packages folders.
priority
ursula is not drafting in with either inputed measures or the standard measurses ursula is not drafting in i m not but i suspect this is the splitting packages folders
1
259,855
8,200,702,870
IssuesEvent
2018-09-01 08:05:08
marvinlabs/customer-area
https://api.github.com/repos/marvinlabs/customer-area
opened
Bug in reset password
Priority - high bug
Some keys are generated with special characters (e.g. `1535687074:$P$Bdz168aJBFAFUOaQpZZHK6oArvdxDk0` ) but the add-on code seems to be removing those characters **before** comparison with in-db user key.
1.0
Bug in reset password - Some keys are generated with special characters (e.g. `1535687074:$P$Bdz168aJBFAFUOaQpZZHK6oArvdxDk0` ) but the add-on code seems to be removing those characters **before** comparison with in-db user key.
priority
bug in reset password some keys are generated with special characters e g p but the add on code seems to be removing those characters before comparison with in db user key
1
619,297
19,521,364,126
IssuesEvent
2021-12-29 19:09:51
Cotalker/documentation
https://api.github.com/repos/Cotalker/documentation
closed
Bug report: Vista de task se pega al cargar
Bug report Bug high priority Bug rejected
### Affected system Cotalker Web Application ### Affected system (other) _No response_ ### Affected environment Production ### Affected environment (other) _No response_ ### App version 17.5.8 ### Details Al cargar la vista de task, se queda la pestaña de chrome pegada ![bug_task_251121](https://user-images.githubusercontent.com/50208672/143498962-d47c3512-db50-408c-ae5e-fd741edf06cc.png) ** No permite desplegar sección ** ![image](https://user-images.githubusercontent.com/50208672/143500301-60ee68db-15b7-45cc-a350-cbbbd6dc15e4.png) ### Steps to reproduce Abrir vista de task en un flujo de trabajo que tenga múltiple task ### Expected result Que no se pegue la página y se carguen de forma correcta las tasks. ### Additional data - Company: Cotalker - Group : Operaciones (sub flujo optask)
1.0
Bug report: Vista de task se pega al cargar - ### Affected system Cotalker Web Application ### Affected system (other) _No response_ ### Affected environment Production ### Affected environment (other) _No response_ ### App version 17.5.8 ### Details Al cargar la vista de task, se queda la pestaña de chrome pegada ![bug_task_251121](https://user-images.githubusercontent.com/50208672/143498962-d47c3512-db50-408c-ae5e-fd741edf06cc.png) ** No permite desplegar sección ** ![image](https://user-images.githubusercontent.com/50208672/143500301-60ee68db-15b7-45cc-a350-cbbbd6dc15e4.png) ### Steps to reproduce Abrir vista de task en un flujo de trabajo que tenga múltiple task ### Expected result Que no se pegue la página y se carguen de forma correcta las tasks. ### Additional data - Company: Cotalker - Group : Operaciones (sub flujo optask)
priority
bug report vista de task se pega al cargar affected system cotalker web application affected system other no response affected environment production affected environment other no response app version details al cargar la vista de task se queda la pestaña de chrome pegada no permite desplegar sección steps to reproduce abrir vista de task en un flujo de trabajo que tenga múltiple task expected result que no se pegue la página y se carguen de forma correcta las tasks additional data company cotalker group operaciones sub flujo optask
1
105,749
4,241,211,507
IssuesEvent
2016-07-06 15:40:25
ccswbs/hjckrrh
https://api.github.com/repos/ccswbs/hjckrrh
closed
G- Establish default front page layouts for pages
feature: general (G) priority: high type: enhancement request
This should be specified (documented requirements)- see Trello card- G14-#240 https://trello.com/c/ZkCKfdnw
1.0
G- Establish default front page layouts for pages - This should be specified (documented requirements)- see Trello card- G14-#240 https://trello.com/c/ZkCKfdnw
priority
g establish default front page layouts for pages this should be specified documented requirements see trello card
1
667,161
22,420,269,662
IssuesEvent
2022-06-20 01:45:26
portefaix/portefaix
https://api.github.com/repos/portefaix/portefaix
closed
AKS: Ingress Application Gateway
priority/high kind/feature area/terraform lifecycle/stale lifecycle/frozen cloud/azure todo
AKS: Ingress Application Gateway - https://github.com/Azure/terraform-azurerm-aks/pull/99 enable_ingress_application_gateway = true ingress_application_gateway_gateway_name = ingress_application_gateway_subnet_cidr = ingress_application_gateway_subnet_id = https://github.com/portefaix/portefaix/blob/739c1d3f62ac9ddb95d4c2cfd684529098216e8c/terraform/azure/aks/modules/aks/aks.tf#L53 ```ruby # rbac_aad_managed = false # rbac_aad_admin_group_object_ids = var.admin_group_object_ids enable_log_analytics_workspace = false enable_auto_scaling = var.enable_auto_scaling enable_kube_dashboard = var.enable_kube_dashboard enable_azure_policy = var.enable_azure_policy enable_http_application_routing = var.enable_http_application_routing # TODO: AKS: Ingress Application Gateway # labels: kind/feature, priority/high, lifecycle/frozen, area/terraform, cloud/azure # https://github.com/Azure/terraform-azurerm-aks/pull/99 # enable_ingress_application_gateway = true # ingress_application_gateway_gateway_name = # ingress_application_gateway_subnet_cidr = # ingress_application_gateway_subnet_id = os_disk_size_gb = var.os_disk_size_gb agents_min_count = var.agents_min_count ``` afce5cfdca8a7775d0e82c8e7e9387f68d5e7736
1.0
AKS: Ingress Application Gateway - AKS: Ingress Application Gateway - https://github.com/Azure/terraform-azurerm-aks/pull/99 enable_ingress_application_gateway = true ingress_application_gateway_gateway_name = ingress_application_gateway_subnet_cidr = ingress_application_gateway_subnet_id = https://github.com/portefaix/portefaix/blob/739c1d3f62ac9ddb95d4c2cfd684529098216e8c/terraform/azure/aks/modules/aks/aks.tf#L53 ```ruby # rbac_aad_managed = false # rbac_aad_admin_group_object_ids = var.admin_group_object_ids enable_log_analytics_workspace = false enable_auto_scaling = var.enable_auto_scaling enable_kube_dashboard = var.enable_kube_dashboard enable_azure_policy = var.enable_azure_policy enable_http_application_routing = var.enable_http_application_routing # TODO: AKS: Ingress Application Gateway # labels: kind/feature, priority/high, lifecycle/frozen, area/terraform, cloud/azure # https://github.com/Azure/terraform-azurerm-aks/pull/99 # enable_ingress_application_gateway = true # ingress_application_gateway_gateway_name = # ingress_application_gateway_subnet_cidr = # ingress_application_gateway_subnet_id = os_disk_size_gb = var.os_disk_size_gb agents_min_count = var.agents_min_count ``` afce5cfdca8a7775d0e82c8e7e9387f68d5e7736
priority
aks ingress application gateway aks ingress application gateway enable ingress application gateway true ingress application gateway gateway name ingress application gateway subnet cidr ingress application gateway subnet id ruby rbac aad managed false  rbac aad admin group object ids var admin group object ids enable log analytics workspace false enable auto scaling var enable auto scaling enable kube dashboard var enable kube dashboard enable azure policy var enable azure policy enable http application routing var enable http application routing todo aks ingress application gateway labels kind feature priority high lifecycle frozen area terraform cloud azure enable ingress application gateway true ingress application gateway gateway name ingress application gateway subnet cidr ingress application gateway subnet id os disk size gb var os disk size gb agents min count var agents min count
1
6,304
2,587,112,727
IssuesEvent
2015-02-17 16:30:01
civio/quienmanda.es
https://api.github.com/repos/civio/quienmanda.es
closed
Implementar 'páginas temáticas'
high_priority in progress
Queremos agrupar una serie de fotos y artículos en una "página temática" que compile y muestre de una forma agradable toda la información sobre un tema. Primer uso: compilar los artículos sobre el Colegio de El Pilar. Comenzaremos mostrando simplemente una serie de fotos y/o artículos etiquetados con una palabra X, pero podría añadirse una foto y/o texto de entrada, dando contexto. Posteriormente esto aplica a temas como 'energía', 'banca'...
1.0
Implementar 'páginas temáticas' - Queremos agrupar una serie de fotos y artículos en una "página temática" que compile y muestre de una forma agradable toda la información sobre un tema. Primer uso: compilar los artículos sobre el Colegio de El Pilar. Comenzaremos mostrando simplemente una serie de fotos y/o artículos etiquetados con una palabra X, pero podría añadirse una foto y/o texto de entrada, dando contexto. Posteriormente esto aplica a temas como 'energía', 'banca'...
priority
implementar páginas temáticas queremos agrupar una serie de fotos y artículos en una página temática que compile y muestre de una forma agradable toda la información sobre un tema primer uso compilar los artículos sobre el colegio de el pilar comenzaremos mostrando simplemente una serie de fotos y o artículos etiquetados con una palabra x pero podría añadirse una foto y o texto de entrada dando contexto posteriormente esto aplica a temas como energía banca
1
283,012
8,712,895,467
IssuesEvent
2018-12-06 23:59:53
aowen87/TicketTester
https://api.github.com/repos/aowen87/TicketTester
closed
VisIt hangs during re-execution prompted by pick.
bug crash likelihood medium priority reviewed severity high wrong results
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps. This seems only to occur with 2 nodes. Multiple processors on single node does not replicate. Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var) On surface: (ensure parallel engine with 2 nodes) Open multi_curv2d.silo Add PC Plot of d Add CoordSwap operator, swap x and y coords Draw. Create a scalar expression d+p, Open Pick window, set variable to d+p. Apply Do a Zone Pick Using Navigation, change the view either by zooming or panning. Do another Zone Pick. Engine will hang, must cancel the engine_par job in order to interact with VisIt again. Information window shows Pick wanting to re-execute, and a merge exception: ++++++++++++++++ VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks. Shortly thereafter, the following occured... Pseudocolor: (InvalidMergeException) viewer: Cannot merge datasets because of an incompatible field 1 and 2. Pick mode now fully ready. +++++++++++++++++++++ This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error: +++++++++++++++++++++++++++++++++++++ This source should not load balance the data. Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2. catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027 ++++++++++++++++++++++++++++++ This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2169 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: VisIt hangs during re-execution prompted by pick. Assigned to: Kathleen Biagas Category: Target version: 2.9.1 Author: Kathleen Biagas Start: 03/03/2015 Due date: % Done: 100 Estimated time: Created: 03/03/2015 04:55 pm Updated: 03/20/2015 05:44 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.8.2 Impact: Expected Use: OS: All Support Group: Any Description: This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps. This seems only to occur with 2 nodes. Multiple processors on single node does not replicate. Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var) On surface: (ensure parallel engine with 2 nodes) Open multi_curv2d.silo Add PC Plot of d Add CoordSwap operator, swap x and y coords Draw. Create a scalar expression d+p, Open Pick window, set variable to d+p. Apply Do a Zone Pick Using Navigation, change the view either by zooming or panning. Do another Zone Pick. Engine will hang, must cancel the engine_par job in order to interact with VisIt again. Information window shows Pick wanting to re-execute, and a merge exception: ++++++++++++++++ VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks. Shortly thereafter, the following occured... Pseudocolor: (InvalidMergeException) viewer: Cannot merge datasets because of an incompatible field 1 and 2. Pick mode now fully ready. +++++++++++++++++++++ This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error: +++++++++++++++++++++++++++++++++++++ This source should not load balance the data. Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2. catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027 ++++++++++++++++++++++++++++++ This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors. Comments: Turns out that Pick was sending the secondary variable request to only 1 processor, which caused an 'Invalid Merge' exception during pipeline re-execution.I modified Pick to request SecondaryVars of all processors.SVN update 25986 (2.9RC), 25988 (trunk)M /src/avt/Queries/Pick/avtPickQuery.C
1.0
VisIt hangs during re-execution prompted by pick. - This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps. This seems only to occur with 2 nodes. Multiple processors on single node does not replicate. Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var) On surface: (ensure parallel engine with 2 nodes) Open multi_curv2d.silo Add PC Plot of d Add CoordSwap operator, swap x and y coords Draw. Create a scalar expression d+p, Open Pick window, set variable to d+p. Apply Do a Zone Pick Using Navigation, change the view either by zooming or panning. Do another Zone Pick. Engine will hang, must cancel the engine_par job in order to interact with VisIt again. Information window shows Pick wanting to re-execute, and a merge exception: ++++++++++++++++ VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks. Shortly thereafter, the following occured... Pseudocolor: (InvalidMergeException) viewer: Cannot merge datasets because of an incompatible field 1 and 2. Pick mode now fully ready. +++++++++++++++++++++ This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error: +++++++++++++++++++++++++++++++++++++ This source should not load balance the data. Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2. catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027 ++++++++++++++++++++++++++++++ This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2169 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: VisIt hangs during re-execution prompted by pick. Assigned to: Kathleen Biagas Category: Target version: 2.9.1 Author: Kathleen Biagas Start: 03/03/2015 Due date: % Done: 100 Estimated time: Created: 03/03/2015 04:55 pm Updated: 03/20/2015 05:44 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.8.2 Impact: Expected Use: OS: All Support Group: Any Description: This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps. This seems only to occur with 2 nodes. Multiple processors on single node does not replicate. Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var) On surface: (ensure parallel engine with 2 nodes) Open multi_curv2d.silo Add PC Plot of d Add CoordSwap operator, swap x and y coords Draw. Create a scalar expression d+p, Open Pick window, set variable to d+p. Apply Do a Zone Pick Using Navigation, change the view either by zooming or panning. Do another Zone Pick. Engine will hang, must cancel the engine_par job in order to interact with VisIt again. Information window shows Pick wanting to re-execute, and a merge exception: ++++++++++++++++ VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks. Shortly thereafter, the following occured... Pseudocolor: (InvalidMergeException) viewer: Cannot merge datasets because of an incompatible field 1 and 2. Pick mode now fully ready. +++++++++++++++++++++ This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error: +++++++++++++++++++++++++++++++++++++ This source should not load balance the data. Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2. catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027 ++++++++++++++++++++++++++++++ This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors. Comments: Turns out that Pick was sending the secondary variable request to only 1 processor, which caused an 'Invalid Merge' exception during pipeline re-execution.I modified Pick to request SecondaryVars of all processors.SVN update 25986 (2.9RC), 25988 (trunk)M /src/avt/Queries/Pick/avtPickQuery.C
priority
visit hangs during re execution prompted by pick this is a bug bruce hammel claims to have been experiencing with visit for years so i ve set the priority as high now that i have reproducible steps this seems only to occur with nodes multiple processors on single node does not replicate also seems only to occur in conjunction with coordswap operator and with pick var being set to an expression and not the pipeline var on surface ensure parallel engine with nodes open multi silo add pc plot of d add coordswap operator swap x and y coords draw create a scalar expression d p open pick window set variable to d p apply do a zone pick using navigation change the view either by zooming or panning do another zone pick engine will hang must cancel the engine par job in order to interact with visit again information window shows pick wanting to re execute and a merge exception visit does not have all the information it needs to perform a pick please wait while the necessary information is calculated all current pick selections have been cached and will be performed when calculations are complete visit will notify you when it is fully ready for more picks shortly thereafter the following occured pseudocolor invalidmergeexception viewer cannot merge datasets because of an incompatible field and pick mode now fully ready this shows error seeming to come from the viewer but if you run with debug then processors log files will show this error this source should not load balance the data exception invalidmergeexception usr tmp brugger aztec visitbuild src avt pipeline data avtdataattributes c line cannot merge datasets because of an incompatible field and catch visitexception usr tmp brugger aztec visitbuild src engine main executors h this is consistent with what i saw with bruce s real data running on muir in pdebug with nodes and processors but also nodes and processors redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject visit hangs during re execution prompted by pick assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description this is a bug bruce hammel claims to have been experiencing with visit for years so i ve set the priority as high now that i have reproducible steps this seems only to occur with nodes multiple processors on single node does not replicate also seems only to occur in conjunction with coordswap operator and with pick var being set to an expression and not the pipeline var on surface ensure parallel engine with nodes open multi silo add pc plot of d add coordswap operator swap x and y coords draw create a scalar expression d p open pick window set variable to d p apply do a zone pick using navigation change the view either by zooming or panning do another zone pick engine will hang must cancel the engine par job in order to interact with visit again information window shows pick wanting to re execute and a merge exception visit does not have all the information it needs to perform a pick please wait while the necessary information is calculated all current pick selections have been cached and will be performed when calculations are complete visit will notify you when it is fully ready for more picks shortly thereafter the following occured pseudocolor invalidmergeexception viewer cannot merge datasets because of an incompatible field and pick mode now fully ready this shows error seeming to come from the viewer but if you run with debug then processors log files will show this error this source should not load balance the data exception invalidmergeexception usr tmp brugger aztec visitbuild src avt pipeline data avtdataattributes c line cannot merge datasets because of an incompatible field and catch visitexception usr tmp brugger aztec visitbuild src engine main executors h this is consistent with what i saw with bruce s real data running on muir in pdebug with nodes and processors but also nodes and processors comments turns out that pick was sending the secondary variable request to only processor which caused an invalid merge exception during pipeline re execution i modified pick to request secondaryvars of all processors svn update trunk m src avt queries pick avtpickquery c
1
673,551
23,018,356,996
IssuesEvent
2022-07-22 00:43:33
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Conversion from bool to float sometimes produces tensor with values 255.0 for True instead of 1.0
high priority triaged module: numpy module: boolean tensor actionable module: correctness (silent)
## 🐛 Bug When I have two seemingly identical bool tensors with values: `tensor([True, True, True, True])`, one of them produces a float tensor with values: `tensor([1.0, 1.0, 1.0, 1.0])`. The other one produces a float tensor with values: `tensor([255., 255., 255., 255.])`. ## To Reproduce Steps to reproduce the behavior: ```python import numpy as np import torch # these are the same data and they look the same when you load them, but they behave subtly differently poisoned = np.frombuffer(b'\xff\xff\xff\xff', dtype=np.bool) # hex 255 clean = np.frombuffer(b'\x01\x01\x01\x01', dtype=np.bool) torch_poisoned = torch.from_numpy(poisoned) torch_clean = torch.from_numpy(clean) print('poisoned == clean -->', (poisoned == clean).all()) # >>> True print('torch_poisoned == torch_clean --> ', (torch_poisoned == torch_clean).all()) # >>> tensor(False) print('torch_poisoned.dtype == torch_clean.dtype --> ', (torch_poisoned.dtype == torch_clean.dtype)) # >>> True print('torch_poisoned.dtype, torch_clean.dtype --> ', torch_poisoned.dtype, torch_clean.dtype) # >>> torch.bool, torch.bool print('torch_clean: ',torch_clean) # >>> tensor([True, True, True, True]) print('torch_clean.float(): ', torch_clean.float()) # >>> tensor([1., 1., 1., 1.]) print('torch_poisoned: ', torch_poisoned) # >>> tensor([True, True, True, True]) print('torch_poisoned.float(): ', torch_poisoned.float()) # >>> tensor([255., 255., 255., 255.]) ``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior Any thing that torch calls a True bool should be converted to a float value of 1.0, not whatever bytes happened to be stored inside the bool. This could just be an abuse of bool, but I came across this bug when using converting a boolean image using PIL. PyTorch should probably prevent users from being able to shoot their foot like this. <!-- A clear and concise description of what you expected to happen. --> ## Environment Collecting environment information... PyTorch version: 1.7.1+cu110 Is debug build: False CUDA used to build PyTorch: 11.0 ROCM used to build PyTorch: N/A OS: Ubuntu 18.04.5 LTS (x86_64) GCC version: (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0 Clang version: Could not collect CMake version: version 3.15.3 Python version: 3.6 (64-bit runtime) Is CUDA available: True CUDA runtime version: 11.2.142 GPU models and configuration: GPU 0: GeForce GTX 1080 Ti Nvidia driver version: 455.45.01 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5 /usr/lib/x86_64-linux-gnu/libcudnn.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.1.1 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.18.5 [pip3] torch==1.7.1+cu110 [pip3] torchaudio==0.7.2 [pip3] torchgeometry==0.1.2 [pip3] torchvision==0.8.2+cu110 [conda] Could not collect ## Additional context <!-- Add any other context about the problem here. --> cc @ezyang @gchanan @zou3519 @mruberry @rgommers @heitorschueroff
1.0
Conversion from bool to float sometimes produces tensor with values 255.0 for True instead of 1.0 - ## 🐛 Bug When I have two seemingly identical bool tensors with values: `tensor([True, True, True, True])`, one of them produces a float tensor with values: `tensor([1.0, 1.0, 1.0, 1.0])`. The other one produces a float tensor with values: `tensor([255., 255., 255., 255.])`. ## To Reproduce Steps to reproduce the behavior: ```python import numpy as np import torch # these are the same data and they look the same when you load them, but they behave subtly differently poisoned = np.frombuffer(b'\xff\xff\xff\xff', dtype=np.bool) # hex 255 clean = np.frombuffer(b'\x01\x01\x01\x01', dtype=np.bool) torch_poisoned = torch.from_numpy(poisoned) torch_clean = torch.from_numpy(clean) print('poisoned == clean -->', (poisoned == clean).all()) # >>> True print('torch_poisoned == torch_clean --> ', (torch_poisoned == torch_clean).all()) # >>> tensor(False) print('torch_poisoned.dtype == torch_clean.dtype --> ', (torch_poisoned.dtype == torch_clean.dtype)) # >>> True print('torch_poisoned.dtype, torch_clean.dtype --> ', torch_poisoned.dtype, torch_clean.dtype) # >>> torch.bool, torch.bool print('torch_clean: ',torch_clean) # >>> tensor([True, True, True, True]) print('torch_clean.float(): ', torch_clean.float()) # >>> tensor([1., 1., 1., 1.]) print('torch_poisoned: ', torch_poisoned) # >>> tensor([True, True, True, True]) print('torch_poisoned.float(): ', torch_poisoned.float()) # >>> tensor([255., 255., 255., 255.]) ``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior Any thing that torch calls a True bool should be converted to a float value of 1.0, not whatever bytes happened to be stored inside the bool. This could just be an abuse of bool, but I came across this bug when using converting a boolean image using PIL. PyTorch should probably prevent users from being able to shoot their foot like this. <!-- A clear and concise description of what you expected to happen. --> ## Environment Collecting environment information... PyTorch version: 1.7.1+cu110 Is debug build: False CUDA used to build PyTorch: 11.0 ROCM used to build PyTorch: N/A OS: Ubuntu 18.04.5 LTS (x86_64) GCC version: (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0 Clang version: Could not collect CMake version: version 3.15.3 Python version: 3.6 (64-bit runtime) Is CUDA available: True CUDA runtime version: 11.2.142 GPU models and configuration: GPU 0: GeForce GTX 1080 Ti Nvidia driver version: 455.45.01 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5 /usr/lib/x86_64-linux-gnu/libcudnn.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.1.1 /usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.1.1 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.18.5 [pip3] torch==1.7.1+cu110 [pip3] torchaudio==0.7.2 [pip3] torchgeometry==0.1.2 [pip3] torchvision==0.8.2+cu110 [conda] Could not collect ## Additional context <!-- Add any other context about the problem here. --> cc @ezyang @gchanan @zou3519 @mruberry @rgommers @heitorschueroff
priority
conversion from bool to float sometimes produces tensor with values for true instead of 🐛 bug when i have two seemingly identical bool tensors with values tensor one of them produces a float tensor with values tensor the other one produces a float tensor with values tensor to reproduce steps to reproduce the behavior python import numpy as np import torch these are the same data and they look the same when you load them but they behave subtly differently poisoned np frombuffer b xff xff xff xff dtype np bool hex clean np frombuffer b dtype np bool torch poisoned torch from numpy poisoned torch clean torch from numpy clean print poisoned clean poisoned clean all true print torch poisoned torch clean torch poisoned torch clean all tensor false print torch poisoned dtype torch clean dtype torch poisoned dtype torch clean dtype true print torch poisoned dtype torch clean dtype torch poisoned dtype torch clean dtype torch bool torch bool print torch clean torch clean tensor print torch clean float torch clean float tensor print torch poisoned torch poisoned tensor print torch poisoned float torch poisoned float tensor expected behavior any thing that torch calls a true bool should be converted to a float value of not whatever bytes happened to be stored inside the bool this could just be an abuse of bool but i came across this bug when using converting a boolean image using pil pytorch should probably prevent users from being able to shoot their foot like this environment collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version python version bit runtime is cuda available true cuda runtime version gpu models and configuration gpu geforce gtx ti nvidia driver version cudnn version probably one of the following usr lib linux gnu libcudnn so usr lib linux gnu libcudnn so usr lib linux gnu libcudnn adv infer so usr lib linux gnu libcudnn adv train so usr lib linux gnu libcudnn cnn infer so usr lib linux gnu libcudnn cnn train so usr lib linux gnu libcudnn ops infer so usr lib linux gnu libcudnn ops train so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy torch torchaudio torchgeometry torchvision could not collect additional context cc ezyang gchanan mruberry rgommers heitorschueroff
1
105,139
4,231,311,731
IssuesEvent
2016-07-04 15:28:27
bromagosa/beetleCloud
https://api.github.com/repos/bromagosa/beetleCloud
closed
"Loading project list" message stays open forever
bug high priority
When you open the projects browser before having saved any projects
1.0
"Loading project list" message stays open forever - When you open the projects browser before having saved any projects
priority
loading project list message stays open forever when you open the projects browser before having saved any projects
1
80,003
3,549,487,259
IssuesEvent
2016-01-20 18:11:57
DDMAL/Rodan
https://api.github.com/repos/DDMAL/Rodan
closed
Coordinate sets for WorkflowJobGroup
Location: Core system Priority: High Type: Feature
We'll either need a new model or edit the existing coordinate set model to take different types of objects. Your call.
1.0
Coordinate sets for WorkflowJobGroup - We'll either need a new model or edit the existing coordinate set model to take different types of objects. Your call.
priority
coordinate sets for workflowjobgroup we ll either need a new model or edit the existing coordinate set model to take different types of objects your call
1
40,286
2,868,385,780
IssuesEvent
2015-06-05 18:31:55
nim-lang/nimble
https://api.github.com/repos/nim-lang/nimble
closed
Empty uninstall
Bug High priority
You can pass an empty string to nimble uninstall. Might want to uninstall the current package if a .nimble is found, similar to nimble install (that's the assumption i had, and that's why i ran this command)? I got scared for a minute after actually pressing "Y" :fearful:. ``` $ nimble uninstall The following packages will be removed: Do you wish to continue? [y/N] ```
1.0
Empty uninstall - You can pass an empty string to nimble uninstall. Might want to uninstall the current package if a .nimble is found, similar to nimble install (that's the assumption i had, and that's why i ran this command)? I got scared for a minute after actually pressing "Y" :fearful:. ``` $ nimble uninstall The following packages will be removed: Do you wish to continue? [y/N] ```
priority
empty uninstall you can pass an empty string to nimble uninstall might want to uninstall the current package if a nimble is found similar to nimble install that s the assumption i had and that s why i ran this command i got scared for a minute after actually pressing y fearful nimble uninstall the following packages will be removed do you wish to continue
1
591,973
17,866,993,361
IssuesEvent
2021-09-06 10:39:51
AbsaOSS/enceladus
https://api.github.com/repos/AbsaOSS/enceladus
closed
Authenticated using keytab, configuring Kerberos on the EMR master node
feature security priority: high cloud
## Background Keytab is used for authentication within Enceladus ## Feature Set-up and verify keytab authentication in EMR
1.0
Authenticated using keytab, configuring Kerberos on the EMR master node - ## Background Keytab is used for authentication within Enceladus ## Feature Set-up and verify keytab authentication in EMR
priority
authenticated using keytab configuring kerberos on the emr master node background keytab is used for authentication within enceladus feature set up and verify keytab authentication in emr
1
588,976
17,687,070,986
IssuesEvent
2021-08-24 04:07:06
woowa-techcamp-2021/store-6
https://api.github.com/repos/woowa-techcamp-2021/store-6
closed
[FE] 헤더 컴포넌트 가장 기본 UI 구현
high priority feat
## :hammer: 기능 설명 드랍다운, 장바구니 API 연동 등을 제외한 가장 기본적인 헤더의 UI를 구현합니다. ## 📑 완료 조건 - [x] 각 버튼/링크를 클릭했을 때 해당 페이지로 이동해야 합니다. - [x] 로그인이 되었을 경우 사용자의 프로필이미지를 기존 아이콘 대신 보여줘야 합니다. ## :thought_balloon: 관련 Backlog > [대분류] - [중분류] - [Backlog 이름] - [FE] 헤더 - 헤더 컴포넌트 - 리팩토링(공통 컴포넌트 분리 등) - [FE] 헤더 - 헤더 컴포넌트 - 로고 클릭 시 홈으로 이동 - [FE] 헤더 - 헤더 컴포넌트 - 장바구니, 좋아요, 마이페이지 버튼 클릭 시 페이지 이동 - [FE] 기타 - 공통 컴포넌트 - SEO를 위한 로고 개선
1.0
[FE] 헤더 컴포넌트 가장 기본 UI 구현 - ## :hammer: 기능 설명 드랍다운, 장바구니 API 연동 등을 제외한 가장 기본적인 헤더의 UI를 구현합니다. ## 📑 완료 조건 - [x] 각 버튼/링크를 클릭했을 때 해당 페이지로 이동해야 합니다. - [x] 로그인이 되었을 경우 사용자의 프로필이미지를 기존 아이콘 대신 보여줘야 합니다. ## :thought_balloon: 관련 Backlog > [대분류] - [중분류] - [Backlog 이름] - [FE] 헤더 - 헤더 컴포넌트 - 리팩토링(공통 컴포넌트 분리 등) - [FE] 헤더 - 헤더 컴포넌트 - 로고 클릭 시 홈으로 이동 - [FE] 헤더 - 헤더 컴포넌트 - 장바구니, 좋아요, 마이페이지 버튼 클릭 시 페이지 이동 - [FE] 기타 - 공통 컴포넌트 - SEO를 위한 로고 개선
priority
헤더 컴포넌트 가장 기본 ui 구현 hammer 기능 설명 드랍다운 장바구니 api 연동 등을 제외한 가장 기본적인 헤더의 ui를 구현합니다 📑 완료 조건 각 버튼 링크를 클릭했을 때 해당 페이지로 이동해야 합니다 로그인이 되었을 경우 사용자의 프로필이미지를 기존 아이콘 대신 보여줘야 합니다 thought balloon 관련 backlog 헤더 헤더 컴포넌트 리팩토링 공통 컴포넌트 분리 등 헤더 헤더 컴포넌트 로고 클릭 시 홈으로 이동 헤더 헤더 컴포넌트 장바구니 좋아요 마이페이지 버튼 클릭 시 페이지 이동 기타 공통 컴포넌트 seo를 위한 로고 개선
1
789,048
27,776,805,147
IssuesEvent
2023-03-16 17:48:09
WFP-VAM/prism-app
https://api.github.com/repos/WFP-VAM/prism-app
closed
[Bug]: Kobo get dates request doesn't implement filters
bug priority:high triage
### What happened? The timeline interface shows that there's data available on a given date, but when I change to that date, I get an error message saying there's no data. I can see that the initial request for available dates adds the filter attribute as expected, but the results do not apply the filter, so dates are being returned when they should not. See this query: https://prism-api.ovio.org/kobo/dates?formId=aMLBShjyyhYeS5ERuRy8s3&datetimeField=Date_report&measureField=NumPeoAff&koboUrl=https://kobo.humanitarianresponse.info/api/v2/assets/&filters=status=Approved As an example, '2023-01-16' is returned as a valid date. If I select that date in the app, the request correctly adds the filter parameter and (correctly) returns an empty dataset. Here's the request: https://prism-api.ovio.org/kobo/forms?beginDateTime=2023-01-16&endDateTime=2023-01-16&formId=aMLBShjyyhYeS5ERuRy8s3&datetimeField=Date_report&measureField=NumPeoAff&koboUrl=https://kobo.humanitarianresponse.info/api/v2/assets/&filters=status=Approved I can see in [api/app/kobo.py ](https://github.com/WFP-VAM/prism-app/blob/master/api/app/kobo.py)get_form_dates() doesn't make use of the filter, but get_form_responses() does. ### Which country / deployment are you running? Cambodia on PR https://github.com/WFP-VAM/prism-app/pull/688 ### Add a screenshot (if relevant) https://prism-khm-new-ux.surge.sh/?hazardLayerIds=flood_report&date=2023-01-16 <img width="1520" alt="Screenshot 2023-03-15 at 22 45 06" src="https://user-images.githubusercontent.com/3343536/225529229-c47bfbe8-8065-48c0-8bfb-6bf5fc50b145.png"> Response from the date request: <img width="355" alt="Screenshot 2023-03-15 at 23 06 23" src="https://user-images.githubusercontent.com/3343536/225529617-fa3b1d5f-f08a-421d-9c6d-efb1c53ea0fc.png">
1.0
[Bug]: Kobo get dates request doesn't implement filters - ### What happened? The timeline interface shows that there's data available on a given date, but when I change to that date, I get an error message saying there's no data. I can see that the initial request for available dates adds the filter attribute as expected, but the results do not apply the filter, so dates are being returned when they should not. See this query: https://prism-api.ovio.org/kobo/dates?formId=aMLBShjyyhYeS5ERuRy8s3&datetimeField=Date_report&measureField=NumPeoAff&koboUrl=https://kobo.humanitarianresponse.info/api/v2/assets/&filters=status=Approved As an example, '2023-01-16' is returned as a valid date. If I select that date in the app, the request correctly adds the filter parameter and (correctly) returns an empty dataset. Here's the request: https://prism-api.ovio.org/kobo/forms?beginDateTime=2023-01-16&endDateTime=2023-01-16&formId=aMLBShjyyhYeS5ERuRy8s3&datetimeField=Date_report&measureField=NumPeoAff&koboUrl=https://kobo.humanitarianresponse.info/api/v2/assets/&filters=status=Approved I can see in [api/app/kobo.py ](https://github.com/WFP-VAM/prism-app/blob/master/api/app/kobo.py)get_form_dates() doesn't make use of the filter, but get_form_responses() does. ### Which country / deployment are you running? Cambodia on PR https://github.com/WFP-VAM/prism-app/pull/688 ### Add a screenshot (if relevant) https://prism-khm-new-ux.surge.sh/?hazardLayerIds=flood_report&date=2023-01-16 <img width="1520" alt="Screenshot 2023-03-15 at 22 45 06" src="https://user-images.githubusercontent.com/3343536/225529229-c47bfbe8-8065-48c0-8bfb-6bf5fc50b145.png"> Response from the date request: <img width="355" alt="Screenshot 2023-03-15 at 23 06 23" src="https://user-images.githubusercontent.com/3343536/225529617-fa3b1d5f-f08a-421d-9c6d-efb1c53ea0fc.png">
priority
kobo get dates request doesn t implement filters what happened the timeline interface shows that there s data available on a given date but when i change to that date i get an error message saying there s no data i can see that the initial request for available dates adds the filter attribute as expected but the results do not apply the filter so dates are being returned when they should not see this query as an example is returned as a valid date if i select that date in the app the request correctly adds the filter parameter and correctly returns an empty dataset here s the request i can see in doesn t make use of the filter but get form responses does which country deployment are you running cambodia on pr add a screenshot if relevant img width alt screenshot at src response from the date request img width alt screenshot at src
1
136,510
5,284,090,531
IssuesEvent
2017-02-07 23:09:28
pymedusa/Medusa
https://api.github.com/repos/pymedusa/Medusa
closed
[APP SUBMITTED]: TypeError: unhashable type: 'list'
Bug Confirmed Priority: 1. High
### INFO **Python Version**: `2.7.12 (v2.7.12:d33e0cf91556, Jun 27 2016, 15:24:40) [MSC v.1500 64 bit (AMD64)]` **Operating System**: `Windows-10-10.0.14393` **Locale**: `cp1252` **Branch**: [develop](../tree/develop) **Database**: `44.4` **Commit**: pymedusa/Medusa@63f66cdd827c391aa500986cd2b1e3b8276f858e **Link to Log**: https://gist.github.com/3ac6567533f759fba5d4d3288d93eb69 ### ERROR <pre> 2017-02-07 15:57:36 ERROR Thread-37 :: [63f66cd] Exception generated: unhashable type: 'list' Traceback (most recent call last): File "C:\Medusa\medusa\server\web\core\base.py", line 259, in async_call result = function(**kwargs) File "C:\Medusa\medusa\server\web\home\handler.py", line 1431, in editShow exceptions = {exceptions} TypeError: unhashable type: 'list' </pre> --- _STAFF NOTIFIED_: @pymedusa/support @pymedusa/moderators
1.0
[APP SUBMITTED]: TypeError: unhashable type: 'list' - ### INFO **Python Version**: `2.7.12 (v2.7.12:d33e0cf91556, Jun 27 2016, 15:24:40) [MSC v.1500 64 bit (AMD64)]` **Operating System**: `Windows-10-10.0.14393` **Locale**: `cp1252` **Branch**: [develop](../tree/develop) **Database**: `44.4` **Commit**: pymedusa/Medusa@63f66cdd827c391aa500986cd2b1e3b8276f858e **Link to Log**: https://gist.github.com/3ac6567533f759fba5d4d3288d93eb69 ### ERROR <pre> 2017-02-07 15:57:36 ERROR Thread-37 :: [63f66cd] Exception generated: unhashable type: 'list' Traceback (most recent call last): File "C:\Medusa\medusa\server\web\core\base.py", line 259, in async_call result = function(**kwargs) File "C:\Medusa\medusa\server\web\home\handler.py", line 1431, in editShow exceptions = {exceptions} TypeError: unhashable type: 'list' </pre> --- _STAFF NOTIFIED_: @pymedusa/support @pymedusa/moderators
priority
typeerror unhashable type list info python version jun operating system windows locale branch tree develop database commit pymedusa medusa link to log error error thread exception generated unhashable type list traceback most recent call last file c medusa medusa server web core base py line in async call result function kwargs file c medusa medusa server web home handler py line in editshow exceptions exceptions typeerror unhashable type list staff notified pymedusa support pymedusa moderators
1
769,894
27,021,546,756
IssuesEvent
2023-02-11 03:42:22
codersforcauses/poops
https://api.github.com/repos/codersforcauses/poops
opened
update admin email domain in firebase functions
backend bug enhancement difficulty::easy priority::high point::1
**Is your feature request related to a problem? Please describe.** The current admin email domain is `poop.org` which is incorrect. **Describe the solution you'd like** Change the domain name in `addAdmin.ts` firebase functions to `poopswa.org.au` . **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
update admin email domain in firebase functions - **Is your feature request related to a problem? Please describe.** The current admin email domain is `poop.org` which is incorrect. **Describe the solution you'd like** Change the domain name in `addAdmin.ts` firebase functions to `poopswa.org.au` . **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
priority
update admin email domain in firebase functions is your feature request related to a problem please describe the current admin email domain is poop org which is incorrect describe the solution you d like change the domain name in addadmin ts firebase functions to poopswa org au describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
1
133,559
5,205,368,774
IssuesEvent
2017-01-24 17:45:39
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
opened
Need to Add Ground Truth Object ID to Depth Sensor
priority: high team: sensors type: feature request
For every depth measurement, it'll be nice to include "ground truth" in terms of identifying what object was measured. This ground truth should ideally take the form of an integer ID, which could then be translated into something meaningful like "car", "bicycle", and "pedestrian".
1.0
Need to Add Ground Truth Object ID to Depth Sensor - For every depth measurement, it'll be nice to include "ground truth" in terms of identifying what object was measured. This ground truth should ideally take the form of an integer ID, which could then be translated into something meaningful like "car", "bicycle", and "pedestrian".
priority
need to add ground truth object id to depth sensor for every depth measurement it ll be nice to include ground truth in terms of identifying what object was measured this ground truth should ideally take the form of an integer id which could then be translated into something meaningful like car bicycle and pedestrian
1
230,024
7,603,426,823
IssuesEvent
2018-04-29 14:27:39
LazyFoxStudio/Project-2
https://api.github.com/repos/LazyFoxStudio/Project-2
opened
Camera bugs (black screen)
Gameplay bug high priority
When you move the camera to the top or left of the map. The screen turns black
1.0
Camera bugs (black screen) - When you move the camera to the top or left of the map. The screen turns black
priority
camera bugs black screen when you move the camera to the top or left of the map the screen turns black
1
338,191
10,225,518,273
IssuesEvent
2019-08-16 15:20:37
GeekyAnts/NativeBase
https://api.github.com/repos/GeekyAnts/NativeBase
closed
v2.12.2 refers to FontAwesome5Pro
high priority
v2.12.2 imports react-native-vector-icons/FontAwesome5Pro which breaks on Expo (probably all Expo versions) where package @expo/vector-icons override react-native-vector-icons. v2.12.1 which has no FontAwesome5Pro import (only FontAwesome5) works. Comment in other issue: https://github.com/GeekyAnts/NativeBase/issues/2731#issuecomment-512796306_ Quick solution: downgrade to v2.12.1.
1.0
v2.12.2 refers to FontAwesome5Pro - v2.12.2 imports react-native-vector-icons/FontAwesome5Pro which breaks on Expo (probably all Expo versions) where package @expo/vector-icons override react-native-vector-icons. v2.12.1 which has no FontAwesome5Pro import (only FontAwesome5) works. Comment in other issue: https://github.com/GeekyAnts/NativeBase/issues/2731#issuecomment-512796306_ Quick solution: downgrade to v2.12.1.
priority
refers to imports react native vector icons which breaks on expo probably all expo versions where package expo vector icons override react native vector icons which has no import only works comment in other issue quick solution downgrade to
1
327,685
9,978,951,100
IssuesEvent
2019-07-09 21:15:22
Baystation12/Baystation12
https://api.github.com/repos/Baystation12/Baystation12
closed
RP-Rev recruiting is not working.
Priority: High ⚠
#### Description of issue Recruiting a revolutioner is not working #### Difference between expected and actual behavior Being recruted and pushing "Yes" button should make you a rev. It doesn't. #### Steps to reproduce >play RP-Rev round >accept the proposition to join revolution #### Specific information for locating It happened with all of headrevs and all of people who tried to accept the proposition. #### Length of time in which bug has been known to occur I've only spotted this during one round because it was first time me playing rev and taking part in conversion. #### Client version, Server revision & Game ID Client Version: 510 Server Revision: 1f54f8a2655ed900b64da48a0d24b1dea24d3129 - dev - Game ID: bMS-cJu5 Current map: SEV Torch #### Issue bingo Please check whatever applies. More checkboxes checked increase your chances of the issue being looked at sooner. - [X] Issue could be reproduced at least once - [X] Issue could be reproduced by different players - [?] Issue could be reproduced in multiple rounds - [X] Issue happened in a recent (less than 7 days ago) round - [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
1.0
RP-Rev recruiting is not working. - #### Description of issue Recruiting a revolutioner is not working #### Difference between expected and actual behavior Being recruted and pushing "Yes" button should make you a rev. It doesn't. #### Steps to reproduce >play RP-Rev round >accept the proposition to join revolution #### Specific information for locating It happened with all of headrevs and all of people who tried to accept the proposition. #### Length of time in which bug has been known to occur I've only spotted this during one round because it was first time me playing rev and taking part in conversion. #### Client version, Server revision & Game ID Client Version: 510 Server Revision: 1f54f8a2655ed900b64da48a0d24b1dea24d3129 - dev - Game ID: bMS-cJu5 Current map: SEV Torch #### Issue bingo Please check whatever applies. More checkboxes checked increase your chances of the issue being looked at sooner. - [X] Issue could be reproduced at least once - [X] Issue could be reproduced by different players - [?] Issue could be reproduced in multiple rounds - [X] Issue happened in a recent (less than 7 days ago) round - [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
priority
rp rev recruiting is not working description of issue recruiting a revolutioner is not working difference between expected and actual behavior being recruted and pushing yes button should make you a rev it doesn t steps to reproduce play rp rev round accept the proposition to join revolution specific information for locating it happened with all of headrevs and all of people who tried to accept the proposition length of time in which bug has been known to occur i ve only spotted this during one round because it was first time me playing rev and taking part in conversion client version server revision game id client version server revision dev game id bms current map sev torch issue bingo please check whatever applies more checkboxes checked increase your chances of the issue being looked at sooner issue could be reproduced at least once issue could be reproduced by different players issue could be reproduced in multiple rounds issue happened in a recent less than days ago round
1
282,213
8,704,534,204
IssuesEvent
2018-12-05 19:40:41
AICrowd/AIcrowd
https://api.github.com/repos/AICrowd/AIcrowd
closed
Add GDPR button
feature high priority
_From @seanfcarroll on May 14, 2018 07:45_ _Copied from original issue: crowdAI/crowdai#784_
1.0
Add GDPR button - _From @seanfcarroll on May 14, 2018 07:45_ _Copied from original issue: crowdAI/crowdai#784_
priority
add gdpr button from seanfcarroll on may copied from original issue crowdai crowdai
1
436,957
12,556,700,456
IssuesEvent
2020-06-07 10:36:57
modernweb-dev/web
https://api.github.com/repos/modernweb-dev/web
opened
[test-runner-*] dedupe syntax errors
enhancement priority: high
Syntax errors are now logged by es-dev-server, so they are not deduped cross browsers. We should pull this logic into WTR and dedupe them.
1.0
[test-runner-*] dedupe syntax errors - Syntax errors are now logged by es-dev-server, so they are not deduped cross browsers. We should pull this logic into WTR and dedupe them.
priority
dedupe syntax errors syntax errors are now logged by es dev server so they are not deduped cross browsers we should pull this logic into wtr and dedupe them
1
608,657
18,845,243,241
IssuesEvent
2021-11-11 14:17:43
boostcampwm-2021/iOS05-Escaper
https://api.github.com/repos/boostcampwm-2021/iOS05-Escaper
closed
탭바, 네비게이션바 세팅 및 iOS13 위치 정보 인증 개선
feature High Priority
- 탭바, 네비게이션바 iOS 13,14 와 15가 backgroud.color가 다름. - 네비게이션바 iOS15에서 스크롤 시 backgroud.color가 변경됨 - 유저 위치 정보 인증 - iOS13에서 첫 접속시 delegate 함수를 진행하지 않음
1.0
탭바, 네비게이션바 세팅 및 iOS13 위치 정보 인증 개선 - - 탭바, 네비게이션바 iOS 13,14 와 15가 backgroud.color가 다름. - 네비게이션바 iOS15에서 스크롤 시 backgroud.color가 변경됨 - 유저 위치 정보 인증 - iOS13에서 첫 접속시 delegate 함수를 진행하지 않음
priority
탭바 네비게이션바 세팅 및 위치 정보 인증 개선 탭바 네비게이션바 ios 와 backgroud color가 다름 네비게이션바 스크롤 시 backgroud color가 변경됨 유저 위치 정보 인증 첫 접속시 delegate 함수를 진행하지 않음
1
375,947
11,136,336,383
IssuesEvent
2019-12-20 16:21:54
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
We should test __cuda_array_interface__ stride handling
high priority module: cuda triage review
https://github.com/pytorch/pytorch/pull/24947 fixed a bug around how PyTorch handles contiguous tensor strides in __cuda_array_interface__, but didn't add a test for it. We should add a test for it. cc @ezyang @gchanan @zou3519 @ngimel
1.0
We should test __cuda_array_interface__ stride handling - https://github.com/pytorch/pytorch/pull/24947 fixed a bug around how PyTorch handles contiguous tensor strides in __cuda_array_interface__, but didn't add a test for it. We should add a test for it. cc @ezyang @gchanan @zou3519 @ngimel
priority
we should test cuda array interface stride handling fixed a bug around how pytorch handles contiguous tensor strides in cuda array interface but didn t add a test for it we should add a test for it cc ezyang gchanan ngimel
1
302,046
9,254,537,614
IssuesEvent
2019-03-16 00:21:28
ESGF/esgf-compute-wps
https://api.github.com/repos/ESGF/esgf-compute-wps
closed
Need to fix the webapp wps execute
area/frontend kind/bug priority/high size/small
**Describe the bug** The Execute request generated by the webapp is not properly formatted, it's using LiteralData when the service is expecting ComplexData. **To Reproduce** Submit a job using the webapp. **Expected behavior** Should successfully submit a properly formatted Execute POST request. **Screenshots** n/a **Additional context** n/a
1.0
Need to fix the webapp wps execute - **Describe the bug** The Execute request generated by the webapp is not properly formatted, it's using LiteralData when the service is expecting ComplexData. **To Reproduce** Submit a job using the webapp. **Expected behavior** Should successfully submit a properly formatted Execute POST request. **Screenshots** n/a **Additional context** n/a
priority
need to fix the webapp wps execute describe the bug the execute request generated by the webapp is not properly formatted it s using literaldata when the service is expecting complexdata to reproduce submit a job using the webapp expected behavior should successfully submit a properly formatted execute post request screenshots n a additional context n a
1
240,930
7,807,424,065
IssuesEvent
2018-06-11 16:48:36
ministryofjustice/cloud-platform-roadmap
https://api.github.com/repos/ministryofjustice/cloud-platform-roadmap
opened
Monitoring capacity of cluster
high-priority
## Background As we add more things to the cluster we need to know when to add more nodes. The first metric to do this on the back of will be the amount of memory/cpu available on the cluster overall. We need to be able to: - measure that metric, - alert when it is nearing capacity and - have a process for adding more resources when we are near capacity ## Proposed user journey - Cloud platform team member gets alert that cluster is at e.g. 80% capacity for CPU and Memory - Team member goes to a run book - Team member provisions a new node - Alert stops firing as the new node comes into service and overall capacity increases ## Approach ## Questions / Assumptions <!-- Additional information to explain approach taken --> ## Definition of done <!-- Checklist for definition of done and acceptance criteria, for example: --> - [ ] must compile - [ ] must pass tests - [ ] must address all the steps of user journey ## Reference [How to write good user stories](https://www.gov.uk/service-manual/agile-delivery/writing-user-stories)
1.0
Monitoring capacity of cluster - ## Background As we add more things to the cluster we need to know when to add more nodes. The first metric to do this on the back of will be the amount of memory/cpu available on the cluster overall. We need to be able to: - measure that metric, - alert when it is nearing capacity and - have a process for adding more resources when we are near capacity ## Proposed user journey - Cloud platform team member gets alert that cluster is at e.g. 80% capacity for CPU and Memory - Team member goes to a run book - Team member provisions a new node - Alert stops firing as the new node comes into service and overall capacity increases ## Approach ## Questions / Assumptions <!-- Additional information to explain approach taken --> ## Definition of done <!-- Checklist for definition of done and acceptance criteria, for example: --> - [ ] must compile - [ ] must pass tests - [ ] must address all the steps of user journey ## Reference [How to write good user stories](https://www.gov.uk/service-manual/agile-delivery/writing-user-stories)
priority
monitoring capacity of cluster background as we add more things to the cluster we need to know when to add more nodes the first metric to do this on the back of will be the amount of memory cpu available on the cluster overall we need to be able to measure that metric alert when it is nearing capacity and have a process for adding more resources when we are near capacity proposed user journey cloud platform team member gets alert that cluster is at e g capacity for cpu and memory team member goes to a run book team member provisions a new node alert stops firing as the new node comes into service and overall capacity increases approach questions assumptions definition of done must compile must pass tests must address all the steps of user journey reference
1
179,250
6,622,878,820
IssuesEvent
2017-09-22 03:02:10
minio/mint
https://api.github.com/repos/minio/mint
closed
failure in aws-sdk-ruby
fixed priority: high
aws-sdk-ruby fails with the log ``` {"name":"aws-sdk-ruby","function":"listBuckets()","args":{},"duration":31.82,"status":"PASS"} {"name":"aws-sdk-ruby","function":"removeObjects(bucket_name)","args":{"bucket_name":"4ed81b3f25bc"},"duration":173.16,"error":"Failed to clean-up bucket '4ed81b3f25bc': undefined method `merge' for \"datafile-1-MB\":String","status":"FAIL"} ```
1.0
failure in aws-sdk-ruby - aws-sdk-ruby fails with the log ``` {"name":"aws-sdk-ruby","function":"listBuckets()","args":{},"duration":31.82,"status":"PASS"} {"name":"aws-sdk-ruby","function":"removeObjects(bucket_name)","args":{"bucket_name":"4ed81b3f25bc"},"duration":173.16,"error":"Failed to clean-up bucket '4ed81b3f25bc': undefined method `merge' for \"datafile-1-MB\":String","status":"FAIL"} ```
priority
failure in aws sdk ruby aws sdk ruby fails with the log name aws sdk ruby function listbuckets args duration status pass name aws sdk ruby function removeobjects bucket name args bucket name duration error failed to clean up bucket undefined method merge for datafile mb string status fail
1
274,002
8,555,996,603
IssuesEvent
2018-11-08 11:44:51
BeamMW/beam
https://api.github.com/repos/BeamMW/beam
opened
Failed transactions on local node
High priority bug
Precon: wallet 1 is connected to local node wallet 2 is connected to testnet node Steps: 1. Run wallet 1 2. While it's syncronyzing send beams from wallet 2 to wallet 1 3. Wait for some minutes Actual: transaction is changed from Pending to Failed 4. Try to send one more transaction Actual: transaction is failed [wallet_18_11_08_13_30_59.log](https://github.com/BeamMW/beam/files/2561547/wallet_18_11_08_13_30_59.log) Transactions are also failed when node is synchronized
1.0
Failed transactions on local node - Precon: wallet 1 is connected to local node wallet 2 is connected to testnet node Steps: 1. Run wallet 1 2. While it's syncronyzing send beams from wallet 2 to wallet 1 3. Wait for some minutes Actual: transaction is changed from Pending to Failed 4. Try to send one more transaction Actual: transaction is failed [wallet_18_11_08_13_30_59.log](https://github.com/BeamMW/beam/files/2561547/wallet_18_11_08_13_30_59.log) Transactions are also failed when node is synchronized
priority
failed transactions on local node precon wallet is connected to local node wallet is connected to testnet node steps run wallet while it s syncronyzing send beams from wallet to wallet wait for some minutes actual transaction is changed from pending to failed try to send one more transaction actual transaction is failed transactions are also failed when node is synchronized
1
810,810
30,262,060,188
IssuesEvent
2023-07-07 08:57:54
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
[Bug]: Goto Definition not working for regexp lang lib
Type/Bug Priority/High Status/Blocked Team/LanguageServer Area/Definition
### Description ![regexp](https://user-images.githubusercontent.com/35211477/214019490-768eac3d-fab5-4035-9424-451165e75899.gif) ### Steps to Reproduce ```ballerina import ballerina/lang.regexp; public function main() { regexp:RegExp reg = re ``; } ``` ### Affected Version(s) 2201.4.0-SNAPSHOT ### OS, DB, other environment details and versions _No response_ ### Related area -> Editor ### Related issue(s) (optional) _No response_ ### Suggested label(s) (optional) _No response_ ### Suggested assignee(s) (optional) _No response_
1.0
[Bug]: Goto Definition not working for regexp lang lib - ### Description ![regexp](https://user-images.githubusercontent.com/35211477/214019490-768eac3d-fab5-4035-9424-451165e75899.gif) ### Steps to Reproduce ```ballerina import ballerina/lang.regexp; public function main() { regexp:RegExp reg = re ``; } ``` ### Affected Version(s) 2201.4.0-SNAPSHOT ### OS, DB, other environment details and versions _No response_ ### Related area -> Editor ### Related issue(s) (optional) _No response_ ### Suggested label(s) (optional) _No response_ ### Suggested assignee(s) (optional) _No response_
priority
goto definition not working for regexp lang lib description steps to reproduce ballerina import ballerina lang regexp public function main regexp regexp reg re affected version s snapshot os db other environment details and versions no response related area editor related issue s optional no response suggested label s optional no response suggested assignee s optional no response
1
296,180
9,105,227,529
IssuesEvent
2019-02-20 20:13:00
ganeti/ganeti
https://api.github.com/repos/ganeti/ganeti
closed
QEMU 3.1 compatibility
Compatibility Priority:High
Debian Buster will be shipping with QEMU 3.1. Currently KVM instances fail to start because we're using `-balloon virtio`, which was removed in 3.1: ``` $ qemu-system-x86_64 -balloon virtio qemu-system-x86_64: -balloon: invalid option $ qemu-system-x86_64 --version QEMU emulator version 3.1.0 (Debian 1:3.1+dfsg-2+b1) Copyright (c) 2003-2018 Fabrice Bellard and the QEMU Project developers ``` We need to make sure Ganeti works with QEMU 3.1, if possible without using any deprecated features.
1.0
QEMU 3.1 compatibility - Debian Buster will be shipping with QEMU 3.1. Currently KVM instances fail to start because we're using `-balloon virtio`, which was removed in 3.1: ``` $ qemu-system-x86_64 -balloon virtio qemu-system-x86_64: -balloon: invalid option $ qemu-system-x86_64 --version QEMU emulator version 3.1.0 (Debian 1:3.1+dfsg-2+b1) Copyright (c) 2003-2018 Fabrice Bellard and the QEMU Project developers ``` We need to make sure Ganeti works with QEMU 3.1, if possible without using any deprecated features.
priority
qemu compatibility debian buster will be shipping with qemu currently kvm instances fail to start because we re using balloon virtio which was removed in qemu system balloon virtio qemu system balloon invalid option qemu system version qemu emulator version debian dfsg copyright c fabrice bellard and the qemu project developers we need to make sure ganeti works with qemu if possible without using any deprecated features
1
404,008
11,850,872,683
IssuesEvent
2020-03-24 17:15:56
BlocKoopa/KnightsCapstone
https://api.github.com/repos/BlocKoopa/KnightsCapstone
opened
Programming - Switches & Gates
Point: 5 Priority: Highest
2 parts: 1. Program Gates to block player. (Program Tiles to recognize if a gate is in the way, and block player). 2. Program Switches to remove corresponding Gates when landed on.
1.0
Programming - Switches & Gates - 2 parts: 1. Program Gates to block player. (Program Tiles to recognize if a gate is in the way, and block player). 2. Program Switches to remove corresponding Gates when landed on.
priority
programming switches gates parts program gates to block player program tiles to recognize if a gate is in the way and block player program switches to remove corresponding gates when landed on
1
614,822
19,190,335,074
IssuesEvent
2021-12-05 22:01:05
RE-SS3D/SS3D
https://api.github.com/repos/RE-SS3D/SS3D
opened
Implement Wire Adjacency Connections
Type: Feature (Addition) Asset: Script Coding: C# Priority: 2 - High Difficulty: 2 - Medium System: Tilemaps
<!-- The notes within these arrows are for you but can be deleted. --> ## Summary Implement a new tilemap adjacency connection script (similar to the others) for "wire connections". This should follow the design located in the link below. ## Goal This will allow for wires to be added to the map via the editor and perform intended connections. https://docs.google.com/document/d/1ful7_gIJo7e74i9LMQuYMpjZMH2V1aT90mwlrQ0hcgE/#heading=h.15jb46xsi8l2
1.0
Implement Wire Adjacency Connections - <!-- The notes within these arrows are for you but can be deleted. --> ## Summary Implement a new tilemap adjacency connection script (similar to the others) for "wire connections". This should follow the design located in the link below. ## Goal This will allow for wires to be added to the map via the editor and perform intended connections. https://docs.google.com/document/d/1ful7_gIJo7e74i9LMQuYMpjZMH2V1aT90mwlrQ0hcgE/#heading=h.15jb46xsi8l2
priority
implement wire adjacency connections summary implement a new tilemap adjacency connection script similar to the others for wire connections this should follow the design located in the link below goal this will allow for wires to be added to the map via the editor and perform intended connections
1
424,250
12,308,298,950
IssuesEvent
2020-05-12 06:55:51
GrassrootsEconomics/CIC-Docs
https://api.github.com/repos/GrassrootsEconomics/CIC-Docs
opened
From and To on reciept messages
enhancement priority:high
**Is your feature request related to a problem? Please describe.** we have old ladies swapping sim cards and they are VERY confused .... because on the same phone from 'Sarafu' they can't tell whom sent what to whom and they are swapping sim cards not all users have phones but they do have a sim **Describe the solution you'd like** we need on the receipt message to say both the sender and receiver... Sent receipt ```Successfully sent a payment of 1000.00 SARAFU to Lydia Wanjiku +254720305812 on 11/05/2020 at 09:45 PM from Jim Wanger +254720303811. New SARAFU balance is 1236.37 ``` received receipt ```Successfully received a payment of 1000.00 SARAFU from Wanjiru Watheka +254702875427 on 11/05/2020 at 09:45 PM to Sayy Hornsby +254720303822. New SARAFU balance is 1809.01```
1.0
From and To on reciept messages - **Is your feature request related to a problem? Please describe.** we have old ladies swapping sim cards and they are VERY confused .... because on the same phone from 'Sarafu' they can't tell whom sent what to whom and they are swapping sim cards not all users have phones but they do have a sim **Describe the solution you'd like** we need on the receipt message to say both the sender and receiver... Sent receipt ```Successfully sent a payment of 1000.00 SARAFU to Lydia Wanjiku +254720305812 on 11/05/2020 at 09:45 PM from Jim Wanger +254720303811. New SARAFU balance is 1236.37 ``` received receipt ```Successfully received a payment of 1000.00 SARAFU from Wanjiru Watheka +254702875427 on 11/05/2020 at 09:45 PM to Sayy Hornsby +254720303822. New SARAFU balance is 1809.01```
priority
from and to on reciept messages is your feature request related to a problem please describe we have old ladies swapping sim cards and they are very confused because on the same phone from sarafu they can t tell whom sent what to whom and they are swapping sim cards not all users have phones but they do have a sim describe the solution you d like we need on the receipt message to say both the sender and receiver sent receipt successfully sent a payment of sarafu to lydia wanjiku on at pm from jim wanger new sarafu balance is received receipt successfully received a payment of sarafu from wanjiru watheka on at pm to sayy hornsby new sarafu balance is
1
54,117
3,060,050,758
IssuesEvent
2015-08-14 18:24:55
loklak/loklak_webclient
https://api.github.com/repos/loklak/loklak_webclient
closed
wall not working on loklak.net
bug Priority 1 - High Twitter Wall - Aneesh
After I moved test.loklak.net to loklak.net, the wall is not working any more. The javascript terminal shows the following error: ``` "Error: [$rootScope:inprog] $apply already in progress http://errors.angularjs.org/1.3.16/$rootScope/inprog?p0=%24apply i/<@http://loklak.net/js/main.js:30:1177 f@http://loklak.net/js/main.js:31:30933 Ht/this.$get</h.prototype.$apply@http://loklak.net/js/main.js:33:2968 ["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</<@http://loklak.net/js/main.js:35:18841 jQuery.event.dispatch@http://code.jquery.com/jquery-2.1.4.js:4434:15 jQuery.event.add/elemData.handle@http://code.jquery.com/jquery-2.1.4.js:4121:6 jQuery.event.trigger@http://code.jquery.com/jquery-2.1.4.js:4350:5 .trigger/<@http://code.jquery.com/jquery-2.1.4.js:4901:4 .each@http://code.jquery.com/jquery-2.1.4.js:374:14 jQuery.prototype.each@http://code.jquery.com/jquery-2.1.4.js:139:10 .trigger@http://code.jquery.com/jquery-2.1.4.js:4900:1 i/e.proceed@http://loklak.net/js/main.js:22:16928 ["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</va.prototype.functionCall/<@http://loklak.net/js/main.js:35:12265 ["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</</a@http://loklak.net/js/main.js:35:18791 Ht/this.$get</h.prototype.$eval@http://loklak.net/js/main.js:33:2755 Ht/this.$get</h.prototype.$apply@http://loklak.net/js/main.js:33:2980 ["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</<@http://loklak.net/js/main.js:35:18841 jQuery.event.dispatch@http://code.jquery.com/jquery-2.1.4.js:4434:15 jQuery.event.add/elemData.handle@http://code.jquery.com/jquery-2.1.4.js:4121:6 " ```
1.0
wall not working on loklak.net - After I moved test.loklak.net to loklak.net, the wall is not working any more. The javascript terminal shows the following error: ``` "Error: [$rootScope:inprog] $apply already in progress http://errors.angularjs.org/1.3.16/$rootScope/inprog?p0=%24apply i/<@http://loklak.net/js/main.js:30:1177 f@http://loklak.net/js/main.js:31:30933 Ht/this.$get</h.prototype.$apply@http://loklak.net/js/main.js:33:2968 ["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</<@http://loklak.net/js/main.js:35:18841 jQuery.event.dispatch@http://code.jquery.com/jquery-2.1.4.js:4434:15 jQuery.event.add/elemData.handle@http://code.jquery.com/jquery-2.1.4.js:4121:6 jQuery.event.trigger@http://code.jquery.com/jquery-2.1.4.js:4350:5 .trigger/<@http://code.jquery.com/jquery-2.1.4.js:4901:4 .each@http://code.jquery.com/jquery-2.1.4.js:374:14 jQuery.prototype.each@http://code.jquery.com/jquery-2.1.4.js:139:10 .trigger@http://code.jquery.com/jquery-2.1.4.js:4900:1 i/e.proceed@http://loklak.net/js/main.js:22:16928 ["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</va.prototype.functionCall/<@http://loklak.net/js/main.js:35:12265 ["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</</a@http://loklak.net/js/main.js:35:18791 Ht/this.$get</h.prototype.$eval@http://loklak.net/js/main.js:33:2755 Ht/this.$get</h.prototype.$apply@http://loklak.net/js/main.js:33:2980 ["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</<@http://loklak.net/js/main.js:35:18841 jQuery.event.dispatch@http://code.jquery.com/jquery-2.1.4.js:4434:15 jQuery.event.add/elemData.handle@http://code.jquery.com/jquery-2.1.4.js:4121:6 " ```
priority
wall not working on loklak net after i moved test loklak net to loklak net the wall is not working any more the javascript terminal shows the following error error apply already in progress i f ht this get h prototype apply or compile jquery event dispatch jquery event add elemdata handle jquery event trigger trigger each jquery prototype each trigger i e proceed va prototype functioncall or compile a ht this get h prototype eval ht this get h prototype apply or compile jquery event dispatch jquery event add elemdata handle
1
637,329
20,625,531,265
IssuesEvent
2022-03-07 22:04:33
microsoft/fluentui
https://api.github.com/repos/microsoft/fluentui
closed
Project: Manual accessibility review for vNext Beta/RC
Area: Accessibility Priority 1: High Fluent UI vNext
## Problem Fluent vNext components could use an accessibility review before we release them in Beta, just to catch any potential issues while they're still easy to change. ## Appetite 2 weeks ## Solution Review components that have UI and are included in the initial beta release, and raise any necessary issues - [x] Accordion #21101 - [x] Avatar #21102 - [x] Badge (includes CounterBadge and PresenceBadge) #21103 - [x] Button (includes CompoundButton, MenuButton, SplitButton, and ToggleButton) #20917 - [ ] Card #21104 - [x] Divider #21105 - [x] Link #19926 - [x] Menu #21106 - [x] Popover #21107 - [x] Tooltip #21108 ## Risks (Rabbit holes) N/A ## Out of scope (No-gos) N/A
1.0
Project: Manual accessibility review for vNext Beta/RC - ## Problem Fluent vNext components could use an accessibility review before we release them in Beta, just to catch any potential issues while they're still easy to change. ## Appetite 2 weeks ## Solution Review components that have UI and are included in the initial beta release, and raise any necessary issues - [x] Accordion #21101 - [x] Avatar #21102 - [x] Badge (includes CounterBadge and PresenceBadge) #21103 - [x] Button (includes CompoundButton, MenuButton, SplitButton, and ToggleButton) #20917 - [ ] Card #21104 - [x] Divider #21105 - [x] Link #19926 - [x] Menu #21106 - [x] Popover #21107 - [x] Tooltip #21108 ## Risks (Rabbit holes) N/A ## Out of scope (No-gos) N/A
priority
project manual accessibility review for vnext beta rc problem fluent vnext components could use an accessibility review before we release them in beta just to catch any potential issues while they re still easy to change appetite weeks solution review components that have ui and are included in the initial beta release and raise any necessary issues accordion avatar badge includes counterbadge and presencebadge button includes compoundbutton menubutton splitbutton and togglebutton card divider link menu popover tooltip risks rabbit holes n a out of scope no gos n a
1
314,430
9,596,903,720
IssuesEvent
2019-05-09 19:51:53
fgpv-vpgf/fgpv-vpgf
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
closed
Design Options for JSON File Data Table
addition: feature feedback: discussion priority: high
Original request can be found in https://github.com/fgpv-vpgf/fgpv-vpgf/issues/1415 Things have changed a bit since that request, and we are now looking at supporting layers that traditionally had no attribute support at all (i.e. adding a table feature instead of just making existing table features faster). Basically, the concept is for a layer we can target a json file to use as a source to populate the data table for that layer. There are a number of considerations for how fancy we make this. ## Level of Integration ### Unlinked In this case, the data table is essentially independent of the layer, with the exception that clicking the layer in the legend will open the table. Any additional functions (e.g. zoom to feature, details, icons) will need to be addressed for the unlinked case (details in sections below). This case is a :bear: minimum to implement. ### Linked In this case, the data table (and all other functions that run off attributes) remain connected to the map layer. The only difference is when the app goes to download the attribute set, it will source it from the json file instead of querying the map service. ObjectIds in the json file must align to the Ids on the map service. Why use a json file when the service already exists? Some benefits include faster load time, and possibly supporting Dynamic Layers on servers that have the `/query` endpoint disabled. The linked case will only work on layers that already support ESRI attributes (i.e. anything that inherits from `AttribRecord` -- mainly Feature Layers and Dynamic Layers). Supporting the linked mode is more work, but lets us hit additional use cases. If we decide not to implement in the first cut of file based tables, we should at least attempt to design the new code to allow support in the future. ## File Schema ### ESRI-like Dictionary This approach stores attributes in the JSON file using a similar/identical structure to how ArcGIS Server provides attributes. The pro of this approach is our current parsing routines can eat it raw. We can also create content from existing ESRI sources with relative ease. The con is the file size will be larger as attribute keys are repeated for every entry. ```text { fields: [ { name: "Field1", type: "esriFieldTypeString", alias: "Field One", length: 200 }, { name: "Field2", type: "esriFieldTypeDouble", alias: "Field Two" } ], features: [ { attributes: { Field1: "Big Duck Pond", Field2: 12.82 } }, { attributes: { Field1: "Germy Lake", Field2: 4.98 } } ] } ``` ### Array Based Structure This approach stores data in an array form (making it look similar to a table). The pro of this approach is the file size is minimized. The con is we need to write parsing code, and possibly it is more effort to translate date into this structure when generating the files. ```text { fields: [ { name: "Field1", type: "esriFieldTypeString", alias: "Field One", length: 200 }, { name: "Field2", type: "esriFieldTypeDouble", alias: "Field Two" } ], features: [ [ "Big Duck Pond", 12.82 ], [ "Germy Lake", 4.98 ] ] } ``` As it stands, this is the recommended approach for speed gains. ### Split Structure This approach separates the bulk data and the metadata into two files. This lets us quickly download table metadata (e.g. field names) up front if we need them. Whether this is necessary depends on how much support we give to non-attribute layer types (e.g. tabe on WMS, table on ESRI Tile). If we essentially disable everything but "table view" then it may be unnecessary. ## Config Structure ### Core Elements Need an optional property on a `layerNode` element to allow a url to a json file to be defined. The presence of a value in that property will trigger the appropriate setup for a file based grid (see sections below). ### Optional Elements Potential for additional properties to define enhanced things. * Marking a lat & long field in the data to enable a zoom to point function * An indicator if the file is linked to the layer, or is independent (e.g. object ids in file align to object ids on service). Should default to unlinked for safety. * Ability to provide an icon for each row. This could be enhanced to fancier structures (essentially providing a class break or unique value renderer type thing); not recommended at this point. * Ability to specify a name field? Would be needed if we want to support a details pane view with a reliable title value (See `Details Pane` section below for more analysis). If we are considering adding options beyond the url, we may want to create a sub-object to contain all the json table properties. E.g. (field names not final): ```text { id: "myLayer", url: "http://myserver.com/fancymap/", jsonTable: { url: "http://myserver.com/fancytable.json", linked: false, icon: "http://myserver.com/happyDuck.png" } } ``` ## Loading Mechanism ### Considerations * Files must be loaded in the Angular code (using `$http.get()`). The request library in geoApi fails with files due to the proxy being cooked into `esriRequest`. (See https://github.com/fgpv-vpgf/fgpv-vpgf/issues/2061#issuecomment-313451883 ) * The results can be provided from the same function within geoApi as normal attributes. See `attribute.js - newLayerPackage()`. * The loading from the server should not begin until `layerPackage.getAttribs()` is first called. * This approach allows for existing workflows regarding the table / attributes to be re-used with minimal impact * The results can be provided via a separate mechanism * Would only support unlinked mode * The "load the table" step would recognize its the json file case and call the alternate method to get data * Allows us to ignore compatibility issues of trying to have traditional attributes and json file attributes existing harmoniously in the same workflows ### Implementations If attempting to integrate with the standard `getAttribs()` approach: * We want the actual web request in angular, so that function will exist in the viewer. However, we trigger our attribute load from the geoApi. * Suggest our new geoApi function to setup the file table mode should accept a loader function. This way, when the geoApi is ready to load the table (usually first time opening the table), it can just execute the registered loader function. * Makes geoApi blissfully unaware that it's using angular. * Allows the option to switch to a different load approach later, as long as the function interface remains the same. * Likely need an alternate form of `newLayerPackage`, or need to overwrite it's `getAttribs` function. * The new function needs to see if `_attribData` exists, and if not, it needs to make a new Promise * Inside new Promise: execute the registered function to trigger the $http.get(), then massage result from json file format into esri format * Resolve Promise with proper object (see `createAttribSet`) * Likely need a third function similar to `loadFileAttribsBuilder` and `loadServerAttribsBuilder` (probably closer to `loadFileAttribsBuilder` in content). * It will set the various metadata properties correctly. * Need to be aware of sources, particularly if we are dealing with a layer that doesn't natively have attributes (e.g. fields would have to come from file, so they may not be set upfront) If taking the separate mechanism route -- we can pretty much do whatever we please. ## Layer Objects This new feature has the potential to complicate our layer hierarchy in geoApi. Attribute support is currently confined to the `AttribRecord` class and all inheritors. File based tables means classes not derived from `AttribRecord` now require some level of support. ### Minimal Approach - Integrated This approach essentially fakes the attribute loader on layers not related to `AttribRecord`. It would not use the class inheritance, and would instead inject appropriate functions onto Record/FC objects to make things look compatible. We can also entertain doing this on `AttribRecord` based classes, where we overwrite existing attribute loader logic with file loader logic on a case-by-case basis. This is a bit risky in that there are quite a few related functions that depend on attributes and may break / act weird. ### Minimal Approach - Separate Mechanism This ties into the separate mechanism approach described in the `Loading Mechanism` section above. Basically the json file to grid flow is separate from all existing structure. This approach will only support the Unlinked mode. So it causes the least amount of chaos, but offers the least amount of value. ### Reeefactor This involves significant changes to the Layer Record / FC object hierarchies. While it would need further analysis if we decide to persue this, at a high level: * Attribute support would need to exist at a low level in the hierarchy. The `LayerRecord` class would make sense. * The `LayerRecord` implementation could have the file based load support. If no file has been defined in the config, it can return an empty array or an error (to be decided which is more appropriate). * The `AttribRecord` class would retain it's code for supporting ArcGIS Server based attribute loading. However, it would need to allow for recognizing the file based case, and let the logic drop through to `LayerRecord`. In other words, it can't just steamroll the attribute logic; it needs to allow both branches of code to execute depending on the configuration. This approach offers good support for all layer types, and best supports the Linked mode (having file that aligns with server data). ## Details Pane In unlinked mode, we have a few choices * Disable the details pane. The data is in the grid anyways so its just showing the same thing but in vertical layout. * Lazy! * Should remove the detail button from the grid. * Push data to the standard zebra-stripe attribute view. * May need to disable things like zoom to feature and icon, depending on what configuration options we've added. * May also have to guess or suppress the "name field" if we have decided to not make that a config option (or it was not provided) * Could make the detail pane a configuration option. In linked mode, everything should work as normal as the data will be aligned to the layer metadata. ## Zoom To Feature In unlinked mode: * Default behavior would be no zoom. Button should not be in the grid. * Can have optional lat-long columns defined in config to zoom to point. Would involve a projection and then the zoom to point pipeline. * If geometry is a line/poly, only real option is to use the lat-long as centroid and accept that zoom level may be a poor choice for the shape. In linked mode, everything should work as normal as the data will be aligned to the layer metadata. ## Hovertips In the case of having a layer that natively supports attributes, and we are using a json sourced table, and the table is not aligned to the service, we would need to disable hovertips, or force them to always use the one-shot quickloader (essentially ignoring the attribute store even though it exists).
1.0
Design Options for JSON File Data Table - Original request can be found in https://github.com/fgpv-vpgf/fgpv-vpgf/issues/1415 Things have changed a bit since that request, and we are now looking at supporting layers that traditionally had no attribute support at all (i.e. adding a table feature instead of just making existing table features faster). Basically, the concept is for a layer we can target a json file to use as a source to populate the data table for that layer. There are a number of considerations for how fancy we make this. ## Level of Integration ### Unlinked In this case, the data table is essentially independent of the layer, with the exception that clicking the layer in the legend will open the table. Any additional functions (e.g. zoom to feature, details, icons) will need to be addressed for the unlinked case (details in sections below). This case is a :bear: minimum to implement. ### Linked In this case, the data table (and all other functions that run off attributes) remain connected to the map layer. The only difference is when the app goes to download the attribute set, it will source it from the json file instead of querying the map service. ObjectIds in the json file must align to the Ids on the map service. Why use a json file when the service already exists? Some benefits include faster load time, and possibly supporting Dynamic Layers on servers that have the `/query` endpoint disabled. The linked case will only work on layers that already support ESRI attributes (i.e. anything that inherits from `AttribRecord` -- mainly Feature Layers and Dynamic Layers). Supporting the linked mode is more work, but lets us hit additional use cases. If we decide not to implement in the first cut of file based tables, we should at least attempt to design the new code to allow support in the future. ## File Schema ### ESRI-like Dictionary This approach stores attributes in the JSON file using a similar/identical structure to how ArcGIS Server provides attributes. The pro of this approach is our current parsing routines can eat it raw. We can also create content from existing ESRI sources with relative ease. The con is the file size will be larger as attribute keys are repeated for every entry. ```text { fields: [ { name: "Field1", type: "esriFieldTypeString", alias: "Field One", length: 200 }, { name: "Field2", type: "esriFieldTypeDouble", alias: "Field Two" } ], features: [ { attributes: { Field1: "Big Duck Pond", Field2: 12.82 } }, { attributes: { Field1: "Germy Lake", Field2: 4.98 } } ] } ``` ### Array Based Structure This approach stores data in an array form (making it look similar to a table). The pro of this approach is the file size is minimized. The con is we need to write parsing code, and possibly it is more effort to translate date into this structure when generating the files. ```text { fields: [ { name: "Field1", type: "esriFieldTypeString", alias: "Field One", length: 200 }, { name: "Field2", type: "esriFieldTypeDouble", alias: "Field Two" } ], features: [ [ "Big Duck Pond", 12.82 ], [ "Germy Lake", 4.98 ] ] } ``` As it stands, this is the recommended approach for speed gains. ### Split Structure This approach separates the bulk data and the metadata into two files. This lets us quickly download table metadata (e.g. field names) up front if we need them. Whether this is necessary depends on how much support we give to non-attribute layer types (e.g. tabe on WMS, table on ESRI Tile). If we essentially disable everything but "table view" then it may be unnecessary. ## Config Structure ### Core Elements Need an optional property on a `layerNode` element to allow a url to a json file to be defined. The presence of a value in that property will trigger the appropriate setup for a file based grid (see sections below). ### Optional Elements Potential for additional properties to define enhanced things. * Marking a lat & long field in the data to enable a zoom to point function * An indicator if the file is linked to the layer, or is independent (e.g. object ids in file align to object ids on service). Should default to unlinked for safety. * Ability to provide an icon for each row. This could be enhanced to fancier structures (essentially providing a class break or unique value renderer type thing); not recommended at this point. * Ability to specify a name field? Would be needed if we want to support a details pane view with a reliable title value (See `Details Pane` section below for more analysis). If we are considering adding options beyond the url, we may want to create a sub-object to contain all the json table properties. E.g. (field names not final): ```text { id: "myLayer", url: "http://myserver.com/fancymap/", jsonTable: { url: "http://myserver.com/fancytable.json", linked: false, icon: "http://myserver.com/happyDuck.png" } } ``` ## Loading Mechanism ### Considerations * Files must be loaded in the Angular code (using `$http.get()`). The request library in geoApi fails with files due to the proxy being cooked into `esriRequest`. (See https://github.com/fgpv-vpgf/fgpv-vpgf/issues/2061#issuecomment-313451883 ) * The results can be provided from the same function within geoApi as normal attributes. See `attribute.js - newLayerPackage()`. * The loading from the server should not begin until `layerPackage.getAttribs()` is first called. * This approach allows for existing workflows regarding the table / attributes to be re-used with minimal impact * The results can be provided via a separate mechanism * Would only support unlinked mode * The "load the table" step would recognize its the json file case and call the alternate method to get data * Allows us to ignore compatibility issues of trying to have traditional attributes and json file attributes existing harmoniously in the same workflows ### Implementations If attempting to integrate with the standard `getAttribs()` approach: * We want the actual web request in angular, so that function will exist in the viewer. However, we trigger our attribute load from the geoApi. * Suggest our new geoApi function to setup the file table mode should accept a loader function. This way, when the geoApi is ready to load the table (usually first time opening the table), it can just execute the registered loader function. * Makes geoApi blissfully unaware that it's using angular. * Allows the option to switch to a different load approach later, as long as the function interface remains the same. * Likely need an alternate form of `newLayerPackage`, or need to overwrite it's `getAttribs` function. * The new function needs to see if `_attribData` exists, and if not, it needs to make a new Promise * Inside new Promise: execute the registered function to trigger the $http.get(), then massage result from json file format into esri format * Resolve Promise with proper object (see `createAttribSet`) * Likely need a third function similar to `loadFileAttribsBuilder` and `loadServerAttribsBuilder` (probably closer to `loadFileAttribsBuilder` in content). * It will set the various metadata properties correctly. * Need to be aware of sources, particularly if we are dealing with a layer that doesn't natively have attributes (e.g. fields would have to come from file, so they may not be set upfront) If taking the separate mechanism route -- we can pretty much do whatever we please. ## Layer Objects This new feature has the potential to complicate our layer hierarchy in geoApi. Attribute support is currently confined to the `AttribRecord` class and all inheritors. File based tables means classes not derived from `AttribRecord` now require some level of support. ### Minimal Approach - Integrated This approach essentially fakes the attribute loader on layers not related to `AttribRecord`. It would not use the class inheritance, and would instead inject appropriate functions onto Record/FC objects to make things look compatible. We can also entertain doing this on `AttribRecord` based classes, where we overwrite existing attribute loader logic with file loader logic on a case-by-case basis. This is a bit risky in that there are quite a few related functions that depend on attributes and may break / act weird. ### Minimal Approach - Separate Mechanism This ties into the separate mechanism approach described in the `Loading Mechanism` section above. Basically the json file to grid flow is separate from all existing structure. This approach will only support the Unlinked mode. So it causes the least amount of chaos, but offers the least amount of value. ### Reeefactor This involves significant changes to the Layer Record / FC object hierarchies. While it would need further analysis if we decide to persue this, at a high level: * Attribute support would need to exist at a low level in the hierarchy. The `LayerRecord` class would make sense. * The `LayerRecord` implementation could have the file based load support. If no file has been defined in the config, it can return an empty array or an error (to be decided which is more appropriate). * The `AttribRecord` class would retain it's code for supporting ArcGIS Server based attribute loading. However, it would need to allow for recognizing the file based case, and let the logic drop through to `LayerRecord`. In other words, it can't just steamroll the attribute logic; it needs to allow both branches of code to execute depending on the configuration. This approach offers good support for all layer types, and best supports the Linked mode (having file that aligns with server data). ## Details Pane In unlinked mode, we have a few choices * Disable the details pane. The data is in the grid anyways so its just showing the same thing but in vertical layout. * Lazy! * Should remove the detail button from the grid. * Push data to the standard zebra-stripe attribute view. * May need to disable things like zoom to feature and icon, depending on what configuration options we've added. * May also have to guess or suppress the "name field" if we have decided to not make that a config option (or it was not provided) * Could make the detail pane a configuration option. In linked mode, everything should work as normal as the data will be aligned to the layer metadata. ## Zoom To Feature In unlinked mode: * Default behavior would be no zoom. Button should not be in the grid. * Can have optional lat-long columns defined in config to zoom to point. Would involve a projection and then the zoom to point pipeline. * If geometry is a line/poly, only real option is to use the lat-long as centroid and accept that zoom level may be a poor choice for the shape. In linked mode, everything should work as normal as the data will be aligned to the layer metadata. ## Hovertips In the case of having a layer that natively supports attributes, and we are using a json sourced table, and the table is not aligned to the service, we would need to disable hovertips, or force them to always use the one-shot quickloader (essentially ignoring the attribute store even though it exists).
priority
design options for json file data table original request can be found in things have changed a bit since that request and we are now looking at supporting layers that traditionally had no attribute support at all i e adding a table feature instead of just making existing table features faster basically the concept is for a layer we can target a json file to use as a source to populate the data table for that layer there are a number of considerations for how fancy we make this level of integration unlinked in this case the data table is essentially independent of the layer with the exception that clicking the layer in the legend will open the table any additional functions e g zoom to feature details icons will need to be addressed for the unlinked case details in sections below this case is a bear minimum to implement linked in this case the data table and all other functions that run off attributes remain connected to the map layer the only difference is when the app goes to download the attribute set it will source it from the json file instead of querying the map service objectids in the json file must align to the ids on the map service why use a json file when the service already exists some benefits include faster load time and possibly supporting dynamic layers on servers that have the query endpoint disabled the linked case will only work on layers that already support esri attributes i e anything that inherits from attribrecord mainly feature layers and dynamic layers supporting the linked mode is more work but lets us hit additional use cases if we decide not to implement in the first cut of file based tables we should at least attempt to design the new code to allow support in the future file schema esri like dictionary this approach stores attributes in the json file using a similar identical structure to how arcgis server provides attributes the pro of this approach is our current parsing routines can eat it raw we can also create content from existing esri sources with relative ease the con is the file size will be larger as attribute keys are repeated for every entry text fields name type esrifieldtypestring alias field one length name type esrifieldtypedouble alias field two features attributes big duck pond attributes germy lake array based structure this approach stores data in an array form making it look similar to a table the pro of this approach is the file size is minimized the con is we need to write parsing code and possibly it is more effort to translate date into this structure when generating the files text fields name type esrifieldtypestring alias field one length name type esrifieldtypedouble alias field two features as it stands this is the recommended approach for speed gains split structure this approach separates the bulk data and the metadata into two files this lets us quickly download table metadata e g field names up front if we need them whether this is necessary depends on how much support we give to non attribute layer types e g tabe on wms table on esri tile if we essentially disable everything but table view then it may be unnecessary config structure core elements need an optional property on a layernode element to allow a url to a json file to be defined the presence of a value in that property will trigger the appropriate setup for a file based grid see sections below optional elements potential for additional properties to define enhanced things marking a lat long field in the data to enable a zoom to point function an indicator if the file is linked to the layer or is independent e g object ids in file align to object ids on service should default to unlinked for safety ability to provide an icon for each row this could be enhanced to fancier structures essentially providing a class break or unique value renderer type thing not recommended at this point ability to specify a name field would be needed if we want to support a details pane view with a reliable title value see details pane section below for more analysis if we are considering adding options beyond the url we may want to create a sub object to contain all the json table properties e g field names not final text id mylayer url jsontable url linked false icon loading mechanism considerations files must be loaded in the angular code using http get the request library in geoapi fails with files due to the proxy being cooked into esrirequest see the results can be provided from the same function within geoapi as normal attributes see attribute js newlayerpackage the loading from the server should not begin until layerpackage getattribs is first called this approach allows for existing workflows regarding the table attributes to be re used with minimal impact the results can be provided via a separate mechanism would only support unlinked mode the load the table step would recognize its the json file case and call the alternate method to get data allows us to ignore compatibility issues of trying to have traditional attributes and json file attributes existing harmoniously in the same workflows implementations if attempting to integrate with the standard getattribs approach we want the actual web request in angular so that function will exist in the viewer however we trigger our attribute load from the geoapi suggest our new geoapi function to setup the file table mode should accept a loader function this way when the geoapi is ready to load the table usually first time opening the table it can just execute the registered loader function makes geoapi blissfully unaware that it s using angular allows the option to switch to a different load approach later as long as the function interface remains the same likely need an alternate form of newlayerpackage or need to overwrite it s getattribs function the new function needs to see if attribdata exists and if not it needs to make a new promise inside new promise execute the registered function to trigger the http get then massage result from json file format into esri format resolve promise with proper object see createattribset likely need a third function similar to loadfileattribsbuilder and loadserverattribsbuilder probably closer to loadfileattribsbuilder in content it will set the various metadata properties correctly need to be aware of sources particularly if we are dealing with a layer that doesn t natively have attributes e g fields would have to come from file so they may not be set upfront if taking the separate mechanism route we can pretty much do whatever we please layer objects this new feature has the potential to complicate our layer hierarchy in geoapi attribute support is currently confined to the attribrecord class and all inheritors file based tables means classes not derived from attribrecord now require some level of support minimal approach integrated this approach essentially fakes the attribute loader on layers not related to attribrecord it would not use the class inheritance and would instead inject appropriate functions onto record fc objects to make things look compatible we can also entertain doing this on attribrecord based classes where we overwrite existing attribute loader logic with file loader logic on a case by case basis this is a bit risky in that there are quite a few related functions that depend on attributes and may break act weird minimal approach separate mechanism this ties into the separate mechanism approach described in the loading mechanism section above basically the json file to grid flow is separate from all existing structure this approach will only support the unlinked mode so it causes the least amount of chaos but offers the least amount of value reeefactor this involves significant changes to the layer record fc object hierarchies while it would need further analysis if we decide to persue this at a high level attribute support would need to exist at a low level in the hierarchy the layerrecord class would make sense the layerrecord implementation could have the file based load support if no file has been defined in the config it can return an empty array or an error to be decided which is more appropriate the attribrecord class would retain it s code for supporting arcgis server based attribute loading however it would need to allow for recognizing the file based case and let the logic drop through to layerrecord in other words it can t just steamroll the attribute logic it needs to allow both branches of code to execute depending on the configuration this approach offers good support for all layer types and best supports the linked mode having file that aligns with server data details pane in unlinked mode we have a few choices disable the details pane the data is in the grid anyways so its just showing the same thing but in vertical layout lazy should remove the detail button from the grid push data to the standard zebra stripe attribute view may need to disable things like zoom to feature and icon depending on what configuration options we ve added may also have to guess or suppress the name field if we have decided to not make that a config option or it was not provided could make the detail pane a configuration option in linked mode everything should work as normal as the data will be aligned to the layer metadata zoom to feature in unlinked mode default behavior would be no zoom button should not be in the grid can have optional lat long columns defined in config to zoom to point would involve a projection and then the zoom to point pipeline if geometry is a line poly only real option is to use the lat long as centroid and accept that zoom level may be a poor choice for the shape in linked mode everything should work as normal as the data will be aligned to the layer metadata hovertips in the case of having a layer that natively supports attributes and we are using a json sourced table and the table is not aligned to the service we would need to disable hovertips or force them to always use the one shot quickloader essentially ignoring the attribute store even though it exists
1
501,443
14,528,424,543
IssuesEvent
2020-12-14 16:31:09
E3SM-Project/scorpio
https://api.github.com/repos/E3SM-Project/scorpio
opened
Scorpio asserts for non-NULL starts/counts when writing scalar variables
High Priority Next Release
Scorpio asserts when the starts/counts are valid when writing out scalar variables. Although passing starts/counts don't make sense when writing out scalar variables libraries like NetCDF allow (and silently ignore) it. From pio_getput_int.c ``` ... if (ndims == 0) { /* This is a scalar var. */ LOG((2, "pnetcdf writing scalar with ncmpi_put_vars_*() file->fh = %d varid = %d", file->fh, varid)); pioassert(!start && !count && !stride, "expected NULLs", __FILE__, __LINE__); ... ``` We will need to remove this assert and include a tighter check and warning to the user. Also see E3SM-Project/E3SM#3988
1.0
Scorpio asserts for non-NULL starts/counts when writing scalar variables - Scorpio asserts when the starts/counts are valid when writing out scalar variables. Although passing starts/counts don't make sense when writing out scalar variables libraries like NetCDF allow (and silently ignore) it. From pio_getput_int.c ``` ... if (ndims == 0) { /* This is a scalar var. */ LOG((2, "pnetcdf writing scalar with ncmpi_put_vars_*() file->fh = %d varid = %d", file->fh, varid)); pioassert(!start && !count && !stride, "expected NULLs", __FILE__, __LINE__); ... ``` We will need to remove this assert and include a tighter check and warning to the user. Also see E3SM-Project/E3SM#3988
priority
scorpio asserts for non null starts counts when writing scalar variables scorpio asserts when the starts counts are valid when writing out scalar variables although passing starts counts don t make sense when writing out scalar variables libraries like netcdf allow and silently ignore it from pio getput int c if ndims this is a scalar var log pnetcdf writing scalar with ncmpi put vars file fh d varid d file fh varid pioassert start count stride expected nulls file line we will need to remove this assert and include a tighter check and warning to the user also see project
1
381,825
11,296,020,360
IssuesEvent
2020-01-17 00:14:02
lokka30/PhantomCombat
https://api.github.com/repos/lokka30/PhantomCombat
closed
plural feature for combat mode boss bar
enhancement priority: high
at the moment it says '1 seconds' but there should be a placeholder so it says '1 second' e.g. '%time% second%s%'
1.0
plural feature for combat mode boss bar - at the moment it says '1 seconds' but there should be a placeholder so it says '1 second' e.g. '%time% second%s%'
priority
plural feature for combat mode boss bar at the moment it says seconds but there should be a placeholder so it says second e g time second s
1
372,448
11,014,791,415
IssuesEvent
2019-12-04 23:41:11
data2health/maturity-model
https://api.github.com/repos/data2health/maturity-model
closed
Develop app login page
Priority: High enhancement
Should - [x] Authenticate with Python Flask API - [x] Check user email and survey entry code
1.0
Develop app login page - Should - [x] Authenticate with Python Flask API - [x] Check user email and survey entry code
priority
develop app login page should authenticate with python flask api check user email and survey entry code
1
193,313
6,883,852,647
IssuesEvent
2017-11-21 10:50:17
wordpress-mobile/AztecEditor-Android
https://api.github.com/repos/wordpress-mobile/AztecEditor-Android
closed
List items broken for RTL languages
bug high priority
The list items (both ordered and unordered) are not rendered properly for RTL languages. The bullets and numbers are not displayed. ![image](https://user-images.githubusercontent.com/1522856/32543884-321dfe96-c477-11e7-9107-e58dc54a5bff.png)
1.0
List items broken for RTL languages - The list items (both ordered and unordered) are not rendered properly for RTL languages. The bullets and numbers are not displayed. ![image](https://user-images.githubusercontent.com/1522856/32543884-321dfe96-c477-11e7-9107-e58dc54a5bff.png)
priority
list items broken for rtl languages the list items both ordered and unordered are not rendered properly for rtl languages the bullets and numbers are not displayed
1
474,653
13,673,223,327
IssuesEvent
2020-09-29 09:33:00
swan-cern/jupyter-extensions
https://api.github.com/repos/swan-cern/jupyter-extensions
closed
Renew of oAuth tokens inside the user session
priority:high type:feature request
We need to support oAuth token for accessing EOS and, in the future, to access the CS3APIs #26. The token can, and will, be pushed while Spawning, but we also need to ensure that is renewed inside the user session. We should be able able to retrieve that information directly from the authenticator renewal process. The best way to retrieve this information is to try to get the user auth_state directly from the JH API. If this information is not available, we create a Authenticated Service that will return this information.
1.0
Renew of oAuth tokens inside the user session - We need to support oAuth token for accessing EOS and, in the future, to access the CS3APIs #26. The token can, and will, be pushed while Spawning, but we also need to ensure that is renewed inside the user session. We should be able able to retrieve that information directly from the authenticator renewal process. The best way to retrieve this information is to try to get the user auth_state directly from the JH API. If this information is not available, we create a Authenticated Service that will return this information.
priority
renew of oauth tokens inside the user session we need to support oauth token for accessing eos and in the future to access the the token can and will be pushed while spawning but we also need to ensure that is renewed inside the user session we should be able able to retrieve that information directly from the authenticator renewal process the best way to retrieve this information is to try to get the user auth state directly from the jh api if this information is not available we create a authenticated service that will return this information
1
495,108
14,272,193,420
IssuesEvent
2020-11-21 15:58:36
UC-Davis-molecular-computing/scadnano
https://api.github.com/repos/UC-Davis-molecular-computing/scadnano
opened
Fix deprecation warning on Release action
bug super high priority
In [this run](https://github.com/UC-Davis-molecular-computing/scadnano/runs/1435916555?check_suite_focus=true), the action "failed" because of the use of deprecated `set-env` variable: ``` The `set-env` command is disabled. Please upgrade to using Environment Files or opt into unsecure command execution by setting the `ACTIONS_ALLOW_UNSECURE_COMMANDS` environment variable to `true`. For more information see: https://github.blog/changelog/2020-10-01-github-actions-deprecating-set-env-and-add-path-commands/ ``` The action was still able to successfully generate the release log, but the workflow is considered a failure.
1.0
Fix deprecation warning on Release action - In [this run](https://github.com/UC-Davis-molecular-computing/scadnano/runs/1435916555?check_suite_focus=true), the action "failed" because of the use of deprecated `set-env` variable: ``` The `set-env` command is disabled. Please upgrade to using Environment Files or opt into unsecure command execution by setting the `ACTIONS_ALLOW_UNSECURE_COMMANDS` environment variable to `true`. For more information see: https://github.blog/changelog/2020-10-01-github-actions-deprecating-set-env-and-add-path-commands/ ``` The action was still able to successfully generate the release log, but the workflow is considered a failure.
priority
fix deprecation warning on release action in the action failed because of the use of deprecated set env variable the set env command is disabled please upgrade to using environment files or opt into unsecure command execution by setting the actions allow unsecure commands environment variable to true for more information see the action was still able to successfully generate the release log but the workflow is considered a failure
1
499,273
14,443,858,154
IssuesEvent
2020-12-07 20:18:43
TheIOFoundation/ProjectLockdown
https://api.github.com/repos/TheIOFoundation/ProjectLockdown
opened
[ADM] Position: Crowdfunding assistant
Level: Member Module: ADM Need: Position Priority: High Project: PLD Stage: Ready Team: RA
![logo](https://user-images.githubusercontent.com/9198668/85232285-68543380-b430-11ea-8353-1aafb79baf78.png) # Open Position ## Position title Crowdfunding Assistant ## Position type Volunteer ## Position description The Crowdfunding assistant will be a member of Team Resource Allocation and will be responsible for the following: - Creating a comprehensive crowdfunding strategic plan - Setting up and maintaining accounts - Working with Team Communications to write the listing content and produce other necessary materials (such as brochures or pitch videos) - Working with Team UXUI to create graphics and other necessary materials - Developing and producing incentives - Updating the crowdfunding platform with latest developments - Building email lists - Report to the Team Resource Allocation Coordinator ## Skills & Qualifications - Requirements: -- Excellent communication skills -- Creative thinker -- Familiarity with platforms for crowdfunding (such as Patreon, GitHub Sponsors, Open Collective GoFundMe, Indiegogo and others) -- Able to participate in a highly dynamic environment as a part of a team -- Able to learn independently and dedicate at least 6 hours a week and attend regular team meetings -- Able to participate in a highly dynamic environment as a part of a team - Desirable -- Previous experience with Crowdfunding - Optional -- Previous experience with an NGO ## Disclaimer This is a completely volunteer, non remunerated position. Please ensure that there are no legal impediments that may enter in conflict with the position and tasks assigned. If interested, please fill the following form: https://TIOF.Click/PLDEoI
1.0
[ADM] Position: Crowdfunding assistant - ![logo](https://user-images.githubusercontent.com/9198668/85232285-68543380-b430-11ea-8353-1aafb79baf78.png) # Open Position ## Position title Crowdfunding Assistant ## Position type Volunteer ## Position description The Crowdfunding assistant will be a member of Team Resource Allocation and will be responsible for the following: - Creating a comprehensive crowdfunding strategic plan - Setting up and maintaining accounts - Working with Team Communications to write the listing content and produce other necessary materials (such as brochures or pitch videos) - Working with Team UXUI to create graphics and other necessary materials - Developing and producing incentives - Updating the crowdfunding platform with latest developments - Building email lists - Report to the Team Resource Allocation Coordinator ## Skills & Qualifications - Requirements: -- Excellent communication skills -- Creative thinker -- Familiarity with platforms for crowdfunding (such as Patreon, GitHub Sponsors, Open Collective GoFundMe, Indiegogo and others) -- Able to participate in a highly dynamic environment as a part of a team -- Able to learn independently and dedicate at least 6 hours a week and attend regular team meetings -- Able to participate in a highly dynamic environment as a part of a team - Desirable -- Previous experience with Crowdfunding - Optional -- Previous experience with an NGO ## Disclaimer This is a completely volunteer, non remunerated position. Please ensure that there are no legal impediments that may enter in conflict with the position and tasks assigned. If interested, please fill the following form: https://TIOF.Click/PLDEoI
priority
position crowdfunding assistant open position position title crowdfunding assistant position type volunteer position description the crowdfunding assistant will be a member of team resource allocation and will be responsible for the following creating a comprehensive crowdfunding strategic plan setting up and maintaining accounts working with team communications to write the listing content and produce other necessary materials such as brochures or pitch videos working with team uxui to create graphics and other necessary materials developing and producing incentives updating the crowdfunding platform with latest developments building email lists report to the team resource allocation coordinator skills qualifications requirements excellent communication skills creative thinker familiarity with platforms for crowdfunding such as patreon github sponsors open collective gofundme indiegogo and others able to participate in a highly dynamic environment as a part of a team able to learn independently and dedicate at least hours a week and attend regular team meetings able to participate in a highly dynamic environment as a part of a team desirable previous experience with crowdfunding optional previous experience with an ngo disclaimer this is a completely volunteer non remunerated position please ensure that there are no legal impediments that may enter in conflict with the position and tasks assigned if interested please fill the following form
1
586,100
17,569,660,273
IssuesEvent
2021-08-14 12:14:18
umple/umple
https://api.github.com/repos/umple/umple
closed
No option to cancel the Task Create Panel after a task is created in UmpleOnline
Component-UmpleOnline Priority-High Diffic-Easy tasks
## Summary Within the UmpleOnline, the users can create a task and save it. Once the task a saved, the system has no option to cancel this panel. Moreover, it does not show the "Load a task" option under the "TASKS" section. Therefore, the user is forced to reload the home page in browser to create a new task. ## Steps to Reproduce 1. Go to the "TASKS", then click on the "Create a Task". ![image](https://user-images.githubusercontent.com/82404135/128386311-5d59de04-6769-4e9d-9e9d-fa36cb3c535f.png) 2. Enter a sample value at task name and instructions. Click "Submit Task". ![image](https://user-images.githubusercontent.com/82404135/128386666-391c6f39-33c7-428c-af22-9d3fd1904922.png) 3. Now, there is no option to hide this task creation panel. This panel is visible at every action that the user does. ![image](https://user-images.githubusercontent.com/82404135/128387177-0892a0aa-4b2e-49d7-aeff-ca88b44286e9.png) ## Expected Feature The system should have a option to close this panel after a task is saved. A similar workflow is implemented when clicked on "Cancel this task response".
1.0
No option to cancel the Task Create Panel after a task is created in UmpleOnline - ## Summary Within the UmpleOnline, the users can create a task and save it. Once the task a saved, the system has no option to cancel this panel. Moreover, it does not show the "Load a task" option under the "TASKS" section. Therefore, the user is forced to reload the home page in browser to create a new task. ## Steps to Reproduce 1. Go to the "TASKS", then click on the "Create a Task". ![image](https://user-images.githubusercontent.com/82404135/128386311-5d59de04-6769-4e9d-9e9d-fa36cb3c535f.png) 2. Enter a sample value at task name and instructions. Click "Submit Task". ![image](https://user-images.githubusercontent.com/82404135/128386666-391c6f39-33c7-428c-af22-9d3fd1904922.png) 3. Now, there is no option to hide this task creation panel. This panel is visible at every action that the user does. ![image](https://user-images.githubusercontent.com/82404135/128387177-0892a0aa-4b2e-49d7-aeff-ca88b44286e9.png) ## Expected Feature The system should have a option to close this panel after a task is saved. A similar workflow is implemented when clicked on "Cancel this task response".
priority
no option to cancel the task create panel after a task is created in umpleonline summary within the umpleonline the users can create a task and save it once the task a saved the system has no option to cancel this panel moreover it does not show the load a task option under the tasks section therefore the user is forced to reload the home page in browser to create a new task steps to reproduce go to the tasks then click on the create a task enter a sample value at task name and instructions click submit task now there is no option to hide this task creation panel this panel is visible at every action that the user does expected feature the system should have a option to close this panel after a task is saved a similar workflow is implemented when clicked on cancel this task response
1
130,678
5,119,295,075
IssuesEvent
2017-01-08 16:31:55
bmoreonrails/rails_tutorial
https://api.github.com/repos/bmoreonrails/rails_tutorial
closed
Make the syntax colors less awful
help wanted high-priority
- [ ] get rid of ugly default colors like the red in the html blocks - [ ] We also need the contrast to work for anyone who might be colorblind. Here's just one tool for checking it - you can take a screen shot and upload it to here: http://www.color-blindness.com/coblis-color-blindness-simulator/ There is a syntax.scss file that I copied from GH in there that you can start with. The markup might be a little inaccurate in spots with regard to the language in the block - so be sure to check which language is indicated and change that if you need to.
1.0
Make the syntax colors less awful - - [ ] get rid of ugly default colors like the red in the html blocks - [ ] We also need the contrast to work for anyone who might be colorblind. Here's just one tool for checking it - you can take a screen shot and upload it to here: http://www.color-blindness.com/coblis-color-blindness-simulator/ There is a syntax.scss file that I copied from GH in there that you can start with. The markup might be a little inaccurate in spots with regard to the language in the block - so be sure to check which language is indicated and change that if you need to.
priority
make the syntax colors less awful get rid of ugly default colors like the red in the html blocks we also need the contrast to work for anyone who might be colorblind here s just one tool for checking it you can take a screen shot and upload it to here there is a syntax scss file that i copied from gh in there that you can start with the markup might be a little inaccurate in spots with regard to the language in the block so be sure to check which language is indicated and change that if you need to
1
471,267
13,563,633,771
IssuesEvent
2020-09-18 08:50:42
decentraland/explorer
https://api.github.com/repos/decentraland/explorer
closed
transparencyMode issues in materials
SDK bug high priority
Here we see a planeshape that appears to have z-fighting issues with what's behind it, even though there's a significant distance in between ![image.png](https://images.zenhubusercontent.com/5d9b940e491c060001c8647b/0ce71f83-2775-4836-9d39-5c8912f83fee) [https://app.zenhub.com/files/220994476/b53ddd64-9638-41d5-ba3a-979778dc26fc/download](https://app.zenhub.com/files/220994476/b53ddd64-9638-41d5-ba3a-979778dc26fc/download) Here the material is using a same image as the alphaTexture and the albedoTexture and transparencyMode is not defined spriteMaterial.albedoTexture = spriteTexture; spriteMaterial.alphaTexture = spriteTexture; If I add an extra line to this scene with a transparencyMode = 1, it fixes it spriteMaterial.transparencyMode = 1 But if I set that to 2 or 3, if fails again. It fails if undefined because 2 is the default value of transparencyMode. * | Value | Type | * | ----- | ---------------------------------------------- | * | 0 | OPAQUE (default) | * | 1 | ALPHATEST | * | 2 | ALPHABLEND | * | 3 | ALPHATESTANDBLEND | * | 4 | AUTO (ALPHABLEND if alpha OPAQUE otherwise | */ So ALPHATEST works, ALPHABLEND and ALPHATESTANDBLEND fail. Here's a second example where an alphaTexture is not working. https://github.com/MetaZoneio/AlphaDemo In this case, we're applying a different texture in alphaTexture from the one in albedoTexture. We should be seeing that a couple of these plains have a round shape being cut out, but the alphaTexture is not affecting what we see in any of the cases. ![Screen Shot 2020-06-30 at 10.15.23.png](https://images.zenhubusercontent.com/5d9b940e491c060001c8647b/d253c4c8-f9f8-4149-9f82-8856ddd27d48) We can edit the transparencyMode property for all of them, and set any value we want there, but nothing changes with any value we set.
1.0
transparencyMode issues in materials - Here we see a planeshape that appears to have z-fighting issues with what's behind it, even though there's a significant distance in between ![image.png](https://images.zenhubusercontent.com/5d9b940e491c060001c8647b/0ce71f83-2775-4836-9d39-5c8912f83fee) [https://app.zenhub.com/files/220994476/b53ddd64-9638-41d5-ba3a-979778dc26fc/download](https://app.zenhub.com/files/220994476/b53ddd64-9638-41d5-ba3a-979778dc26fc/download) Here the material is using a same image as the alphaTexture and the albedoTexture and transparencyMode is not defined spriteMaterial.albedoTexture = spriteTexture; spriteMaterial.alphaTexture = spriteTexture; If I add an extra line to this scene with a transparencyMode = 1, it fixes it spriteMaterial.transparencyMode = 1 But if I set that to 2 or 3, if fails again. It fails if undefined because 2 is the default value of transparencyMode. * | Value | Type | * | ----- | ---------------------------------------------- | * | 0 | OPAQUE (default) | * | 1 | ALPHATEST | * | 2 | ALPHABLEND | * | 3 | ALPHATESTANDBLEND | * | 4 | AUTO (ALPHABLEND if alpha OPAQUE otherwise | */ So ALPHATEST works, ALPHABLEND and ALPHATESTANDBLEND fail. Here's a second example where an alphaTexture is not working. https://github.com/MetaZoneio/AlphaDemo In this case, we're applying a different texture in alphaTexture from the one in albedoTexture. We should be seeing that a couple of these plains have a round shape being cut out, but the alphaTexture is not affecting what we see in any of the cases. ![Screen Shot 2020-06-30 at 10.15.23.png](https://images.zenhubusercontent.com/5d9b940e491c060001c8647b/d253c4c8-f9f8-4149-9f82-8856ddd27d48) We can edit the transparencyMode property for all of them, and set any value we want there, but nothing changes with any value we set.
priority
transparencymode issues in materials here we see a planeshape that appears to have z fighting issues with what s behind it even though there s a significant distance in between here the material is using a same image as the alphatexture and the albedotexture and transparencymode is not defined spritematerial albedotexture spritetexture spritematerial alphatexture spritetexture if i add an extra line to this scene with a transparencymode it fixes it spritematerial transparencymode but if i set that to or if fails again it fails if undefined because is the default value of transparencymode value type opaque default alphatest alphablend alphatestandblend auto alphablend if alpha opaque otherwise so alphatest works alphablend and alphatestandblend fail here s a second example where an alphatexture is not working in this case we re applying a different texture in alphatexture from the one in albedotexture we should be seeing that a couple of these plains have a round shape being cut out but the alphatexture is not affecting what we see in any of the cases we can edit the transparencymode property for all of them and set any value we want there but nothing changes with any value we set
1
434,504
12,519,339,817
IssuesEvent
2020-06-03 14:18:17
luna/ide
https://api.github.com/repos/luna/ide
closed
Nodes and edges view
Category: IDE Change: Non-Breaking Difficulty: Core Contributor Epic Priority: Highest Type: Enhancement
- Collapsed and expanded (no widgets yet) - With labels above edges - With text above nodes - Ability to change the expression - flags
1.0
Nodes and edges view - - Collapsed and expanded (no widgets yet) - With labels above edges - With text above nodes - Ability to change the expression - flags
priority
nodes and edges view collapsed and expanded no widgets yet with labels above edges with text above nodes ability to change the expression flags
1
401,906
11,799,849,195
IssuesEvent
2020-03-18 16:32:40
geneontology/minerva
https://api.github.com/repos/geneontology/minerva
opened
API - add search models by ontology terms other than GO
high priority
Task requirement from Noctua Landing Page Project - [ ] include ontologies such as Uberon, CL in blazegraph - [ ] make sure one can perform a closure search (eg retrieve models also containing the more specific/children terms of the specified term) - [ ] Provide the API route for NLP UI
1.0
API - add search models by ontology terms other than GO - Task requirement from Noctua Landing Page Project - [ ] include ontologies such as Uberon, CL in blazegraph - [ ] make sure one can perform a closure search (eg retrieve models also containing the more specific/children terms of the specified term) - [ ] Provide the API route for NLP UI
priority
api add search models by ontology terms other than go task requirement from noctua landing page project include ontologies such as uberon cl in blazegraph make sure one can perform a closure search eg retrieve models also containing the more specific children terms of the specified term provide the api route for nlp ui
1
614,397
19,181,887,324
IssuesEvent
2021-12-04 14:50:25
BlueBubblesApp/bluebubbles-app
https://api.github.com/repos/BlueBubblesApp/bluebubbles-app
opened
Fix issue where image disappears after sending. Comes back after leave and re enter
Bug priority: high Alpha Difficulty: Medium
Not sure if this will work for you, but here is what I did: 1. Send an image with text 2. Before the image fully sends, leave the app 3. Wait a sec 4. Re enter the app 5. Image seems to disappear or flicker 6. Send a message and it comes back
1.0
Fix issue where image disappears after sending. Comes back after leave and re enter - Not sure if this will work for you, but here is what I did: 1. Send an image with text 2. Before the image fully sends, leave the app 3. Wait a sec 4. Re enter the app 5. Image seems to disappear or flicker 6. Send a message and it comes back
priority
fix issue where image disappears after sending comes back after leave and re enter not sure if this will work for you but here is what i did send an image with text before the image fully sends leave the app wait a sec re enter the app image seems to disappear or flicker send a message and it comes back
1
307,961
9,424,405,039
IssuesEvent
2019-04-11 14:02:52
level73/membernet
https://api.github.com/repos/level73/membernet
opened
Separate EMENA from Global results in membership fees
Priority: High Type: Development
Currently there is no option for viewing EMENA membership fees only (i presume they are grouped under global?). We'd like to separate them.
1.0
Separate EMENA from Global results in membership fees - Currently there is no option for viewing EMENA membership fees only (i presume they are grouped under global?). We'd like to separate them.
priority
separate emena from global results in membership fees currently there is no option for viewing emena membership fees only i presume they are grouped under global we d like to separate them
1
239,790
7,800,024,243
IssuesEvent
2018-06-09 03:37:37
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
closed
0006802: Error on setting a Flags message filter.
Bug Felamimail Mantis high priority
**Reported by esakol on 23 Jul 2012 21:05** **Version:** Milan (2012.03.5) When selecting a Flags filter in Felamimail&#039;s messages filter interface, the value ComboBox is empty. And when clicking at the value combo box, the following error is reported on firebug&#039;s console. this.selectionWidget is null this.selectionWidget.on(&#039;select&#039;, this.onRecordSelect, this); on Tinebase/js/widgets/grid/PickerFilter.js (line 303)
1.0
0006802: Error on setting a Flags message filter. - **Reported by esakol on 23 Jul 2012 21:05** **Version:** Milan (2012.03.5) When selecting a Flags filter in Felamimail&#039;s messages filter interface, the value ComboBox is empty. And when clicking at the value combo box, the following error is reported on firebug&#039;s console. this.selectionWidget is null this.selectionWidget.on(&#039;select&#039;, this.onRecordSelect, this); on Tinebase/js/widgets/grid/PickerFilter.js (line 303)
priority
error on setting a flags message filter reported by esakol on jul version milan when selecting a flags filter in felamimail s messages filter interface the value combobox is empty and when clicking at the value combo box the following error is reported on firebug s console this selectionwidget is null this selectionwidget on select this onrecordselect this on tinebase js widgets grid pickerfilter js line
1
340,562
10,273,310,344
IssuesEvent
2019-08-23 18:54:56
X-Plane/XPlane2Blender
https://api.github.com/repos/X-Plane/XPlane2Blender
closed
249: Convert Aircraft and Cockpit properties
2.49 Converter priority high
The first alpha of the converter should be Airplane only (with anything else a plus) What's left to implement: ### ATTR - [x] ATTR_draw_disable - [x] ATTR_solid_camera (game prop only, not TexFace version) - [x] ATTR_light_level (and _v1, and _v2 and lit_level) ### GLOBAL - [x] NORMAL_METALNESS - [x] BLEND_GLASS ~- GLOBAL_cockpit_lit (Isn't this simply always on now?)~ This was not a global prop, is true by default, and cockpit_lit only comes into play if the export version is less than `1100` - [x] GLOBAL_no_blend - [x] GLOBAL_specular - [x] GLOBAL_shadow_blend - [x] GLOBAL_tint Plus, of course, unit tests - [x] ATTR properties only - [x] GLOBAL_properties only These get applied before the mesh splits, since it previous applied to the whole Object anyway. It is just easier that way.
1.0
249: Convert Aircraft and Cockpit properties - The first alpha of the converter should be Airplane only (with anything else a plus) What's left to implement: ### ATTR - [x] ATTR_draw_disable - [x] ATTR_solid_camera (game prop only, not TexFace version) - [x] ATTR_light_level (and _v1, and _v2 and lit_level) ### GLOBAL - [x] NORMAL_METALNESS - [x] BLEND_GLASS ~- GLOBAL_cockpit_lit (Isn't this simply always on now?)~ This was not a global prop, is true by default, and cockpit_lit only comes into play if the export version is less than `1100` - [x] GLOBAL_no_blend - [x] GLOBAL_specular - [x] GLOBAL_shadow_blend - [x] GLOBAL_tint Plus, of course, unit tests - [x] ATTR properties only - [x] GLOBAL_properties only These get applied before the mesh splits, since it previous applied to the whole Object anyway. It is just easier that way.
priority
convert aircraft and cockpit properties the first alpha of the converter should be airplane only with anything else a plus what s left to implement attr attr draw disable attr solid camera game prop only not texface version attr light level and and and lit level global normal metalness blend glass global cockpit lit isn t this simply always on now this was not a global prop is true by default and cockpit lit only comes into play if the export version is less than global no blend global specular global shadow blend global tint plus of course unit tests attr properties only global properties only these get applied before the mesh splits since it previous applied to the whole object anyway it is just easier that way
1
461,111
13,223,867,159
IssuesEvent
2020-08-17 18:02:08
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
closed
Order details page not displaying variation/attribute name
bug component: variation has pull request priority: high
The other bug report for this was closed incorrectly and is being ignored, but the bug is very much present. On numerous of my client's sites, the order details do not say the variation's name in the title. I've tested this on WC 3.2.6 and 3.8, they both have the same issue. Occasionally it'll display the variation's name, and occasionally not. I can't work out why this happens sporadically. **To Reproduce** Check your order details. If it's happening to several of our sites it'd presumably appear on yours too. The below screenshots are 2 separate stores. You can clearly see there's a variation ID, but no variation name in the title. I'm getting angry emails from my clients and I haven't a clue what to suggest: [https://i.imgur.com/oqfZ13r.png](https://i.imgur.com/oqfZ13r.png) [https://i.imgur.com/5OlvvgQ.png](https://i.imgur.com/5OlvvgQ.png) **Expected behavior** ...that it'd say what the variation is in the title. **Isolating the problem (mark completed items with an [x]):** - [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active. - [x] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/). - [x] I can reproduce this bug consistently using the steps above. I'm leaving out the status report as this is affecting numerous WC instances on different versions/themes/etc. It's everywhere.
1.0
Order details page not displaying variation/attribute name - The other bug report for this was closed incorrectly and is being ignored, but the bug is very much present. On numerous of my client's sites, the order details do not say the variation's name in the title. I've tested this on WC 3.2.6 and 3.8, they both have the same issue. Occasionally it'll display the variation's name, and occasionally not. I can't work out why this happens sporadically. **To Reproduce** Check your order details. If it's happening to several of our sites it'd presumably appear on yours too. The below screenshots are 2 separate stores. You can clearly see there's a variation ID, but no variation name in the title. I'm getting angry emails from my clients and I haven't a clue what to suggest: [https://i.imgur.com/oqfZ13r.png](https://i.imgur.com/oqfZ13r.png) [https://i.imgur.com/5OlvvgQ.png](https://i.imgur.com/5OlvvgQ.png) **Expected behavior** ...that it'd say what the variation is in the title. **Isolating the problem (mark completed items with an [x]):** - [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active. - [x] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/). - [x] I can reproduce this bug consistently using the steps above. I'm leaving out the status report as this is affecting numerous WC instances on different versions/themes/etc. It's everywhere.
priority
order details page not displaying variation attribute name the other bug report for this was closed incorrectly and is being ignored but the bug is very much present on numerous of my client s sites the order details do not say the variation s name in the title i ve tested this on wc and they both have the same issue occasionally it ll display the variation s name and occasionally not i can t work out why this happens sporadically to reproduce check your order details if it s happening to several of our sites it d presumably appear on yours too the below screenshots are separate stores you can clearly see there s a variation id but no variation name in the title i m getting angry emails from my clients and i haven t a clue what to suggest expected behavior that it d say what the variation is in the title isolating the problem mark completed items with an i have deactivated other plugins and confirmed this bug occurs when only woocommerce plugin is active this bug happens with a default wordpress theme active or i can reproduce this bug consistently using the steps above i m leaving out the status report as this is affecting numerous wc instances on different versions themes etc it s everywhere
1
770,837
27,058,552,301
IssuesEvent
2023-02-13 17:53:17
union-platform/union-mobile-app
https://api.github.com/repos/union-platform/union-mobile-app
opened
User wants to apply to the team, because he believes that it is suitable for him
priority: high type: feature
**Scope of action:** Team Profile Screen **Precondition:** **Design:** https://www.figma.com/file/2St3zSul4fHnLffqy3WK7P/union-mobile?node-id=1481%3A777 ## Use cases: 1. The user clicks on the "submit an application" button 2. The system sends a request to the team ----- **Область действия:** Экран профиля команды **Предусловие:** **Дизайн:** https://www.figma.com/file/2St3zSul4fHnLffqy3WK7P/union-mobile?node-id=1481%3A777 ## Сценарий: 1. Пользователь нажимает на кнопку "подать заявку" 2. Система отправляет заявку в команду
1.0
User wants to apply to the team, because he believes that it is suitable for him - **Scope of action:** Team Profile Screen **Precondition:** **Design:** https://www.figma.com/file/2St3zSul4fHnLffqy3WK7P/union-mobile?node-id=1481%3A777 ## Use cases: 1. The user clicks on the "submit an application" button 2. The system sends a request to the team ----- **Область действия:** Экран профиля команды **Предусловие:** **Дизайн:** https://www.figma.com/file/2St3zSul4fHnLffqy3WK7P/union-mobile?node-id=1481%3A777 ## Сценарий: 1. Пользователь нажимает на кнопку "подать заявку" 2. Система отправляет заявку в команду
priority
user wants to apply to the team because he believes that it is suitable for him scope of action team profile screen precondition design use cases the user clicks on the submit an application button the system sends a request to the team область действия экран профиля команды предусловие дизайн сценарий пользователь нажимает на кнопку подать заявку система отправляет заявку в команду
1
594,395
18,044,731,285
IssuesEvent
2021-09-18 17:44:26
AY2122S1-CS2103T-W16-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-W16-2/tp
closed
Update about us page in project website
type.Task priority.High
The about us needs to be updated with the correct github link, photos, etc for every team member. Update Information for the following team members - [x] Ajay - [x] Eunice - [x] Geng Ning - [x] Imran - [x] Siva
1.0
Update about us page in project website - The about us needs to be updated with the correct github link, photos, etc for every team member. Update Information for the following team members - [x] Ajay - [x] Eunice - [x] Geng Ning - [x] Imran - [x] Siva
priority
update about us page in project website the about us needs to be updated with the correct github link photos etc for every team member update information for the following team members ajay eunice geng ning imran siva
1
486,355
14,007,856,620
IssuesEvent
2020-10-28 22:22:28
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
opened
Sign up process tells user "reset your password" and "forgot your password"
Priority: High Priority: Urgent bug
Something changed at one point in the sign up process. Instead of welcoming users who have just signed up, we tell them "reset your password" and direct them to a page "forgot your password". Expected: 1. Users who sign up should receive an email: "Hello, thank you for signing up on eventyay.com. Now please verify your email address and set a password for your account. Please click here [link]." ![Screenshot from 2020-10-28 23-14-22](https://user-images.githubusercontent.com/1583873/97502911-5e10eb80-1974-11eb-8292-8e1e7223fff9.png) 2. On the page that users are directed to then please show the following: * Thank you for verifying your email! Please create a password for your account. * Instead of "reset password" show "Set Password". * Do not show "new to the site? create an account..." ![Screenshot from 2020-10-28 23-14-41](https://user-images.githubusercontent.com/1583873/97502893-56514700-1974-11eb-9c44-59fa8787fcd8.png) 3. Ensure the changes do not impact the way passwords are reset for existing accounts. For existing verified accounts, the password reset process should stay as is.
2.0
Sign up process tells user "reset your password" and "forgot your password" - Something changed at one point in the sign up process. Instead of welcoming users who have just signed up, we tell them "reset your password" and direct them to a page "forgot your password". Expected: 1. Users who sign up should receive an email: "Hello, thank you for signing up on eventyay.com. Now please verify your email address and set a password for your account. Please click here [link]." ![Screenshot from 2020-10-28 23-14-22](https://user-images.githubusercontent.com/1583873/97502911-5e10eb80-1974-11eb-8292-8e1e7223fff9.png) 2. On the page that users are directed to then please show the following: * Thank you for verifying your email! Please create a password for your account. * Instead of "reset password" show "Set Password". * Do not show "new to the site? create an account..." ![Screenshot from 2020-10-28 23-14-41](https://user-images.githubusercontent.com/1583873/97502893-56514700-1974-11eb-9c44-59fa8787fcd8.png) 3. Ensure the changes do not impact the way passwords are reset for existing accounts. For existing verified accounts, the password reset process should stay as is.
priority
sign up process tells user reset your password and forgot your password something changed at one point in the sign up process instead of welcoming users who have just signed up we tell them reset your password and direct them to a page forgot your password expected users who sign up should receive an email hello thank you for signing up on eventyay com now please verify your email address and set a password for your account please click here on the page that users are directed to then please show the following thank you for verifying your email please create a password for your account instead of reset password show set password do not show new to the site create an account ensure the changes do not impact the way passwords are reset for existing accounts for existing verified accounts the password reset process should stay as is
1
765,677
26,856,985,808
IssuesEvent
2023-02-03 15:19:25
seddonym/grimp
https://api.github.com/repos/seddonym/grimp
closed
Better handling of modules that are not in the graph
high-priority
For example: ``` File ".../grimp/adaptors/graph.py", line 201, in find_shortest_chain target=imported)) File ".../networkx/algorithms/shortest_paths/generic.py", line 170, in shortest_path paths = nx.bidirectional_shortest_path(G, source, target) File ".../networkx/algorithms/shortest_paths/unweighted.py", line 223, in bidirectional_shortest_path raise nx.NodeNotFound(msg.format(source, target)) networkx.exception.NodeNotFound: Either source myproject.foo or target myproject.bar is not in G ``` This should raise a specific grimp exception class `ModuleNotInGraph` with a nicer error message.
1.0
Better handling of modules that are not in the graph - For example: ``` File ".../grimp/adaptors/graph.py", line 201, in find_shortest_chain target=imported)) File ".../networkx/algorithms/shortest_paths/generic.py", line 170, in shortest_path paths = nx.bidirectional_shortest_path(G, source, target) File ".../networkx/algorithms/shortest_paths/unweighted.py", line 223, in bidirectional_shortest_path raise nx.NodeNotFound(msg.format(source, target)) networkx.exception.NodeNotFound: Either source myproject.foo or target myproject.bar is not in G ``` This should raise a specific grimp exception class `ModuleNotInGraph` with a nicer error message.
priority
better handling of modules that are not in the graph for example file grimp adaptors graph py line in find shortest chain target imported file networkx algorithms shortest paths generic py line in shortest path paths nx bidirectional shortest path g source target file networkx algorithms shortest paths unweighted py line in bidirectional shortest path raise nx nodenotfound msg format source target networkx exception nodenotfound either source myproject foo or target myproject bar is not in g this should raise a specific grimp exception class modulenotingraph with a nicer error message
1
247,410
7,918,423,435
IssuesEvent
2018-07-04 13:16:24
openaq/openaq-fetch
https://api.github.com/repos/openaq/openaq-fetch
closed
China - Sources (not useable yet)
help wanted high priority new data question
There are a ton of China air quality data sources. Here are some: aepb.gov.cn bjmemc.com.cn cdemc.cn cfhb.gov.cn cepb.gov.cn dl.gov.cn dyhb.gov.cn nbemc.gov.cn sdein.gov.cn fjepb.gov.cn gsep.gansu.gov.cn qhepb.gov.cn gdep.gov.cn gxepb.gov.cn ghb.gov.cn dloer.gov.cn hebei.gov.cn hljdep.gov.cn hnep.gov.cn hbepb.gov.cn hbt.hunan.gov.cn nmgepb.gov.cn jshb.gov.cn jxepb.gov.cn shbj.klmy.gov.cn lnemc.cn lzhb.gov.cn nnems.gov.cn nxep.gov.cn ordoshb.gov.cn semc.gov.cn sxhjjcz.com.cn szhec.gov.cn tjemc.org.cn xzep.gov.cn wlmqhb.gov.cn whepb.gov.cn xianemc.gov.cn xnepb.gov.cn xjepb.gov.cn ynepb.gov.cn zjepb.gov.cn Sources from: aqicn.org/sources There is also a site that provides an API to Chinese AQ data: http://pm25.in/ However the issue is that from what I can tell, these data are not shared in their raw format only AQI (I have not clicked on each and everyone however). Anyone know differently or find ones above that are shared our in raw format?
1.0
China - Sources (not useable yet) - There are a ton of China air quality data sources. Here are some: aepb.gov.cn bjmemc.com.cn cdemc.cn cfhb.gov.cn cepb.gov.cn dl.gov.cn dyhb.gov.cn nbemc.gov.cn sdein.gov.cn fjepb.gov.cn gsep.gansu.gov.cn qhepb.gov.cn gdep.gov.cn gxepb.gov.cn ghb.gov.cn dloer.gov.cn hebei.gov.cn hljdep.gov.cn hnep.gov.cn hbepb.gov.cn hbt.hunan.gov.cn nmgepb.gov.cn jshb.gov.cn jxepb.gov.cn shbj.klmy.gov.cn lnemc.cn lzhb.gov.cn nnems.gov.cn nxep.gov.cn ordoshb.gov.cn semc.gov.cn sxhjjcz.com.cn szhec.gov.cn tjemc.org.cn xzep.gov.cn wlmqhb.gov.cn whepb.gov.cn xianemc.gov.cn xnepb.gov.cn xjepb.gov.cn ynepb.gov.cn zjepb.gov.cn Sources from: aqicn.org/sources There is also a site that provides an API to Chinese AQ data: http://pm25.in/ However the issue is that from what I can tell, these data are not shared in their raw format only AQI (I have not clicked on each and everyone however). Anyone know differently or find ones above that are shared our in raw format?
priority
china sources not useable yet there are a ton of china air quality data sources here are some aepb gov cn bjmemc com cn cdemc cn cfhb gov cn cepb gov cn dl gov cn dyhb gov cn nbemc gov cn sdein gov cn fjepb gov cn gsep gansu gov cn qhepb gov cn gdep gov cn gxepb gov cn ghb gov cn dloer gov cn hebei gov cn hljdep gov cn hnep gov cn hbepb gov cn hbt hunan gov cn nmgepb gov cn jshb gov cn jxepb gov cn shbj klmy gov cn lnemc cn lzhb gov cn nnems gov cn nxep gov cn ordoshb gov cn semc gov cn sxhjjcz com cn szhec gov cn tjemc org cn xzep gov cn wlmqhb gov cn whepb gov cn xianemc gov cn xnepb gov cn xjepb gov cn ynepb gov cn zjepb gov cn sources from aqicn org sources there is also a site that provides an api to chinese aq data however the issue is that from what i can tell these data are not shared in their raw format only aqi i have not clicked on each and everyone however anyone know differently or find ones above that are shared our in raw format
1
106,487
4,273,247,792
IssuesEvent
2016-07-13 16:43:31
TerraTex-Community/TerraTex-Reallife-Reloaded
https://api.github.com/repos/TerraTex-Community/TerraTex-Reallife-Reloaded
closed
Bug mit Drogen und Materialsverkäufer
bug EXTREMLY HIGH PRIORITY / EMERGENCY FIX NEEDED
Wenn man jeweils die Fenster auf hat, bei denen man sich Drogen oder Mats kaufen kann, und dabei die Läden verlässt, bleiben diese auf. Man kann also auf der gesamten Karte Materials und Drogen kaufen. Ebenfalls steigt dadurch der Preis der Gegenstände für einen selbst nicht an. Jedoch für alle anderen. Auch braucht man keine Masse an Drogen mehr mit sich herum tragen. und könte damit Kontrollen umgehen. ich schreibe es hier herein da ich irgendwie keinen Bugtracker gefunden habe und da Stalker mich dann auf das Github hingewiesen hat
1.0
Bug mit Drogen und Materialsverkäufer - Wenn man jeweils die Fenster auf hat, bei denen man sich Drogen oder Mats kaufen kann, und dabei die Läden verlässt, bleiben diese auf. Man kann also auf der gesamten Karte Materials und Drogen kaufen. Ebenfalls steigt dadurch der Preis der Gegenstände für einen selbst nicht an. Jedoch für alle anderen. Auch braucht man keine Masse an Drogen mehr mit sich herum tragen. und könte damit Kontrollen umgehen. ich schreibe es hier herein da ich irgendwie keinen Bugtracker gefunden habe und da Stalker mich dann auf das Github hingewiesen hat
priority
bug mit drogen und materialsverkäufer wenn man jeweils die fenster auf hat bei denen man sich drogen oder mats kaufen kann und dabei die läden verlässt bleiben diese auf man kann also auf der gesamten karte materials und drogen kaufen ebenfalls steigt dadurch der preis der gegenstände für einen selbst nicht an jedoch für alle anderen auch braucht man keine masse an drogen mehr mit sich herum tragen und könte damit kontrollen umgehen ich schreibe es hier herein da ich irgendwie keinen bugtracker gefunden habe und da stalker mich dann auf das github hingewiesen hat
1
468,748
13,489,753,229
IssuesEvent
2020-09-11 14:16:09
pxblue/angular-component-library
https://api.github.com/repos/pxblue/angular-component-library
closed
Components should set display rule for first rendered elements
bug high-priority info-list-item
#### Describe the bug The first element in our component DOM trees do not always set a display type. This can cause unexpected behavior when trying to style certain components. This is a known issue in the InfoListItem, but we should verify the other components as well. #### What is the expected behavior? First element in component trees should explicitly set the desired display type. #### Anything else to add? When the browser doesn't recognize a HTML tag (such as for custom components), it renders them with display:inline by default (with height/width of 0 and certain CSS rules having no effect). When the first element in the DOM of a component is also an unrecognized browser tag (such as `<mat-list-item>`), that seems to be when the issues arise - we just need to make sure these root elements have an explicit display type. Do this with #111
1.0
Components should set display rule for first rendered elements - #### Describe the bug The first element in our component DOM trees do not always set a display type. This can cause unexpected behavior when trying to style certain components. This is a known issue in the InfoListItem, but we should verify the other components as well. #### What is the expected behavior? First element in component trees should explicitly set the desired display type. #### Anything else to add? When the browser doesn't recognize a HTML tag (such as for custom components), it renders them with display:inline by default (with height/width of 0 and certain CSS rules having no effect). When the first element in the DOM of a component is also an unrecognized browser tag (such as `<mat-list-item>`), that seems to be when the issues arise - we just need to make sure these root elements have an explicit display type. Do this with #111
priority
components should set display rule for first rendered elements describe the bug the first element in our component dom trees do not always set a display type this can cause unexpected behavior when trying to style certain components this is a known issue in the infolistitem but we should verify the other components as well what is the expected behavior first element in component trees should explicitly set the desired display type anything else to add when the browser doesn t recognize a html tag such as for custom components it renders them with display inline by default with height width of and certain css rules having no effect when the first element in the dom of a component is also an unrecognized browser tag such as that seems to be when the issues arise we just need to make sure these root elements have an explicit display type do this with
1
371,674
10,979,908,571
IssuesEvent
2019-11-30 09:55:43
OperationCode/front-end
https://api.github.com/repos/OperationCode/front-end
closed
Cannot register with apostrophies in name
Priority: High Type: Bug beginner friendly good first issue
# Bug Report ## What is the current behavior? Doing names like: N'golo or O'Hara are impossible due to incorrect validation of the name inputs. ## What is the expected behavior? The name fields should have no validation. ## What steps did you take to get this behavior? Noticed a Cypress test failing with randomly generated data ### Screenshots ![Screen Shot 2019-11-18 at 2 13 01 PM](https://user-images.githubusercontent.com/9523719/69098589-bffb6e00-0a0d-11ea-8efb-c703d20848c4.png)
1.0
Cannot register with apostrophies in name - # Bug Report ## What is the current behavior? Doing names like: N'golo or O'Hara are impossible due to incorrect validation of the name inputs. ## What is the expected behavior? The name fields should have no validation. ## What steps did you take to get this behavior? Noticed a Cypress test failing with randomly generated data ### Screenshots ![Screen Shot 2019-11-18 at 2 13 01 PM](https://user-images.githubusercontent.com/9523719/69098589-bffb6e00-0a0d-11ea-8efb-c703d20848c4.png)
priority
cannot register with apostrophies in name bug report what is the current behavior doing names like n golo or o hara are impossible due to incorrect validation of the name inputs what is the expected behavior the name fields should have no validation what steps did you take to get this behavior noticed a cypress test failing with randomly generated data screenshots
1
82,592
3,615,800,127
IssuesEvent
2016-02-07 00:46:04
haskell/cabal
https://api.github.com/repos/haskell/cabal
closed
manual cabal flag toggled by modular solver
bug high-priority solver
With GHC 7.9/7.10, only `deepseq>=1.4.0.0` is installable. `semigroups-0.16` has the following flag defined: ``` flag deepseq description: You can disable the use of the `deepseq` package using `-f-deepseq`. . Disabing this is an unsupported configuration, but it may be useful for accelerating builds in sandboxes for expert users. default: True manual: True ``` and it's used for toggling a build-dep: ``` if flag(deepseq) build-depends: deepseq >= 1.1 && < 1.4 ``` So it's expected that `semigroups-0.16` isn't installable with GHC 7.9/7.10. However, Cabal decides to toggle the `deepseq` flag, even though it isn't allowed to: ``` $ cabal install semigroups-0.16 -v2 Reading available packages... Choosing modular solver. Resolving dependencies... Ready to install semigroups-0.16 Waiting for install task to finish... Extracting /home/hvr/.cabal/packages/hackage.haskell.org/semigroups/0.16/semigroups-0.16.tar.gz to /tmp/semigroups-0.16-25781... Updating semigroups.cabal with the latest revision from the index. Configuring semigroups-0.16... Flags chosen: unordered-containers=True, text=True, deepseq=False, containers=True, bytestring=True, hashable=True Dependency base ==4.8.0.0: using base-4.8.0.0 Dependency bytestring ==0.10.5.0: using bytestring-0.10.5.0 Dependency containers ==0.5.6.1: using containers-0.5.6.1 Dependency hashable ==1.2.3.0: using hashable-1.2.3.0 Dependency nats ==1: using nats-1 Dependency text ==1.2.0.3: using text-1.2.0.3 Dependency unordered-containers ==0.2.5.1: using unordered-containers-0.2.5.1 Using Cabal-1.21.1.0 compiled by ghc-7.4 Using compiler: ghc-7.9.20141216 Using install prefix: /home/hvr/.cabal Binaries installed in: /home/hvr/.cabal/bin Libraries installed in: ... ``` as you can see in the logging output above, surprisingly `deepseq=False` is set.
1.0
manual cabal flag toggled by modular solver - With GHC 7.9/7.10, only `deepseq>=1.4.0.0` is installable. `semigroups-0.16` has the following flag defined: ``` flag deepseq description: You can disable the use of the `deepseq` package using `-f-deepseq`. . Disabing this is an unsupported configuration, but it may be useful for accelerating builds in sandboxes for expert users. default: True manual: True ``` and it's used for toggling a build-dep: ``` if flag(deepseq) build-depends: deepseq >= 1.1 && < 1.4 ``` So it's expected that `semigroups-0.16` isn't installable with GHC 7.9/7.10. However, Cabal decides to toggle the `deepseq` flag, even though it isn't allowed to: ``` $ cabal install semigroups-0.16 -v2 Reading available packages... Choosing modular solver. Resolving dependencies... Ready to install semigroups-0.16 Waiting for install task to finish... Extracting /home/hvr/.cabal/packages/hackage.haskell.org/semigroups/0.16/semigroups-0.16.tar.gz to /tmp/semigroups-0.16-25781... Updating semigroups.cabal with the latest revision from the index. Configuring semigroups-0.16... Flags chosen: unordered-containers=True, text=True, deepseq=False, containers=True, bytestring=True, hashable=True Dependency base ==4.8.0.0: using base-4.8.0.0 Dependency bytestring ==0.10.5.0: using bytestring-0.10.5.0 Dependency containers ==0.5.6.1: using containers-0.5.6.1 Dependency hashable ==1.2.3.0: using hashable-1.2.3.0 Dependency nats ==1: using nats-1 Dependency text ==1.2.0.3: using text-1.2.0.3 Dependency unordered-containers ==0.2.5.1: using unordered-containers-0.2.5.1 Using Cabal-1.21.1.0 compiled by ghc-7.4 Using compiler: ghc-7.9.20141216 Using install prefix: /home/hvr/.cabal Binaries installed in: /home/hvr/.cabal/bin Libraries installed in: ... ``` as you can see in the logging output above, surprisingly `deepseq=False` is set.
priority
manual cabal flag toggled by modular solver with ghc only deepseq is installable semigroups has the following flag defined flag deepseq description you can disable the use of the deepseq package using f deepseq disabing this is an unsupported configuration but it may be useful for accelerating builds in sandboxes for expert users default true manual true and it s used for toggling a build dep if flag deepseq build depends deepseq so it s expected that semigroups isn t installable with ghc however cabal decides to toggle the deepseq flag even though it isn t allowed to cabal install semigroups reading available packages choosing modular solver resolving dependencies ready to install semigroups waiting for install task to finish extracting home hvr cabal packages hackage haskell org semigroups semigroups tar gz to tmp semigroups updating semigroups cabal with the latest revision from the index configuring semigroups flags chosen unordered containers true text true deepseq false containers true bytestring true hashable true dependency base using base dependency bytestring using bytestring dependency containers using containers dependency hashable using hashable dependency nats using nats dependency text using text dependency unordered containers using unordered containers using cabal compiled by ghc using compiler ghc using install prefix home hvr cabal binaries installed in home hvr cabal bin libraries installed in as you can see in the logging output above surprisingly deepseq false is set
1
354,546
10,569,210,203
IssuesEvent
2019-10-06 18:05:26
AY1920S1-CS2113-T14-1/main
https://api.github.com/repos/AY1920S1-CS2113-T14-1/main
opened
Generate final UML diagrams
component.Model priority.High type.Task
When preparing for final release, use the [Sketch It plugin for Intellij] (https://plugins.jetbrains.com/plugin/10387-sketch-it-) to generate final UML class diagrams for the Developer Guide.
1.0
Generate final UML diagrams - When preparing for final release, use the [Sketch It plugin for Intellij] (https://plugins.jetbrains.com/plugin/10387-sketch-it-) to generate final UML class diagrams for the Developer Guide.
priority
generate final uml diagrams when preparing for final release use the to generate final uml class diagrams for the developer guide
1
131,012
5,141,073,820
IssuesEvent
2017-01-12 08:39:50
hugopeek/romanesco-patterns
https://api.github.com/repos/hugopeek/romanesco-patterns
closed
setUserPlaceholders throws fatal error if user ID doesn't exist
Confirmed bug High priority
Fatal error: Call to a member function getOne() on null in /var/www/ddgc-merge/core/cache/includes/elements/modsnippet/83.include.cache.php on line 10
1.0
setUserPlaceholders throws fatal error if user ID doesn't exist - Fatal error: Call to a member function getOne() on null in /var/www/ddgc-merge/core/cache/includes/elements/modsnippet/83.include.cache.php on line 10
priority
setuserplaceholders throws fatal error if user id doesn t exist fatal error call to a member function getone on null in var www ddgc merge core cache includes elements modsnippet include cache php on line
1
815,616
30,564,530,419
IssuesEvent
2023-07-20 16:44:23
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
closed
Valor points are all sorts of fucked up.
Core Status: Need Info Priority: High
[//]: # (REMBEMBER! Add links to things related to the bug using for example:) [//]: # (http://wowhead.com/) [//]: # (cata-twinhead.twinstar.cz) **Description:** Raids are giving inconsistent amounts of valor points. They're all over the place. (no chars have done any dungeons this reset or anything else to gain valor + chars have been logged out and in to ensure its not a visual bug) Char 1: Completed BH10/BH25 = 175/1250 Valor Char 2: Completed BH10/BH25/Tot4W10/Tot4W25 = 280/1250 Valor Char 3: Completed BH25 = 70/1250 Valor Char 4: Completed BH10 = 70/1250 Valor Char 4 went into a BH25 and it's total went to 140/1250 (+70), completed the same raids as Char 1 and has 35 Valor less. Char 1 then went into a Tot4W10. After the first boss the chat window said +70 and the Valor points went to 210/1250 (an actual increase of 35) After the second boss the chat window said +70 and the Valor points went to 210/1250 **How to reproduce:** Go into raids. **How it should work:** From what I can see online it should be 75 per 10 man boss and 105 per 25 man boss (links below) although I'm not sure if this was changed at a later date, regardless of this the points given are inconsistent. **Database links:** https://www.wowhead.com/news/cataclysm-justice-valor-point-drop-quantities-170048 https://www.engadget.com/2010-10-06-cataclysm-justice-and-valor-point-drop-rates-revealed.html
1.0
Valor points are all sorts of fucked up. - [//]: # (REMBEMBER! Add links to things related to the bug using for example:) [//]: # (http://wowhead.com/) [//]: # (cata-twinhead.twinstar.cz) **Description:** Raids are giving inconsistent amounts of valor points. They're all over the place. (no chars have done any dungeons this reset or anything else to gain valor + chars have been logged out and in to ensure its not a visual bug) Char 1: Completed BH10/BH25 = 175/1250 Valor Char 2: Completed BH10/BH25/Tot4W10/Tot4W25 = 280/1250 Valor Char 3: Completed BH25 = 70/1250 Valor Char 4: Completed BH10 = 70/1250 Valor Char 4 went into a BH25 and it's total went to 140/1250 (+70), completed the same raids as Char 1 and has 35 Valor less. Char 1 then went into a Tot4W10. After the first boss the chat window said +70 and the Valor points went to 210/1250 (an actual increase of 35) After the second boss the chat window said +70 and the Valor points went to 210/1250 **How to reproduce:** Go into raids. **How it should work:** From what I can see online it should be 75 per 10 man boss and 105 per 25 man boss (links below) although I'm not sure if this was changed at a later date, regardless of this the points given are inconsistent. **Database links:** https://www.wowhead.com/news/cataclysm-justice-valor-point-drop-quantities-170048 https://www.engadget.com/2010-10-06-cataclysm-justice-and-valor-point-drop-rates-revealed.html
priority
valor points are all sorts of fucked up rembember add links to things related to the bug using for example cata twinhead twinstar cz description raids are giving inconsistent amounts of valor points they re all over the place no chars have done any dungeons this reset or anything else to gain valor chars have been logged out and in to ensure its not a visual bug char completed valor char completed valor char completed valor char completed valor char went into a and it s total went to completed the same raids as char and has valor less char then went into a after the first boss the chat window said and the valor points went to an actual increase of after the second boss the chat window said and the valor points went to how to reproduce go into raids how it should work from what i can see online it should be per man boss and per man boss links below although i m not sure if this was changed at a later date regardless of this the points given are inconsistent database links
1
343,573
10,332,845,312
IssuesEvent
2019-09-03 02:25:32
mateowatson/timetracker
https://api.github.com/repos/mateowatson/timetracker
closed
JavaScript implementation doesn’t support the back button
bug high-priority
I uncovered this issue while testing Turbolinks. Turbolinks _does_ solve this, but my inclination is to try to fix this in the current system because I don’t want to rush a wholesale update to Turbolinks. And this bug should be fixed ASAP.
1.0
JavaScript implementation doesn’t support the back button - I uncovered this issue while testing Turbolinks. Turbolinks _does_ solve this, but my inclination is to try to fix this in the current system because I don’t want to rush a wholesale update to Turbolinks. And this bug should be fixed ASAP.
priority
javascript implementation doesn’t support the back button i uncovered this issue while testing turbolinks turbolinks does solve this but my inclination is to try to fix this in the current system because i don’t want to rush a wholesale update to turbolinks and this bug should be fixed asap
1
412,993
12,059,067,581
IssuesEvent
2020-04-15 18:36:01
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
[TASK] Pre-loading scripts for tenant
Developer-Portal Priority/High Type/New Feature Type/Task
**Description:** The objective of this task is to write scripts to load templates such as IDP_TEMPLATE, APPLICATION_TEMPLATE, etc to the cloud by default.
1.0
[TASK] Pre-loading scripts for tenant - **Description:** The objective of this task is to write scripts to load templates such as IDP_TEMPLATE, APPLICATION_TEMPLATE, etc to the cloud by default.
priority
pre loading scripts for tenant description the objective of this task is to write scripts to load templates such as idp template application template etc to the cloud by default
1
512,763
14,908,934,274
IssuesEvent
2021-01-22 07:00:27
ckiplab/ckip-transformers
https://api.github.com/repos/ckiplab/ckip-transformers
opened
Implement custom Chinese tokenizer.
Priority: High
We may implement our own tokenizer rather than using BertTokenizerFast. Our own tokenizer should have the following features: - Disable word piece. Convert text to token ID character by character (e.g. `tokenizer.convert_tokens_to_ids(list(input_text))`) - Reimplement `clean_up_tokenization` method. The default method is implemented for English only. Our method may remove whitespaces and convert half-width punctuations to full-width ones.
1.0
Implement custom Chinese tokenizer. - We may implement our own tokenizer rather than using BertTokenizerFast. Our own tokenizer should have the following features: - Disable word piece. Convert text to token ID character by character (e.g. `tokenizer.convert_tokens_to_ids(list(input_text))`) - Reimplement `clean_up_tokenization` method. The default method is implemented for English only. Our method may remove whitespaces and convert half-width punctuations to full-width ones.
priority
implement custom chinese tokenizer we may implement our own tokenizer rather than using berttokenizerfast our own tokenizer should have the following features disable word piece convert text to token id character by character e g tokenizer convert tokens to ids list input text reimplement clean up tokenization method the default method is implemented for english only our method may remove whitespaces and convert half width punctuations to full width ones
1