Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
855
| labels
stringlengths 4
721
| body
stringlengths 1
261k
| index
stringclasses 13
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
240k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
87,075
| 3,736,717,427
|
IssuesEvent
|
2016-03-08 16:48:35
|
Fermat-ORG/fermat-org
|
https://api.github.com/repos/Fermat-ORG/fermat-org
|
closed
|
Update developer list
|
Priority: HIGH question server
|
Caso de estudio:
Miguelcldn es un nuevo developer que quiere entrar en Fermat, nunca tuvo una ficha asignada.
- Entra en la página
- Inicia sesión
- Crea un componente
- Pone "miguelcldn" en el nombre del developer
- Se crea la ficha y se pone en su posición, pero... ¿aparecerá Miguelcldn en la lista de developers al presionar F5?
|
1.0
|
Update developer list - Caso de estudio:
Miguelcldn es un nuevo developer que quiere entrar en Fermat, nunca tuvo una ficha asignada.
- Entra en la página
- Inicia sesión
- Crea un componente
- Pone "miguelcldn" en el nombre del developer
- Se crea la ficha y se pone en su posición, pero... ¿aparecerá Miguelcldn en la lista de developers al presionar F5?
|
priority
|
update developer list caso de estudio miguelcldn es un nuevo developer que quiere entrar en fermat nunca tuvo una ficha asignada entra en la página inicia sesión crea un componente pone miguelcldn en el nombre del developer se crea la ficha y se pone en su posición pero ¿aparecerá miguelcldn en la lista de developers al presionar
| 1
|
403,934
| 11,849,278,852
|
IssuesEvent
|
2020-03-24 15:01:02
|
Monika-After-Story/MonikaModDev
|
https://api.github.com/repos/Monika-After-Story/MonikaModDev
|
closed
|
In-game explanation of gifting mechanic
|
enhancement high priority
|
A quick search of code shows no instance of "characters folder" or "characters/" in any dialogue relating to gifting, which implies there is no in-game explanation of the gifting mechanic. This is probably why new players don't understand or miss out on things as it was initially introduced in a 9-22 event as an external file.
We should have some in-game, repeatable explanation of this mechanic (story-events, with a say prompt like "How do I gift you things?"). This is especially important given consumables.
This should be also noted in intro, like how we explain hotkeys and games.
Additionally, I'd say an update script to queue the topic is unnecessary, a new prompt will draw attention in unseen anyway.
@multimokia I am assigning this to you but you can triage as you see fit. Just get this in before next release.
|
1.0
|
In-game explanation of gifting mechanic - A quick search of code shows no instance of "characters folder" or "characters/" in any dialogue relating to gifting, which implies there is no in-game explanation of the gifting mechanic. This is probably why new players don't understand or miss out on things as it was initially introduced in a 9-22 event as an external file.
We should have some in-game, repeatable explanation of this mechanic (story-events, with a say prompt like "How do I gift you things?"). This is especially important given consumables.
This should be also noted in intro, like how we explain hotkeys and games.
Additionally, I'd say an update script to queue the topic is unnecessary, a new prompt will draw attention in unseen anyway.
@multimokia I am assigning this to you but you can triage as you see fit. Just get this in before next release.
|
priority
|
in game explanation of gifting mechanic a quick search of code shows no instance of characters folder or characters in any dialogue relating to gifting which implies there is no in game explanation of the gifting mechanic this is probably why new players don t understand or miss out on things as it was initially introduced in a event as an external file we should have some in game repeatable explanation of this mechanic story events with a say prompt like how do i gift you things this is especially important given consumables this should be also noted in intro like how we explain hotkeys and games additionally i d say an update script to queue the topic is unnecessary a new prompt will draw attention in unseen anyway multimokia i am assigning this to you but you can triage as you see fit just get this in before next release
| 1
|
321,773
| 9,808,860,529
|
IssuesEvent
|
2019-06-12 16:31:52
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
opened
|
IS : Custom claims lost when updating user profile
|
Complexity/High Component/OAuth Priority/High
|
Moved from: https://wso2.org/jira/browse/IDENTITY-7335
Hi WSO2 team,
WSO2 Identity Server offers some extension points for instance ability to add custom claims.
This is normally donc by writing a custom claim handler as described in one wso2 post :
http://pushpalankajaya.blogspot.com/2014/07/adding-custom-claims-to-saml-response.html
When using WSO2 OAuth2.0 Playground sample and OAuth2.0 Authorization code grant type, custom claim handler is invoked normally when resource owner is authenticated.
Then at the end of process you get an access token enabling to retrieve, for instance, user info with customs claim ("computed before when custom claim handler is invoked).
If you update some information of user profile and request again user info from user info endpoint with access_token, all the +custom + claims are lost in provided response.
Actually, when you update user profile, information that were previously cached are emptied so when user info endpoint is called with access_token, user claims (and their corresponding values) are computed from userstore perspective only so that why custom claims are lost.
Would it be possible in that scenario of retrieving claims from userstore that custom claim handler would be also called to ensure that customs claims are available in answer?
Regards,
Franck
|
1.0
|
IS : Custom claims lost when updating user profile - Moved from: https://wso2.org/jira/browse/IDENTITY-7335
Hi WSO2 team,
WSO2 Identity Server offers some extension points for instance ability to add custom claims.
This is normally donc by writing a custom claim handler as described in one wso2 post :
http://pushpalankajaya.blogspot.com/2014/07/adding-custom-claims-to-saml-response.html
When using WSO2 OAuth2.0 Playground sample and OAuth2.0 Authorization code grant type, custom claim handler is invoked normally when resource owner is authenticated.
Then at the end of process you get an access token enabling to retrieve, for instance, user info with customs claim ("computed before when custom claim handler is invoked).
If you update some information of user profile and request again user info from user info endpoint with access_token, all the +custom + claims are lost in provided response.
Actually, when you update user profile, information that were previously cached are emptied so when user info endpoint is called with access_token, user claims (and their corresponding values) are computed from userstore perspective only so that why custom claims are lost.
Would it be possible in that scenario of retrieving claims from userstore that custom claim handler would be also called to ensure that customs claims are available in answer?
Regards,
Franck
|
priority
|
is custom claims lost when updating user profile moved from hi team identity server offers some extension points for instance ability to add custom claims this is normally donc by writing a custom claim handler as described in one post when using playground sample and authorization code grant type custom claim handler is invoked normally when resource owner is authenticated then at the end of process you get an access token enabling to retrieve for instance user info with customs claim computed before when custom claim handler is invoked if you update some information of user profile and request again user info from user info endpoint with access token all the custom claims are lost in provided response actually when you update user profile information that were previously cached are emptied so when user info endpoint is called with access token user claims and their corresponding values are computed from userstore perspective only so that why custom claims are lost would it be possible in that scenario of retrieving claims from userstore that custom claim handler would be also called to ensure that customs claims are available in answer regards franck
| 1
|
567,375
| 16,857,149,375
|
IssuesEvent
|
2021-06-21 08:19:34
|
FEDMix/eshmun
|
https://api.github.com/repos/FEDMix/eshmun
|
closed
|
Orthogonal view to view DICOM Images
|
Brachytherapy FEDmix High Priority Modir User Story
|
## Story
As a Clinician
I want to view DICOM images of a selected patient
So that I can see a preview of the choice I am trying to make
## Proposed work
Within the application it should be possible to view DICOM images.
There should be 3 views that are orthogonal to each other (x, y and z directions).
The user should be able to pan and zoom the views.
## Acceptance Criteria
- [x] 3 Orthogonal Views
- [x] Panning & Zooming
## Designs

|
1.0
|
Orthogonal view to view DICOM Images - ## Story
As a Clinician
I want to view DICOM images of a selected patient
So that I can see a preview of the choice I am trying to make
## Proposed work
Within the application it should be possible to view DICOM images.
There should be 3 views that are orthogonal to each other (x, y and z directions).
The user should be able to pan and zoom the views.
## Acceptance Criteria
- [x] 3 Orthogonal Views
- [x] Panning & Zooming
## Designs

|
priority
|
orthogonal view to view dicom images story as a clinician i want to view dicom images of a selected patient so that i can see a preview of the choice i am trying to make proposed work within the application it should be possible to view dicom images there should be views that are orthogonal to each other x y and z directions the user should be able to pan and zoom the views acceptance criteria orthogonal views panning zooming designs
| 1
|
692,450
| 23,735,308,479
|
IssuesEvent
|
2022-08-31 07:33:57
|
vignetteapp/sekai
|
https://api.github.com/repos/vignetteapp/sekai
|
closed
|
Implement a basic rendering system
|
enhancement priority:high
|
As our rendering engine is primarily for 3D, we'd need a basic rendering system that shows 3D models with lighting and shadows.
|
1.0
|
Implement a basic rendering system - As our rendering engine is primarily for 3D, we'd need a basic rendering system that shows 3D models with lighting and shadows.
|
priority
|
implement a basic rendering system as our rendering engine is primarily for we d need a basic rendering system that shows models with lighting and shadows
| 1
|
404,490
| 11,858,245,224
|
IssuesEvent
|
2020-03-25 11:04:21
|
AY1920S2-CS2103T-F10-2/main
|
https://api.github.com/repos/AY1920S2-CS2103T-F10-2/main
|
closed
|
Implement Reminder feature
|
priority.High type.Epic
|
Update the default list view of the Internship Diary to be shown according to urgency of the application, in terms of the application deadline followed by interview(s) date(s).
|
1.0
|
Implement Reminder feature - Update the default list view of the Internship Diary to be shown according to urgency of the application, in terms of the application deadline followed by interview(s) date(s).
|
priority
|
implement reminder feature update the default list view of the internship diary to be shown according to urgency of the application in terms of the application deadline followed by interview s date s
| 1
|
229,631
| 7,582,167,904
|
IssuesEvent
|
2018-04-25 02:29:17
|
Myoats/preprod
|
https://api.github.com/repos/Myoats/preprod
|
opened
|
The Nav on small on medium devices on the should just be the hamburger
|
Highest Priority
|
It should look like this

|
1.0
|
The Nav on small on medium devices on the should just be the hamburger - It should look like this

|
priority
|
the nav on small on medium devices on the should just be the hamburger it should look like this
| 1
|
359,071
| 10,659,685,809
|
IssuesEvent
|
2019-10-18 08:17:45
|
AY1920S1-CS2113T-W13-2/main
|
https://api.github.com/repos/AY1920S1-CS2113T-W13-2/main
|
opened
|
Create Review function for Quiz
|
priority.High type.Task
|
Allows user to review their answers and the correct answer at the end of a quiz session (CLI).
|
1.0
|
Create Review function for Quiz - Allows user to review their answers and the correct answer at the end of a quiz session (CLI).
|
priority
|
create review function for quiz allows user to review their answers and the correct answer at the end of a quiz session cli
| 1
|
86,978
| 3,735,649,127
|
IssuesEvent
|
2016-03-08 13:08:54
|
asterics/AsTeRICS
|
https://api.github.com/repos/asterics/AsTeRICS
|
closed
|
JNativeHook error message when starting two instances of AsTeRICS
|
high priority
|
When starting two instances of AsTeRICS the jnativehook service bundle cannot be activated and an error message is thrown.
The debug console shows the following root exception:
Caused by: java.lang.RuntimeException: C:\Users\mad\AppData\Local\Temp\JNativeHook-1.2.RC2.dll (Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird)
at org.jnativehook.GlobalScreen.loadNativeLibrary(GlobalScreen.java:609)
at org.jnativehook.GlobalScreen.<init>(GlobalScreen.java:86)
at org.jnativehook.GlobalScreen.<clinit>(GlobalScreen.java:67)
... 28 more
Caused by: java.io.FileNotFoundException: C:\Users\mad\AppData\Local\Temp\JNativeHook-1.2.RC2.dll (Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird)
at java.io.FileOutputStream.open(Native Method)
at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
at java.io.FileOutputStream.<init>(FileOutputStream.java:171)
at org.jnativehook.GlobalScreen.loadNativeLibrary(GlobalScreen.java:570)
... 30 more
Obviously the native dll file is busy and cannot be accessed two times.
Possible solutions:
upgrade to newer version of jnativehook, maybe it's already fixed?
workaround:
Manually delete .dll file after startup of bundle.
|
1.0
|
JNativeHook error message when starting two instances of AsTeRICS - When starting two instances of AsTeRICS the jnativehook service bundle cannot be activated and an error message is thrown.
The debug console shows the following root exception:
Caused by: java.lang.RuntimeException: C:\Users\mad\AppData\Local\Temp\JNativeHook-1.2.RC2.dll (Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird)
at org.jnativehook.GlobalScreen.loadNativeLibrary(GlobalScreen.java:609)
at org.jnativehook.GlobalScreen.<init>(GlobalScreen.java:86)
at org.jnativehook.GlobalScreen.<clinit>(GlobalScreen.java:67)
... 28 more
Caused by: java.io.FileNotFoundException: C:\Users\mad\AppData\Local\Temp\JNativeHook-1.2.RC2.dll (Der Prozess kann nicht auf die Datei zugreifen, da sie von einem anderen Prozess verwendet wird)
at java.io.FileOutputStream.open(Native Method)
at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
at java.io.FileOutputStream.<init>(FileOutputStream.java:171)
at org.jnativehook.GlobalScreen.loadNativeLibrary(GlobalScreen.java:570)
... 30 more
Obviously the native dll file is busy and cannot be accessed two times.
Possible solutions:
upgrade to newer version of jnativehook, maybe it's already fixed?
workaround:
Manually delete .dll file after startup of bundle.
|
priority
|
jnativehook error message when starting two instances of asterics when starting two instances of asterics the jnativehook service bundle cannot be activated and an error message is thrown the debug console shows the following root exception caused by java lang runtimeexception c users mad appdata local temp jnativehook dll der prozess kann nicht auf die datei zugreifen da sie von einem anderen prozess verwendet wird at org jnativehook globalscreen loadnativelibrary globalscreen java at org jnativehook globalscreen globalscreen java at org jnativehook globalscreen globalscreen java more caused by java io filenotfoundexception c users mad appdata local temp jnativehook dll der prozess kann nicht auf die datei zugreifen da sie von einem anderen prozess verwendet wird at java io fileoutputstream open native method at java io fileoutputstream fileoutputstream java at java io fileoutputstream fileoutputstream java at org jnativehook globalscreen loadnativelibrary globalscreen java more obviously the native dll file is busy and cannot be accessed two times possible solutions upgrade to newer version of jnativehook maybe it s already fixed workaround manually delete dll file after startup of bundle
| 1
|
725,264
| 24,956,556,151
|
IssuesEvent
|
2022-11-01 12:18:11
|
AY2223S1-CS2113-T18-1b/tp
|
https://api.github.com/repos/AY2223S1-CS2113-T18-1b/tp
|
closed
|
[PE-D][Tester C] sort should come configurable ordering to make sense - i.e. asc or desc. Otherwise, fix current defaults as some orderings do not make sense
|
priority.High
|
For example, the current sort for review is ascending, which is unusual as most people sort shows by descending rating
<!--session: 1666946737319-de45f0f8-5014-4481-873f-b0f1008853e3--><!--Version: Web v3.4.4-->
-------------
Labels: `type.FeatureFlaw` `severity.Low`
original: winston-lim/ped#7
|
1.0
|
[PE-D][Tester C] sort should come configurable ordering to make sense - i.e. asc or desc. Otherwise, fix current defaults as some orderings do not make sense - For example, the current sort for review is ascending, which is unusual as most people sort shows by descending rating
<!--session: 1666946737319-de45f0f8-5014-4481-873f-b0f1008853e3--><!--Version: Web v3.4.4-->
-------------
Labels: `type.FeatureFlaw` `severity.Low`
original: winston-lim/ped#7
|
priority
|
sort should come configurable ordering to make sense i e asc or desc otherwise fix current defaults as some orderings do not make sense for example the current sort for review is ascending which is unusual as most people sort shows by descending rating labels type featureflaw severity low original winston lim ped
| 1
|
351,271
| 10,514,749,872
|
IssuesEvent
|
2019-09-28 03:17:59
|
xournalpp/xournalpp
|
https://api.github.com/repos/xournalpp/xournalpp
|
closed
|
Crash on opening files.
|
Crash PR available bug confirmed difficulty:easy priority: high
|
**Affects versions :**
- OS: [Linux Mint 19.2 Cinnamon]
- Version of Xournal++ [current master]
**Describe the bug**
Pressing the open button crashes the software
**To Reproduce**
compile with clang (in DEBUG Mode), open xournal++ try to open a file
**Expected behavior**
Do not crash.
fixed it already in a local branch, all `std::strings` in `Settings` are returned by value thus they only live after the line if not assigned. There was a .c_str() call on a temporary which is destructed immediately.
Replaced all returns of a `std::string` by value in `Settings` with a `string const&`.
PR following.
|
1.0
|
Crash on opening files. - **Affects versions :**
- OS: [Linux Mint 19.2 Cinnamon]
- Version of Xournal++ [current master]
**Describe the bug**
Pressing the open button crashes the software
**To Reproduce**
compile with clang (in DEBUG Mode), open xournal++ try to open a file
**Expected behavior**
Do not crash.
fixed it already in a local branch, all `std::strings` in `Settings` are returned by value thus they only live after the line if not assigned. There was a .c_str() call on a temporary which is destructed immediately.
Replaced all returns of a `std::string` by value in `Settings` with a `string const&`.
PR following.
|
priority
|
crash on opening files affects versions os version of xournal describe the bug pressing the open button crashes the software to reproduce compile with clang in debug mode open xournal try to open a file expected behavior do not crash fixed it already in a local branch all std strings in settings are returned by value thus they only live after the line if not assigned there was a c str call on a temporary which is destructed immediately replaced all returns of a std string by value in settings with a string const pr following
| 1
|
204,129
| 7,084,095,754
|
IssuesEvent
|
2018-01-11 04:32:06
|
EFForg/privacybadger
|
https://api.github.com/repos/EFForg/privacybadger
|
reopened
|
Error reports for www.youtube.com
|
broken site high priority
|
We are getting lots of error reports for `www.youtube.com`; it's our most-reported-on domain for the latest Privacy Badger (version 2017.7.24).
Here is some of what our users say:
>video won't load
>Youtube video shows: 'Request blocked by extension'. Fixed by switching on s0.2mdn.net
>video said an extension blocked requests to the server
>Videos laden anfangs nicht - Meldung "Anfragen an den Server wurden durch Erweiterung blockiert"
>Ads don't appear instead it says an extension is blocking the ad content. Ad blocks still appear just without the actual ad.
>preroll Advertisement on Youtube "blocked by an extension".
|
1.0
|
Error reports for www.youtube.com - We are getting lots of error reports for `www.youtube.com`; it's our most-reported-on domain for the latest Privacy Badger (version 2017.7.24).
Here is some of what our users say:
>video won't load
>Youtube video shows: 'Request blocked by extension'. Fixed by switching on s0.2mdn.net
>video said an extension blocked requests to the server
>Videos laden anfangs nicht - Meldung "Anfragen an den Server wurden durch Erweiterung blockiert"
>Ads don't appear instead it says an extension is blocking the ad content. Ad blocks still appear just without the actual ad.
>preroll Advertisement on Youtube "blocked by an extension".
|
priority
|
error reports for we are getting lots of error reports for it s our most reported on domain for the latest privacy badger version here is some of what our users say video won t load youtube video shows request blocked by extension fixed by switching on net video said an extension blocked requests to the server videos laden anfangs nicht meldung anfragen an den server wurden durch erweiterung blockiert ads don t appear instead it says an extension is blocking the ad content ad blocks still appear just without the actual ad preroll advertisement on youtube blocked by an extension
| 1
|
810,681
| 30,254,112,094
|
IssuesEvent
|
2023-07-07 00:00:59
|
hwgilbert16/scholarsome
|
https://api.github.com/repos/hwgilbert16/scholarsome
|
closed
|
Exclude HTML tags from written response answers
|
high priority
|
Written response quiz questions check for exact strings. Cards created using the new rich text editor include `<p>` tags, which are checked for.
These need to be removed before creating a quiz, or we need to filter the set beforehand to exclude these from being quizzed on.
|
1.0
|
Exclude HTML tags from written response answers - Written response quiz questions check for exact strings. Cards created using the new rich text editor include `<p>` tags, which are checked for.
These need to be removed before creating a quiz, or we need to filter the set beforehand to exclude these from being quizzed on.
|
priority
|
exclude html tags from written response answers written response quiz questions check for exact strings cards created using the new rich text editor include tags which are checked for these need to be removed before creating a quiz or we need to filter the set beforehand to exclude these from being quizzed on
| 1
|
727,942
| 25,060,630,893
|
IssuesEvent
|
2022-11-07 00:56:45
|
AY2223S1-CS2103T-W15-4/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-W15-4/tp
|
closed
|
[PE-D][Tester A] Mark INDEX marked 2 students
|
type.Bug priority.High severity.High
|
1. add n/DavId Li
2. add n/Davlid Lin
3. sort t/d
4. mark 1 will pass mastery check for both
5. unmark similarly removes pass for both
<i><video controls><source src="https://raw.githubusercontent.com/KSHan29/ped/main/files/024d6dca-4e30-48ff-a3c0-15dbe2a8bf03.mov" type="video/mp4">Your browser does not support the video tag.</video><br>video:https://raw.githubusercontent.com/KSHan29/ped/main/files/024d6dca-4e30-48ff-a3c0-15dbe2a8bf03.mov</i>
<!--session: 1666944123338-1a1c6713-a846-47ba-a447-d3dd287bc156-->
<!--Version: Web v3.4.4-->
-------------
Labels: `severity.Medium` `type.FunctionalityBug`
original: KSHan29/ped#1
|
1.0
|
[PE-D][Tester A] Mark INDEX marked 2 students - 1. add n/DavId Li
2. add n/Davlid Lin
3. sort t/d
4. mark 1 will pass mastery check for both
5. unmark similarly removes pass for both
<i><video controls><source src="https://raw.githubusercontent.com/KSHan29/ped/main/files/024d6dca-4e30-48ff-a3c0-15dbe2a8bf03.mov" type="video/mp4">Your browser does not support the video tag.</video><br>video:https://raw.githubusercontent.com/KSHan29/ped/main/files/024d6dca-4e30-48ff-a3c0-15dbe2a8bf03.mov</i>
<!--session: 1666944123338-1a1c6713-a846-47ba-a447-d3dd287bc156-->
<!--Version: Web v3.4.4-->
-------------
Labels: `severity.Medium` `type.FunctionalityBug`
original: KSHan29/ped#1
|
priority
|
mark index marked students add n david li add n davlid lin sort t d mark will pass mastery check for both unmark similarly removes pass for both your browser does not support the video tag video labels severity medium type functionalitybug original ped
| 1
|
402,408
| 11,809,307,977
|
IssuesEvent
|
2020-03-19 14:46:01
|
wso2/docker-is
|
https://api.github.com/repos/wso2/docker-is
|
closed
|
Use WSO2 product pack downloadable links to binaries available at GitHub release pages
|
Priority/Highest Type/Task
|
**Description:**
Currently, WSO2 product binaries available at JFrog Bintray are used for the Docker image builds.
It has been suggested to use WSO2 product pack downloadable links to binaries available at GitHub product release pages, as the default.
**Affected Product Version:**
Docker resources for WSO2 IAM v5.10.0 and beyond
**Sub Tasks:**
- [x] Integrate to Alpine based Docker resources
- [x] Integrate to CentOS based Docker resources
- [x] Integrate to Ubuntu based Docker resources
- [x] Code/Peer review
|
1.0
|
Use WSO2 product pack downloadable links to binaries available at GitHub release pages - **Description:**
Currently, WSO2 product binaries available at JFrog Bintray are used for the Docker image builds.
It has been suggested to use WSO2 product pack downloadable links to binaries available at GitHub product release pages, as the default.
**Affected Product Version:**
Docker resources for WSO2 IAM v5.10.0 and beyond
**Sub Tasks:**
- [x] Integrate to Alpine based Docker resources
- [x] Integrate to CentOS based Docker resources
- [x] Integrate to Ubuntu based Docker resources
- [x] Code/Peer review
|
priority
|
use product pack downloadable links to binaries available at github release pages description currently product binaries available at jfrog bintray are used for the docker image builds it has been suggested to use product pack downloadable links to binaries available at github product release pages as the default affected product version docker resources for iam and beyond sub tasks integrate to alpine based docker resources integrate to centos based docker resources integrate to ubuntu based docker resources code peer review
| 1
|
205,670
| 7,104,589,060
|
IssuesEvent
|
2018-01-16 10:29:33
|
AnSyn/ansyn
|
https://api.github.com/repos/AnSyn/ansyn
|
closed
|
Bug- shadow mouse- cursor on inactive screen
|
Bug Priority: High Severity: Medium
|
**Current behavior**
When hovering an inactive screen with the cursor- the user can see both the shadow mouse cross and the cursor in the same screen.
**Expected behavior**
the shadow mouse cross should not appear, so the user wont be confused.
**Minimal reproduction of the problem with instructions**
open more then one screen
activate shadow mouse
hover the inactive screen
|
1.0
|
Bug- shadow mouse- cursor on inactive screen -
**Current behavior**
When hovering an inactive screen with the cursor- the user can see both the shadow mouse cross and the cursor in the same screen.
**Expected behavior**
the shadow mouse cross should not appear, so the user wont be confused.
**Minimal reproduction of the problem with instructions**
open more then one screen
activate shadow mouse
hover the inactive screen
|
priority
|
bug shadow mouse cursor on inactive screen current behavior when hovering an inactive screen with the cursor the user can see both the shadow mouse cross and the cursor in the same screen expected behavior the shadow mouse cross should not appear so the user wont be confused minimal reproduction of the problem with instructions open more then one screen activate shadow mouse hover the inactive screen
| 1
|
30,980
| 2,730,528,396
|
IssuesEvent
|
2015-04-16 15:20:01
|
Esri/briefing-book
|
https://api.github.com/repos/Esri/briefing-book
|
closed
|
Comment the codebase
|
bug develop21drop High Priority
|
Aside from the config.js, the majority of the code base isn't commented.
Please provide quality commenting within the code base.
|
1.0
|
Comment the codebase - Aside from the config.js, the majority of the code base isn't commented.
Please provide quality commenting within the code base.
|
priority
|
comment the codebase aside from the config js the majority of the code base isn t commented please provide quality commenting within the code base
| 1
|
481,243
| 13,882,518,034
|
IssuesEvent
|
2020-10-18 07:23:11
|
gileli121/WindowTop
|
https://api.github.com/repos/gileli121/WindowTop
|
closed
|
Modern toolbar UI is not working after saving window configuration on window with Shrink mode
|
bug fixed high-priority
|
A user reported that the new toolbar UI is not showing even when enabling the modern toolbar feature.
I asked the user for the WindowTop.settings file, and he sent it.
I found that the configuration that saved in this file was reproduced the issue.
While debugging the issue, I found that the configuration key that reproduces it is:
```
[WindowsSettings]
=ArrowXPos=0.45
```
It should not be like that. it should be:
```
[WindowsSettings]
some_process.exe=ArrowXPos=0.45
```
I reviewed the code, and according to the code review, this configuration issue is caused in the following scenario:
1. Disable the modern toolbar
2. Shrink some window
3. Resize the shrink box so the arrow button will show when you move to the top area...
4. Drag the arrow button to some another position
5. Unshrink the window or exit WindowTop
At this point, this configuration issue will reproduce.
Because this configuration issue affects only the modern toolbar UI, you will need to enable the modern toolbar UI feature to reproduce it.
Seems that the user did this scenario in older version or he did it before enabaling the modern toolbar feature.
|
1.0
|
Modern toolbar UI is not working after saving window configuration on window with Shrink mode - A user reported that the new toolbar UI is not showing even when enabling the modern toolbar feature.
I asked the user for the WindowTop.settings file, and he sent it.
I found that the configuration that saved in this file was reproduced the issue.
While debugging the issue, I found that the configuration key that reproduces it is:
```
[WindowsSettings]
=ArrowXPos=0.45
```
It should not be like that. it should be:
```
[WindowsSettings]
some_process.exe=ArrowXPos=0.45
```
I reviewed the code, and according to the code review, this configuration issue is caused in the following scenario:
1. Disable the modern toolbar
2. Shrink some window
3. Resize the shrink box so the arrow button will show when you move to the top area...
4. Drag the arrow button to some another position
5. Unshrink the window or exit WindowTop
At this point, this configuration issue will reproduce.
Because this configuration issue affects only the modern toolbar UI, you will need to enable the modern toolbar UI feature to reproduce it.
Seems that the user did this scenario in older version or he did it before enabaling the modern toolbar feature.
|
priority
|
modern toolbar ui is not working after saving window configuration on window with shrink mode a user reported that the new toolbar ui is not showing even when enabling the modern toolbar feature i asked the user for the windowtop settings file and he sent it i found that the configuration that saved in this file was reproduced the issue while debugging the issue i found that the configuration key that reproduces it is arrowxpos it should not be like that it should be some process exe arrowxpos i reviewed the code and according to the code review this configuration issue is caused in the following scenario disable the modern toolbar shrink some window resize the shrink box so the arrow button will show when you move to the top area drag the arrow button to some another position unshrink the window or exit windowtop at this point this configuration issue will reproduce because this configuration issue affects only the modern toolbar ui you will need to enable the modern toolbar ui feature to reproduce it seems that the user did this scenario in older version or he did it before enabaling the modern toolbar feature
| 1
|
804,372
| 29,485,251,542
|
IssuesEvent
|
2023-06-02 09:14:22
|
Avaiga/taipy-core
|
https://api.github.com/repos/Avaiga/taipy-core
|
closed
|
Event: fire an event when scenario name is modified
|
🟧 Priority: High 📈 Improvement
|
**Description**
An event should be fired when scenario name is modified.
|
1.0
|
Event: fire an event when scenario name is modified - **Description**
An event should be fired when scenario name is modified.
|
priority
|
event fire an event when scenario name is modified description an event should be fired when scenario name is modified
| 1
|
708,856
| 24,357,537,377
|
IssuesEvent
|
2022-10-03 08:49:19
|
NethermindEth/nethermind
|
https://api.github.com/repos/NethermindEth/nethermind
|
closed
|
[Peers] When Peers number drop to 0, nethermind have problems to recover
|
high priority stability
|
**Describe the bug**
After some network issues peers number drops to 0. For about 20 minutes it was logging "Waiting for Peers" and nothing interesting happened. After restart of nethermind app, peers came back properly.
**To Reproduce**
Steps to reproduce the behavior:
Disconnect the network for a moment and check if peers dropped to 0 and then recover network connection.
**Expected behavior**
After reconnecting we should be able to recover and find peers back.
**Additional context**
Add any other context about the problem here.
|
1.0
|
[Peers] When Peers number drop to 0, nethermind have problems to recover - **Describe the bug**
After some network issues peers number drops to 0. For about 20 minutes it was logging "Waiting for Peers" and nothing interesting happened. After restart of nethermind app, peers came back properly.
**To Reproduce**
Steps to reproduce the behavior:
Disconnect the network for a moment and check if peers dropped to 0 and then recover network connection.
**Expected behavior**
After reconnecting we should be able to recover and find peers back.
**Additional context**
Add any other context about the problem here.
|
priority
|
when peers number drop to nethermind have problems to recover describe the bug after some network issues peers number drops to for about minutes it was logging waiting for peers and nothing interesting happened after restart of nethermind app peers came back properly to reproduce steps to reproduce the behavior disconnect the network for a moment and check if peers dropped to and then recover network connection expected behavior after reconnecting we should be able to recover and find peers back additional context add any other context about the problem here
| 1
|
321,249
| 9,795,819,627
|
IssuesEvent
|
2019-06-11 05:36:34
|
apache/skywalking
|
https://api.github.com/repos/apache/skywalking
|
opened
|
Time series ElasticSearch implementation bug of record type
|
OAP-backend bug high priority
|
As the following screenshot shows, the segment(also alarm and other record types) don't create index by day, but in hour wrongly.

This needs to be fixed and make sure the TTL works for it(not working today, possible because of wrong table name).
|
1.0
|
Time series ElasticSearch implementation bug of record type - As the following screenshot shows, the segment(also alarm and other record types) don't create index by day, but in hour wrongly.

This needs to be fixed and make sure the TTL works for it(not working today, possible because of wrong table name).
|
priority
|
time series elasticsearch implementation bug of record type as the following screenshot shows the segment also alarm and other record types don t create index by day but in hour wrongly this needs to be fixed and make sure the ttl works for it not working today possible because of wrong table name
| 1
|
409,446
| 11,962,589,959
|
IssuesEvent
|
2020-04-05 13:01:11
|
traffic-control-fyp-aub/ns3-gym
|
https://api.github.com/repos/traffic-control-fyp-aub/ns3-gym
|
closed
|
System Avg Speed [ Performance ]
|
High Priority
|
Study the effect of different vehicle control percentages (incorporate human driver control) on the system wide average speed and whether it increases or decreases.
|
1.0
|
System Avg Speed [ Performance ] - Study the effect of different vehicle control percentages (incorporate human driver control) on the system wide average speed and whether it increases or decreases.
|
priority
|
system avg speed study the effect of different vehicle control percentages incorporate human driver control on the system wide average speed and whether it increases or decreases
| 1
|
440,283
| 12,697,315,600
|
IssuesEvent
|
2020-06-22 11:36:19
|
chiyadev/MudaeFarm
|
https://api.github.com/repos/chiyadev/MudaeFarm
|
closed
|
Bot breaks and spams chat
|
bug high priority
|
At some point the farm falls apart and fails to parse anything Mudaebot sends, even normal rolls. It seems insane to me that MudaeFarm would try every ~30 seconds indefinitely and create hundreds of messages. Unsanitized logs below, timestamp is 2020-06-19T21:59:41.2842398-05:00
[log_2020-06-19 06.22.24Z-20200619.txt](https://github.com/chiyadev/MudaeFarm/files/4807768/log_2020-06-19.06.22.24Z-20200619.txt)
EDIT: I just realized these logs aren't verbose. Given it's been happening every night, I'll have some soon
|
1.0
|
Bot breaks and spams chat - At some point the farm falls apart and fails to parse anything Mudaebot sends, even normal rolls. It seems insane to me that MudaeFarm would try every ~30 seconds indefinitely and create hundreds of messages. Unsanitized logs below, timestamp is 2020-06-19T21:59:41.2842398-05:00
[log_2020-06-19 06.22.24Z-20200619.txt](https://github.com/chiyadev/MudaeFarm/files/4807768/log_2020-06-19.06.22.24Z-20200619.txt)
EDIT: I just realized these logs aren't verbose. Given it's been happening every night, I'll have some soon
|
priority
|
bot breaks and spams chat at some point the farm falls apart and fails to parse anything mudaebot sends even normal rolls it seems insane to me that mudaefarm would try every seconds indefinitely and create hundreds of messages unsanitized logs below timestamp is edit i just realized these logs aren t verbose given it s been happening every night i ll have some soon
| 1
|
678,068
| 23,186,038,094
|
IssuesEvent
|
2022-08-01 08:28:08
|
netdata/netdata-cloud
|
https://api.github.com/repos/netdata/netdata-cloud
|
closed
|
Enhance the Alerts drawer / modal in the Active alerts and Alerts Configuration Page
|
priority/high cloud-frontend cloud-backend alerts-team Q2 GOAL feature request
|
### Problem
The current Alerts drawer / modal does not display the details of node instances where this alert has been raised and how it has been aggregated at the node level.
### Description
The alert drawer is currently being used in the Active Alerts tab and will need to be used also in the "Alerts Configuration" / "Manage Alerts" tab. The drawer / model needs to be enhanced to provide more clear information on the alert itself:
- The Chart relevant to the alert (already available)
- Details of the alert event as a vertical bar with value (already available for Warning and Critical) also for the Clear state.
- Table with Node instances where the alert was raised (with criticality)
- Alert configurations in a more easily readable and understandable way for each node instance
- CTA to view the alert configuration at the node level (on a dedicated page?)
This can possibly be extended to also show historical logs relevant to the specific alert within the drawer.
cc: @ktsaou @amalkov @vinnygats @car12o @YaroslavDev @novykh @jacekkolasa
### Importance
must have
### Value proposition
1. The current alert drawer does not convey a clear description of the alert
2. The user will need to see details at the node instance level
3. Possible future extension to also show the historical logs relevant to the alert
...
### Proposed implementation
_No response_
|
1.0
|
Enhance the Alerts drawer / modal in the Active alerts and Alerts Configuration Page - ### Problem
The current Alerts drawer / modal does not display the details of node instances where this alert has been raised and how it has been aggregated at the node level.
### Description
The alert drawer is currently being used in the Active Alerts tab and will need to be used also in the "Alerts Configuration" / "Manage Alerts" tab. The drawer / model needs to be enhanced to provide more clear information on the alert itself:
- The Chart relevant to the alert (already available)
- Details of the alert event as a vertical bar with value (already available for Warning and Critical) also for the Clear state.
- Table with Node instances where the alert was raised (with criticality)
- Alert configurations in a more easily readable and understandable way for each node instance
- CTA to view the alert configuration at the node level (on a dedicated page?)
This can possibly be extended to also show historical logs relevant to the specific alert within the drawer.
cc: @ktsaou @amalkov @vinnygats @car12o @YaroslavDev @novykh @jacekkolasa
### Importance
must have
### Value proposition
1. The current alert drawer does not convey a clear description of the alert
2. The user will need to see details at the node instance level
3. Possible future extension to also show the historical logs relevant to the alert
...
### Proposed implementation
_No response_
|
priority
|
enhance the alerts drawer modal in the active alerts and alerts configuration page problem the current alerts drawer modal does not display the details of node instances where this alert has been raised and how it has been aggregated at the node level description the alert drawer is currently being used in the active alerts tab and will need to be used also in the alerts configuration manage alerts tab the drawer model needs to be enhanced to provide more clear information on the alert itself the chart relevant to the alert already available details of the alert event as a vertical bar with value already available for warning and critical also for the clear state table with node instances where the alert was raised with criticality alert configurations in a more easily readable and understandable way for each node instance cta to view the alert configuration at the node level on a dedicated page this can possibly be extended to also show historical logs relevant to the specific alert within the drawer cc ktsaou amalkov vinnygats yaroslavdev novykh jacekkolasa importance must have value proposition the current alert drawer does not convey a clear description of the alert the user will need to see details at the node instance level possible future extension to also show the historical logs relevant to the alert proposed implementation no response
| 1
|
528,619
| 15,370,896,137
|
IssuesEvent
|
2021-03-02 09:22:16
|
enthought/enable
|
https://api.github.com/repos/enthought/enable
|
closed
|
draw_rect behaves weirdly with QPainter backend
|
priority: high type: bug
|
|agg|QPainter|
|---|---|
|<img width="256" alt="kiva agg draw_rect 2x" src="https://user-images.githubusercontent.com/1926457/109619792-1623f000-7b31-11eb-90e4-b5590aba5d9c.png">|<img width="256" alt="qpainter draw_rect 2x" src="https://user-images.githubusercontent.com/1926457/109619815-18864a00-7b31-11eb-900d-cae1da311d98.png">|
The above images have been generated using the new `enable.gcbench` on windows with PyQt5.
|
1.0
|
draw_rect behaves weirdly with QPainter backend - |agg|QPainter|
|---|---|
|<img width="256" alt="kiva agg draw_rect 2x" src="https://user-images.githubusercontent.com/1926457/109619792-1623f000-7b31-11eb-90e4-b5590aba5d9c.png">|<img width="256" alt="qpainter draw_rect 2x" src="https://user-images.githubusercontent.com/1926457/109619815-18864a00-7b31-11eb-900d-cae1da311d98.png">|
The above images have been generated using the new `enable.gcbench` on windows with PyQt5.
|
priority
|
draw rect behaves weirdly with qpainter backend agg qpainter img width alt kiva agg draw rect src width alt qpainter draw rect src the above images have been generated using the new enable gcbench on windows with
| 1
|
413,788
| 12,092,155,253
|
IssuesEvent
|
2020-04-19 14:35:40
|
lorenzwalthert/precommit
|
https://api.github.com/repos/lorenzwalthert/precommit
|
closed
|
Allow to choose installation environment
|
Complexity: Medium Priority: High Status: WIP Type: Enhancement
|
As mentioned in [#113 ](https://github.com/lorenzwalthert/precommit/issues/113#issuecomment-603808455). Recently I've tried to run `keras::install_keras()` inside the Docker image with already installed precommit. Looks like both keras and tensorflow are now [installed by default to `r-reticulate`](https://github.com/rstudio/keras/issues/1014).
This results in the following error:
>Collecting package metadata (current_repodata.json): ...working... done
Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
Solving environment: ...working... failed with repodata from current_repodata.json, will retry with next repodata source.
Collecting package metadata (repodata.json): ...working... done
Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
Examining setuptools: 7%|▋ | 2/30 [00:00<00:00, 2603.54it/s]
Comparing specs that have this dependency: 0%| | 0/15 [00:00<?, ?it/s]
Finding conflict paths: 0%| | 0/2 [00:00<?, ?it/s]
Finding shortest conflict path for setuptools: 0%| | 0/2 [00:00<?, ?it/s]
Finding shortest conflict path for setuptools: 50%|█████ | 1/2 [00:01<00:01, 1.99s/it]
Finding shortest conflict path for setuptools: 100%|██████████| 2/2 [00:01<00:00, 1.01it/s]
...truncated...
|
1.0
|
Allow to choose installation environment - As mentioned in [#113 ](https://github.com/lorenzwalthert/precommit/issues/113#issuecomment-603808455). Recently I've tried to run `keras::install_keras()` inside the Docker image with already installed precommit. Looks like both keras and tensorflow are now [installed by default to `r-reticulate`](https://github.com/rstudio/keras/issues/1014).
This results in the following error:
>Collecting package metadata (current_repodata.json): ...working... done
Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
Solving environment: ...working... failed with repodata from current_repodata.json, will retry with next repodata source.
Collecting package metadata (repodata.json): ...working... done
Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
Examining setuptools: 7%|▋ | 2/30 [00:00<00:00, 2603.54it/s]
Comparing specs that have this dependency: 0%| | 0/15 [00:00<?, ?it/s]
Finding conflict paths: 0%| | 0/2 [00:00<?, ?it/s]
Finding shortest conflict path for setuptools: 0%| | 0/2 [00:00<?, ?it/s]
Finding shortest conflict path for setuptools: 50%|█████ | 1/2 [00:01<00:01, 1.99s/it]
Finding shortest conflict path for setuptools: 100%|██████████| 2/2 [00:01<00:00, 1.01it/s]
...truncated...
|
priority
|
allow to choose installation environment as mentioned in recently i ve tried to run keras install keras inside the docker image with already installed precommit looks like both keras and tensorflow are now this results in the following error collecting package metadata current repodata json working done solving environment working failed with initial frozen solve retrying with flexible solve solving environment working failed with repodata from current repodata json will retry with next repodata source collecting package metadata repodata json working done solving environment working failed with initial frozen solve retrying with flexible solve examining setuptools ▋ comparing specs that have this dependency finding conflict paths finding shortest conflict path for setuptools finding shortest conflict path for setuptools █████ finding shortest conflict path for setuptools ██████████ truncated
| 1
|
813,717
| 30,469,039,169
|
IssuesEvent
|
2023-07-17 12:29:53
|
tinkoff-ai/etna
|
https://api.github.com/repos/tinkoff-ai/etna
|
opened
|
LimitTransform
|
enhancement priority/high
|
### 🚀 Feature Request
Create a transform that limits values of some feature between the borders.
### Proposal
Create `LimitTransform`.
Parameters:
- `in_column`: column to make transformation on;
- `lower_bound`: lower bound for the value of the column; -infty by default;
- `upper_bound`: upper bound for the value of the column; +infty by default;
- If there is value out of limit the exception should be raised.
- NaNs should be ignored.
Reference: [Ensure time series forecasts stay within limits](https://datasciencestunt.com/time-series-forecasting-within-limits/).
To discuss:
- Should this transform work for non-target column?
- Should this transform have `inplace` parameter for working in non-inplace mode?
### Test cases
What should be checked:
- Working on target / non-target column
- Working with set/unset lower/upper values
- Exception on out-of-limit value
- Full pipeline that predicts some arbitrary values can be used
Don't forget to add inference tests into `tests/test_models/test_inference/`.
### Additional context
_No response_
|
1.0
|
LimitTransform - ### 🚀 Feature Request
Create a transform that limits values of some feature between the borders.
### Proposal
Create `LimitTransform`.
Parameters:
- `in_column`: column to make transformation on;
- `lower_bound`: lower bound for the value of the column; -infty by default;
- `upper_bound`: upper bound for the value of the column; +infty by default;
- If there is value out of limit the exception should be raised.
- NaNs should be ignored.
Reference: [Ensure time series forecasts stay within limits](https://datasciencestunt.com/time-series-forecasting-within-limits/).
To discuss:
- Should this transform work for non-target column?
- Should this transform have `inplace` parameter for working in non-inplace mode?
### Test cases
What should be checked:
- Working on target / non-target column
- Working with set/unset lower/upper values
- Exception on out-of-limit value
- Full pipeline that predicts some arbitrary values can be used
Don't forget to add inference tests into `tests/test_models/test_inference/`.
### Additional context
_No response_
|
priority
|
limittransform 🚀 feature request create a transform that limits values of some feature between the borders proposal create limittransform parameters in column column to make transformation on lower bound lower bound for the value of the column infty by default upper bound upper bound for the value of the column infty by default if there is value out of limit the exception should be raised nans should be ignored reference to discuss should this transform work for non target column should this transform have inplace parameter for working in non inplace mode test cases what should be checked working on target non target column working with set unset lower upper values exception on out of limit value full pipeline that predicts some arbitrary values can be used don t forget to add inference tests into tests test models test inference additional context no response
| 1
|
480,829
| 13,876,853,552
|
IssuesEvent
|
2020-10-17 01:02:37
|
webiny/webiny-js
|
https://api.github.com/repos/webiny/webiny-js
|
closed
|
Headless CMS - Access Token unable to access listContentModels
|
bug cms no-issue-activity priority: high
|
This is:
- Bug
## Expected Behavior
Using Access Token with /read API should allow me to use the `listConentModels` field.
## Actual Behavior
Throws `Not Authorized` error.

|
1.0
|
Headless CMS - Access Token unable to access listContentModels - This is:
- Bug
## Expected Behavior
Using Access Token with /read API should allow me to use the `listConentModels` field.
## Actual Behavior
Throws `Not Authorized` error.

|
priority
|
headless cms access token unable to access listcontentmodels this is bug expected behavior using access token with read api should allow me to use the listconentmodels field actual behavior throws not authorized error
| 1
|
420,768
| 12,243,764,121
|
IssuesEvent
|
2020-05-05 09:50:43
|
pokt-network/pocket-core
|
https://api.github.com/repos/pokt-network/pocket-core
|
opened
|
Remove HTTP Round Trip for Queries
|
enhancement high priority optimization performance
|
**Is your feature request related to a problem? Please describe.**
Remove HTTP calls for queries, change to a programmatic interface. Remove all QueryWithData calls and replace them with a direct programmatic interface.
**This is high priority for relays/dispatches only**
|
1.0
|
Remove HTTP Round Trip for Queries - **Is your feature request related to a problem? Please describe.**
Remove HTTP calls for queries, change to a programmatic interface. Remove all QueryWithData calls and replace them with a direct programmatic interface.
**This is high priority for relays/dispatches only**
|
priority
|
remove http round trip for queries is your feature request related to a problem please describe remove http calls for queries change to a programmatic interface remove all querywithdata calls and replace them with a direct programmatic interface this is high priority for relays dispatches only
| 1
|
443,215
| 12,761,475,610
|
IssuesEvent
|
2020-06-29 11:32:29
|
getkirby/kirby
|
https://api.github.com/repos/getkirby/kirby
|
closed
|
CURLOPT_SSL_VERIFYPEER disabled in Remote::fetch()
|
priority: high 🔥 type: bug 🐛
|
Hello,
I just took a look at `Remote::fetch()` due to some forum question and I noticed that `CURLOPT_SSL_VERIFYPEER` is set to `FALSE` by default. Is there a reason why this security feature is disabled by default? I would have expected SSL cert validation to be enabled by default or to find some information why this is disabled but I found nothing about this.
See this line: https://github.com/getkirby/kirby/blob/6bd14bc8099d6c9dbe67ef0582b8c0a2d4c0888c/src/Http/Remote.php#L152
I'm sorry if this should have been posted in the forum or if it's an obvious misunderstanding on my side.
Best Regards,
|
1.0
|
CURLOPT_SSL_VERIFYPEER disabled in Remote::fetch() - Hello,
I just took a look at `Remote::fetch()` due to some forum question and I noticed that `CURLOPT_SSL_VERIFYPEER` is set to `FALSE` by default. Is there a reason why this security feature is disabled by default? I would have expected SSL cert validation to be enabled by default or to find some information why this is disabled but I found nothing about this.
See this line: https://github.com/getkirby/kirby/blob/6bd14bc8099d6c9dbe67ef0582b8c0a2d4c0888c/src/Http/Remote.php#L152
I'm sorry if this should have been posted in the forum or if it's an obvious misunderstanding on my side.
Best Regards,
|
priority
|
curlopt ssl verifypeer disabled in remote fetch hello i just took a look at remote fetch due to some forum question and i noticed that curlopt ssl verifypeer is set to false by default is there a reason why this security feature is disabled by default i would have expected ssl cert validation to be enabled by default or to find some information why this is disabled but i found nothing about this see this line i m sorry if this should have been posted in the forum or if it s an obvious misunderstanding on my side best regards
| 1
|
684,578
| 23,422,835,725
|
IssuesEvent
|
2022-08-14 00:28:54
|
ChaosInitiative/Portal-2-Community-Edition
|
https://api.github.com/repos/ChaosInitiative/Portal-2-Community-Edition
|
closed
|
Bug: Multiplayer players spawn at info_player_start instead of info_coop_spawn
|
Type: bug Component: gameplay Priority 2: High Status: resolved Focus: Co-Op Size 5: Tiny
|
### Describe the bug
Self-explanatory...
This is in the most recent nightly build. Despite the issue being super obvious, I didn't hear anyone talk about it yet so I just wanted to make an issue for it.
Happens regardless of player team.
### Issue Map
All maps.
### To Reproduce
1. Set ConVar **mp_dev_wait_for_other_player** to 0
2. Start a multiplayer game through **map mp_coop_2paints_1bridge**
3. Player spawns at info_player_start
### Operating System
Tested on Windows 10
|
1.0
|
Bug: Multiplayer players spawn at info_player_start instead of info_coop_spawn - ### Describe the bug
Self-explanatory...
This is in the most recent nightly build. Despite the issue being super obvious, I didn't hear anyone talk about it yet so I just wanted to make an issue for it.
Happens regardless of player team.
### Issue Map
All maps.
### To Reproduce
1. Set ConVar **mp_dev_wait_for_other_player** to 0
2. Start a multiplayer game through **map mp_coop_2paints_1bridge**
3. Player spawns at info_player_start
### Operating System
Tested on Windows 10
|
priority
|
bug multiplayer players spawn at info player start instead of info coop spawn describe the bug self explanatory this is in the most recent nightly build despite the issue being super obvious i didn t hear anyone talk about it yet so i just wanted to make an issue for it happens regardless of player team issue map all maps to reproduce set convar mp dev wait for other player to start a multiplayer game through map mp coop player spawns at info player start operating system tested on windows
| 1
|
312,831
| 9,553,708,880
|
IssuesEvent
|
2019-05-02 20:00:17
|
eJourn-al/eJournal
|
https://api.github.com/repos/eJourn-al/eJournal
|
closed
|
Non-required file upload throws an error when left empty
|
Priority: High Status: Review Needed Type: Bug Workload: Low
|
**Describe the bug**
The back end always denies requests that contain `None` for a file upload field.
**To Reproduce**
Steps to reproduce the behavior:
1. Make a template with a non-required file upload field
2. Try to post it as a student
3. Observe that the server returns a bad request stating `One of your files was not correctly uploaded, please try gain.`
**Expected behavior**
Non-required fields to be non-required.
|
1.0
|
Non-required file upload throws an error when left empty - **Describe the bug**
The back end always denies requests that contain `None` for a file upload field.
**To Reproduce**
Steps to reproduce the behavior:
1. Make a template with a non-required file upload field
2. Try to post it as a student
3. Observe that the server returns a bad request stating `One of your files was not correctly uploaded, please try gain.`
**Expected behavior**
Non-required fields to be non-required.
|
priority
|
non required file upload throws an error when left empty describe the bug the back end always denies requests that contain none for a file upload field to reproduce steps to reproduce the behavior make a template with a non required file upload field try to post it as a student observe that the server returns a bad request stating one of your files was not correctly uploaded please try gain expected behavior non required fields to be non required
| 1
|
668,301
| 22,577,064,654
|
IssuesEvent
|
2022-06-28 08:19:23
|
freesewing/freesewing
|
https://api.github.com/repos/freesewing/freesewing
|
closed
|
Ursula is not drafting in 2.21.0
|
:bug: bug :rotating_light: high priority :package: ursula
|
With either inputed measures or the "standard measurses* Ursula is not drafting in 2.21.0. I'm not 100 but I suspect this is the splitting packages folders.
|
1.0
|
Ursula is not drafting in 2.21.0 - With either inputed measures or the "standard measurses* Ursula is not drafting in 2.21.0. I'm not 100 but I suspect this is the splitting packages folders.
|
priority
|
ursula is not drafting in with either inputed measures or the standard measurses ursula is not drafting in i m not but i suspect this is the splitting packages folders
| 1
|
259,855
| 8,200,702,870
|
IssuesEvent
|
2018-09-01 08:05:08
|
marvinlabs/customer-area
|
https://api.github.com/repos/marvinlabs/customer-area
|
opened
|
Bug in reset password
|
Priority - high bug
|
Some keys are generated with special characters (e.g. `1535687074:$P$Bdz168aJBFAFUOaQpZZHK6oArvdxDk0` ) but the add-on code seems to be removing those characters **before** comparison with in-db user key.
|
1.0
|
Bug in reset password - Some keys are generated with special characters (e.g. `1535687074:$P$Bdz168aJBFAFUOaQpZZHK6oArvdxDk0` ) but the add-on code seems to be removing those characters **before** comparison with in-db user key.
|
priority
|
bug in reset password some keys are generated with special characters e g p but the add on code seems to be removing those characters before comparison with in db user key
| 1
|
619,297
| 19,521,364,126
|
IssuesEvent
|
2021-12-29 19:09:51
|
Cotalker/documentation
|
https://api.github.com/repos/Cotalker/documentation
|
closed
|
Bug report: Vista de task se pega al cargar
|
Bug report Bug high priority Bug rejected
|
### Affected system
Cotalker Web Application
### Affected system (other)
_No response_
### Affected environment
Production
### Affected environment (other)
_No response_
### App version
17.5.8
### Details
Al cargar la vista de task, se queda la pestaña de chrome pegada

** No permite desplegar sección **

### Steps to reproduce
Abrir vista de task en un flujo de trabajo que tenga múltiple task
### Expected result
Que no se pegue la página y se carguen de forma correcta las tasks.
### Additional data
- Company: Cotalker
- Group : Operaciones (sub flujo optask)
|
1.0
|
Bug report: Vista de task se pega al cargar - ### Affected system
Cotalker Web Application
### Affected system (other)
_No response_
### Affected environment
Production
### Affected environment (other)
_No response_
### App version
17.5.8
### Details
Al cargar la vista de task, se queda la pestaña de chrome pegada

** No permite desplegar sección **

### Steps to reproduce
Abrir vista de task en un flujo de trabajo que tenga múltiple task
### Expected result
Que no se pegue la página y se carguen de forma correcta las tasks.
### Additional data
- Company: Cotalker
- Group : Operaciones (sub flujo optask)
|
priority
|
bug report vista de task se pega al cargar affected system cotalker web application affected system other no response affected environment production affected environment other no response app version details al cargar la vista de task se queda la pestaña de chrome pegada no permite desplegar sección steps to reproduce abrir vista de task en un flujo de trabajo que tenga múltiple task expected result que no se pegue la página y se carguen de forma correcta las tasks additional data company cotalker group operaciones sub flujo optask
| 1
|
105,749
| 4,241,211,507
|
IssuesEvent
|
2016-07-06 15:40:25
|
ccswbs/hjckrrh
|
https://api.github.com/repos/ccswbs/hjckrrh
|
closed
|
G- Establish default front page layouts for pages
|
feature: general (G) priority: high type: enhancement request
|
This should be specified (documented requirements)- see Trello card- G14-#240
https://trello.com/c/ZkCKfdnw
|
1.0
|
G- Establish default front page layouts for pages - This should be specified (documented requirements)- see Trello card- G14-#240
https://trello.com/c/ZkCKfdnw
|
priority
|
g establish default front page layouts for pages this should be specified documented requirements see trello card
| 1
|
667,161
| 22,420,269,662
|
IssuesEvent
|
2022-06-20 01:45:26
|
portefaix/portefaix
|
https://api.github.com/repos/portefaix/portefaix
|
closed
|
AKS: Ingress Application Gateway
|
priority/high kind/feature area/terraform lifecycle/stale lifecycle/frozen cloud/azure todo
|
AKS: Ingress Application Gateway
- https://github.com/Azure/terraform-azurerm-aks/pull/99
enable_ingress_application_gateway = true
ingress_application_gateway_gateway_name =
ingress_application_gateway_subnet_cidr =
ingress_application_gateway_subnet_id =
https://github.com/portefaix/portefaix/blob/739c1d3f62ac9ddb95d4c2cfd684529098216e8c/terraform/azure/aks/modules/aks/aks.tf#L53
```ruby
# rbac_aad_managed = false
# rbac_aad_admin_group_object_ids = var.admin_group_object_ids
enable_log_analytics_workspace = false
enable_auto_scaling = var.enable_auto_scaling
enable_kube_dashboard = var.enable_kube_dashboard
enable_azure_policy = var.enable_azure_policy
enable_http_application_routing = var.enable_http_application_routing
# TODO: AKS: Ingress Application Gateway
# labels: kind/feature, priority/high, lifecycle/frozen, area/terraform, cloud/azure
# https://github.com/Azure/terraform-azurerm-aks/pull/99
# enable_ingress_application_gateway = true
# ingress_application_gateway_gateway_name =
# ingress_application_gateway_subnet_cidr =
# ingress_application_gateway_subnet_id =
os_disk_size_gb = var.os_disk_size_gb
agents_min_count = var.agents_min_count
```
afce5cfdca8a7775d0e82c8e7e9387f68d5e7736
|
1.0
|
AKS: Ingress Application Gateway - AKS: Ingress Application Gateway
- https://github.com/Azure/terraform-azurerm-aks/pull/99
enable_ingress_application_gateway = true
ingress_application_gateway_gateway_name =
ingress_application_gateway_subnet_cidr =
ingress_application_gateway_subnet_id =
https://github.com/portefaix/portefaix/blob/739c1d3f62ac9ddb95d4c2cfd684529098216e8c/terraform/azure/aks/modules/aks/aks.tf#L53
```ruby
# rbac_aad_managed = false
# rbac_aad_admin_group_object_ids = var.admin_group_object_ids
enable_log_analytics_workspace = false
enable_auto_scaling = var.enable_auto_scaling
enable_kube_dashboard = var.enable_kube_dashboard
enable_azure_policy = var.enable_azure_policy
enable_http_application_routing = var.enable_http_application_routing
# TODO: AKS: Ingress Application Gateway
# labels: kind/feature, priority/high, lifecycle/frozen, area/terraform, cloud/azure
# https://github.com/Azure/terraform-azurerm-aks/pull/99
# enable_ingress_application_gateway = true
# ingress_application_gateway_gateway_name =
# ingress_application_gateway_subnet_cidr =
# ingress_application_gateway_subnet_id =
os_disk_size_gb = var.os_disk_size_gb
agents_min_count = var.agents_min_count
```
afce5cfdca8a7775d0e82c8e7e9387f68d5e7736
|
priority
|
aks ingress application gateway aks ingress application gateway enable ingress application gateway true ingress application gateway gateway name ingress application gateway subnet cidr ingress application gateway subnet id ruby rbac aad managed false rbac aad admin group object ids var admin group object ids enable log analytics workspace false enable auto scaling var enable auto scaling enable kube dashboard var enable kube dashboard enable azure policy var enable azure policy enable http application routing var enable http application routing todo aks ingress application gateway labels kind feature priority high lifecycle frozen area terraform cloud azure enable ingress application gateway true ingress application gateway gateway name ingress application gateway subnet cidr ingress application gateway subnet id os disk size gb var os disk size gb agents min count var agents min count
| 1
|
6,304
| 2,587,112,727
|
IssuesEvent
|
2015-02-17 16:30:01
|
civio/quienmanda.es
|
https://api.github.com/repos/civio/quienmanda.es
|
closed
|
Implementar 'páginas temáticas'
|
high_priority in progress
|
Queremos agrupar una serie de fotos y artículos en una "página temática" que compile y muestre de una forma agradable toda la información sobre un tema.
Primer uso: compilar los artículos sobre el Colegio de El Pilar. Comenzaremos mostrando simplemente una serie de fotos y/o artículos etiquetados con una palabra X, pero podría añadirse una foto y/o texto de entrada, dando contexto. Posteriormente esto aplica a temas como 'energía', 'banca'...
|
1.0
|
Implementar 'páginas temáticas' - Queremos agrupar una serie de fotos y artículos en una "página temática" que compile y muestre de una forma agradable toda la información sobre un tema.
Primer uso: compilar los artículos sobre el Colegio de El Pilar. Comenzaremos mostrando simplemente una serie de fotos y/o artículos etiquetados con una palabra X, pero podría añadirse una foto y/o texto de entrada, dando contexto. Posteriormente esto aplica a temas como 'energía', 'banca'...
|
priority
|
implementar páginas temáticas queremos agrupar una serie de fotos y artículos en una página temática que compile y muestre de una forma agradable toda la información sobre un tema primer uso compilar los artículos sobre el colegio de el pilar comenzaremos mostrando simplemente una serie de fotos y o artículos etiquetados con una palabra x pero podría añadirse una foto y o texto de entrada dando contexto posteriormente esto aplica a temas como energía banca
| 1
|
283,012
| 8,712,895,467
|
IssuesEvent
|
2018-12-06 23:59:53
|
aowen87/TicketTester
|
https://api.github.com/repos/aowen87/TicketTester
|
closed
|
VisIt hangs during re-execution prompted by pick.
|
bug crash likelihood medium priority reviewed severity high wrong results
|
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps.
This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.
Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var)
On surface: (ensure parallel engine with 2 nodes)
Open multi_curv2d.silo
Add PC Plot of d
Add CoordSwap operator, swap x and y coords
Draw.
Create a scalar expression d+p,
Open Pick window, set variable to d+p.
Apply
Do a Zone Pick
Using Navigation, change the view either by zooming or panning.
Do another Zone Pick.
Engine will hang, must cancel the engine_par job in order to interact with VisIt again.
Information window shows Pick wanting to re-execute, and a merge exception:
++++++++++++++++
VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks.
Shortly thereafter, the following occured...
Pseudocolor: (InvalidMergeException)
viewer: Cannot merge datasets because of an incompatible field 1 and 2.
Pick mode now fully ready.
+++++++++++++++++++++
This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error:
+++++++++++++++++++++++++++++++++++++
This source should not load balance the data.
Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.
catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027
++++++++++++++++++++++++++++++
This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2169
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: VisIt hangs during re-execution prompted by pick.
Assigned to: Kathleen Biagas
Category:
Target version: 2.9.1
Author: Kathleen Biagas
Start: 03/03/2015
Due date:
% Done: 100
Estimated time:
Created: 03/03/2015 04:55 pm
Updated: 03/20/2015 05:44 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.8.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps.
This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.
Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var)
On surface: (ensure parallel engine with 2 nodes)
Open multi_curv2d.silo
Add PC Plot of d
Add CoordSwap operator, swap x and y coords
Draw.
Create a scalar expression d+p,
Open Pick window, set variable to d+p.
Apply
Do a Zone Pick
Using Navigation, change the view either by zooming or panning.
Do another Zone Pick.
Engine will hang, must cancel the engine_par job in order to interact with VisIt again.
Information window shows Pick wanting to re-execute, and a merge exception:
++++++++++++++++
VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks.
Shortly thereafter, the following occured...
Pseudocolor: (InvalidMergeException)
viewer: Cannot merge datasets because of an incompatible field 1 and 2.
Pick mode now fully ready.
+++++++++++++++++++++
This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error:
+++++++++++++++++++++++++++++++++++++
This source should not load balance the data.
Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.
catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027
++++++++++++++++++++++++++++++
This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
Comments:
Turns out that Pick was sending the secondary variable request to only 1 processor, which caused an 'Invalid Merge' exception during pipeline re-execution.I modified Pick to request SecondaryVars of all processors.SVN update 25986 (2.9RC), 25988 (trunk)M /src/avt/Queries/Pick/avtPickQuery.C
|
1.0
|
VisIt hangs during re-execution prompted by pick. - This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps.
This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.
Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var)
On surface: (ensure parallel engine with 2 nodes)
Open multi_curv2d.silo
Add PC Plot of d
Add CoordSwap operator, swap x and y coords
Draw.
Create a scalar expression d+p,
Open Pick window, set variable to d+p.
Apply
Do a Zone Pick
Using Navigation, change the view either by zooming or panning.
Do another Zone Pick.
Engine will hang, must cancel the engine_par job in order to interact with VisIt again.
Information window shows Pick wanting to re-execute, and a merge exception:
++++++++++++++++
VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks.
Shortly thereafter, the following occured...
Pseudocolor: (InvalidMergeException)
viewer: Cannot merge datasets because of an incompatible field 1 and 2.
Pick mode now fully ready.
+++++++++++++++++++++
This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error:
+++++++++++++++++++++++++++++++++++++
This source should not load balance the data.
Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.
catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027
++++++++++++++++++++++++++++++
This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2169
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: VisIt hangs during re-execution prompted by pick.
Assigned to: Kathleen Biagas
Category:
Target version: 2.9.1
Author: Kathleen Biagas
Start: 03/03/2015
Due date:
% Done: 100
Estimated time:
Created: 03/03/2015 04:55 pm
Updated: 03/20/2015 05:44 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.8.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps.
This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.
Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var)
On surface: (ensure parallel engine with 2 nodes)
Open multi_curv2d.silo
Add PC Plot of d
Add CoordSwap operator, swap x and y coords
Draw.
Create a scalar expression d+p,
Open Pick window, set variable to d+p.
Apply
Do a Zone Pick
Using Navigation, change the view either by zooming or panning.
Do another Zone Pick.
Engine will hang, must cancel the engine_par job in order to interact with VisIt again.
Information window shows Pick wanting to re-execute, and a merge exception:
++++++++++++++++
VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks.
Shortly thereafter, the following occured...
Pseudocolor: (InvalidMergeException)
viewer: Cannot merge datasets because of an incompatible field 1 and 2.
Pick mode now fully ready.
+++++++++++++++++++++
This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error:
+++++++++++++++++++++++++++++++++++++
This source should not load balance the data.
Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.
catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027
++++++++++++++++++++++++++++++
This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
Comments:
Turns out that Pick was sending the secondary variable request to only 1 processor, which caused an 'Invalid Merge' exception during pipeline re-execution.I modified Pick to request SecondaryVars of all processors.SVN update 25986 (2.9RC), 25988 (trunk)M /src/avt/Queries/Pick/avtPickQuery.C
|
priority
|
visit hangs during re execution prompted by pick this is a bug bruce hammel claims to have been experiencing with visit for years so i ve set the priority as high now that i have reproducible steps this seems only to occur with nodes multiple processors on single node does not replicate also seems only to occur in conjunction with coordswap operator and with pick var being set to an expression and not the pipeline var on surface ensure parallel engine with nodes open multi silo add pc plot of d add coordswap operator swap x and y coords draw create a scalar expression d p open pick window set variable to d p apply do a zone pick using navigation change the view either by zooming or panning do another zone pick engine will hang must cancel the engine par job in order to interact with visit again information window shows pick wanting to re execute and a merge exception visit does not have all the information it needs to perform a pick please wait while the necessary information is calculated all current pick selections have been cached and will be performed when calculations are complete visit will notify you when it is fully ready for more picks shortly thereafter the following occured pseudocolor invalidmergeexception viewer cannot merge datasets because of an incompatible field and pick mode now fully ready this shows error seeming to come from the viewer but if you run with debug then processors log files will show this error this source should not load balance the data exception invalidmergeexception usr tmp brugger aztec visitbuild src avt pipeline data avtdataattributes c line cannot merge datasets because of an incompatible field and catch visitexception usr tmp brugger aztec visitbuild src engine main executors h this is consistent with what i saw with bruce s real data running on muir in pdebug with nodes and processors but also nodes and processors redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject visit hangs during re execution prompted by pick assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description this is a bug bruce hammel claims to have been experiencing with visit for years so i ve set the priority as high now that i have reproducible steps this seems only to occur with nodes multiple processors on single node does not replicate also seems only to occur in conjunction with coordswap operator and with pick var being set to an expression and not the pipeline var on surface ensure parallel engine with nodes open multi silo add pc plot of d add coordswap operator swap x and y coords draw create a scalar expression d p open pick window set variable to d p apply do a zone pick using navigation change the view either by zooming or panning do another zone pick engine will hang must cancel the engine par job in order to interact with visit again information window shows pick wanting to re execute and a merge exception visit does not have all the information it needs to perform a pick please wait while the necessary information is calculated all current pick selections have been cached and will be performed when calculations are complete visit will notify you when it is fully ready for more picks shortly thereafter the following occured pseudocolor invalidmergeexception viewer cannot merge datasets because of an incompatible field and pick mode now fully ready this shows error seeming to come from the viewer but if you run with debug then processors log files will show this error this source should not load balance the data exception invalidmergeexception usr tmp brugger aztec visitbuild src avt pipeline data avtdataattributes c line cannot merge datasets because of an incompatible field and catch visitexception usr tmp brugger aztec visitbuild src engine main executors h this is consistent with what i saw with bruce s real data running on muir in pdebug with nodes and processors but also nodes and processors comments turns out that pick was sending the secondary variable request to only processor which caused an invalid merge exception during pipeline re execution i modified pick to request secondaryvars of all processors svn update trunk m src avt queries pick avtpickquery c
| 1
|
673,551
| 23,018,356,996
|
IssuesEvent
|
2022-07-22 00:43:33
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
Conversion from bool to float sometimes produces tensor with values 255.0 for True instead of 1.0
|
high priority triaged module: numpy module: boolean tensor actionable module: correctness (silent)
|
## 🐛 Bug
When I have two seemingly identical bool tensors with values: `tensor([True, True, True, True])`, one of them produces a float tensor with values: `tensor([1.0, 1.0, 1.0, 1.0])`. The other one produces a float tensor with values: `tensor([255., 255., 255., 255.])`.
## To Reproduce
Steps to reproduce the behavior:
```python
import numpy as np
import torch
# these are the same data and they look the same when you load them, but they behave subtly differently
poisoned = np.frombuffer(b'\xff\xff\xff\xff', dtype=np.bool) # hex 255
clean = np.frombuffer(b'\x01\x01\x01\x01', dtype=np.bool)
torch_poisoned = torch.from_numpy(poisoned)
torch_clean = torch.from_numpy(clean)
print('poisoned == clean -->', (poisoned == clean).all()) # >>> True
print('torch_poisoned == torch_clean --> ', (torch_poisoned == torch_clean).all()) # >>> tensor(False)
print('torch_poisoned.dtype == torch_clean.dtype --> ', (torch_poisoned.dtype == torch_clean.dtype)) # >>> True
print('torch_poisoned.dtype, torch_clean.dtype --> ', torch_poisoned.dtype, torch_clean.dtype) # >>> torch.bool, torch.bool
print('torch_clean: ',torch_clean) # >>> tensor([True, True, True, True])
print('torch_clean.float(): ', torch_clean.float()) # >>> tensor([1., 1., 1., 1.])
print('torch_poisoned: ', torch_poisoned) # >>> tensor([True, True, True, True])
print('torch_poisoned.float(): ', torch_poisoned.float()) # >>> tensor([255., 255., 255., 255.])
```
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
## Expected behavior
Any thing that torch calls a True bool should be converted to a float value of 1.0, not whatever bytes happened to be stored inside the bool.
This could just be an abuse of bool, but I came across this bug when using converting a boolean image using PIL. PyTorch should probably prevent users from being able to shoot their foot like this.
<!-- A clear and concise description of what you expected to happen. -->
## Environment
Collecting environment information...
PyTorch version: 1.7.1+cu110
Is debug build: False
CUDA used to build PyTorch: 11.0
ROCM used to build PyTorch: N/A
OS: Ubuntu 18.04.5 LTS (x86_64)
GCC version: (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0
Clang version: Could not collect
CMake version: version 3.15.3
Python version: 3.6 (64-bit runtime)
Is CUDA available: True
CUDA runtime version: 11.2.142
GPU models and configuration: GPU 0: GeForce GTX 1080 Ti
Nvidia driver version: 455.45.01
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.1.1
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.18.5
[pip3] torch==1.7.1+cu110
[pip3] torchaudio==0.7.2
[pip3] torchgeometry==0.1.2
[pip3] torchvision==0.8.2+cu110
[conda] Could not collect
## Additional context
<!-- Add any other context about the problem here. -->
cc @ezyang @gchanan @zou3519 @mruberry @rgommers @heitorschueroff
|
1.0
|
Conversion from bool to float sometimes produces tensor with values 255.0 for True instead of 1.0 - ## 🐛 Bug
When I have two seemingly identical bool tensors with values: `tensor([True, True, True, True])`, one of them produces a float tensor with values: `tensor([1.0, 1.0, 1.0, 1.0])`. The other one produces a float tensor with values: `tensor([255., 255., 255., 255.])`.
## To Reproduce
Steps to reproduce the behavior:
```python
import numpy as np
import torch
# these are the same data and they look the same when you load them, but they behave subtly differently
poisoned = np.frombuffer(b'\xff\xff\xff\xff', dtype=np.bool) # hex 255
clean = np.frombuffer(b'\x01\x01\x01\x01', dtype=np.bool)
torch_poisoned = torch.from_numpy(poisoned)
torch_clean = torch.from_numpy(clean)
print('poisoned == clean -->', (poisoned == clean).all()) # >>> True
print('torch_poisoned == torch_clean --> ', (torch_poisoned == torch_clean).all()) # >>> tensor(False)
print('torch_poisoned.dtype == torch_clean.dtype --> ', (torch_poisoned.dtype == torch_clean.dtype)) # >>> True
print('torch_poisoned.dtype, torch_clean.dtype --> ', torch_poisoned.dtype, torch_clean.dtype) # >>> torch.bool, torch.bool
print('torch_clean: ',torch_clean) # >>> tensor([True, True, True, True])
print('torch_clean.float(): ', torch_clean.float()) # >>> tensor([1., 1., 1., 1.])
print('torch_poisoned: ', torch_poisoned) # >>> tensor([True, True, True, True])
print('torch_poisoned.float(): ', torch_poisoned.float()) # >>> tensor([255., 255., 255., 255.])
```
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
## Expected behavior
Any thing that torch calls a True bool should be converted to a float value of 1.0, not whatever bytes happened to be stored inside the bool.
This could just be an abuse of bool, but I came across this bug when using converting a boolean image using PIL. PyTorch should probably prevent users from being able to shoot their foot like this.
<!-- A clear and concise description of what you expected to happen. -->
## Environment
Collecting environment information...
PyTorch version: 1.7.1+cu110
Is debug build: False
CUDA used to build PyTorch: 11.0
ROCM used to build PyTorch: N/A
OS: Ubuntu 18.04.5 LTS (x86_64)
GCC version: (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0
Clang version: Could not collect
CMake version: version 3.15.3
Python version: 3.6 (64-bit runtime)
Is CUDA available: True
CUDA runtime version: 11.2.142
GPU models and configuration: GPU 0: GeForce GTX 1080 Ti
Nvidia driver version: 455.45.01
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.7.6.5
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.1.1
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.1.1
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.18.5
[pip3] torch==1.7.1+cu110
[pip3] torchaudio==0.7.2
[pip3] torchgeometry==0.1.2
[pip3] torchvision==0.8.2+cu110
[conda] Could not collect
## Additional context
<!-- Add any other context about the problem here. -->
cc @ezyang @gchanan @zou3519 @mruberry @rgommers @heitorschueroff
|
priority
|
conversion from bool to float sometimes produces tensor with values for true instead of 🐛 bug when i have two seemingly identical bool tensors with values tensor one of them produces a float tensor with values tensor the other one produces a float tensor with values tensor to reproduce steps to reproduce the behavior python import numpy as np import torch these are the same data and they look the same when you load them but they behave subtly differently poisoned np frombuffer b xff xff xff xff dtype np bool hex clean np frombuffer b dtype np bool torch poisoned torch from numpy poisoned torch clean torch from numpy clean print poisoned clean poisoned clean all true print torch poisoned torch clean torch poisoned torch clean all tensor false print torch poisoned dtype torch clean dtype torch poisoned dtype torch clean dtype true print torch poisoned dtype torch clean dtype torch poisoned dtype torch clean dtype torch bool torch bool print torch clean torch clean tensor print torch clean float torch clean float tensor print torch poisoned torch poisoned tensor print torch poisoned float torch poisoned float tensor expected behavior any thing that torch calls a true bool should be converted to a float value of not whatever bytes happened to be stored inside the bool this could just be an abuse of bool but i came across this bug when using converting a boolean image using pil pytorch should probably prevent users from being able to shoot their foot like this environment collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version python version bit runtime is cuda available true cuda runtime version gpu models and configuration gpu geforce gtx ti nvidia driver version cudnn version probably one of the following usr lib linux gnu libcudnn so usr lib linux gnu libcudnn so usr lib linux gnu libcudnn adv infer so usr lib linux gnu libcudnn adv train so usr lib linux gnu libcudnn cnn infer so usr lib linux gnu libcudnn cnn train so usr lib linux gnu libcudnn ops infer so usr lib linux gnu libcudnn ops train so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy torch torchaudio torchgeometry torchvision could not collect additional context cc ezyang gchanan mruberry rgommers heitorschueroff
| 1
|
105,139
| 4,231,311,731
|
IssuesEvent
|
2016-07-04 15:28:27
|
bromagosa/beetleCloud
|
https://api.github.com/repos/bromagosa/beetleCloud
|
closed
|
"Loading project list" message stays open forever
|
bug high priority
|
When you open the projects browser before having saved any projects
|
1.0
|
"Loading project list" message stays open forever - When you open the projects browser before having saved any projects
|
priority
|
loading project list message stays open forever when you open the projects browser before having saved any projects
| 1
|
80,003
| 3,549,487,259
|
IssuesEvent
|
2016-01-20 18:11:57
|
DDMAL/Rodan
|
https://api.github.com/repos/DDMAL/Rodan
|
closed
|
Coordinate sets for WorkflowJobGroup
|
Location: Core system Priority: High Type: Feature
|
We'll either need a new model or edit the existing coordinate set model to take different types of objects. Your call.
|
1.0
|
Coordinate sets for WorkflowJobGroup - We'll either need a new model or edit the existing coordinate set model to take different types of objects. Your call.
|
priority
|
coordinate sets for workflowjobgroup we ll either need a new model or edit the existing coordinate set model to take different types of objects your call
| 1
|
40,286
| 2,868,385,780
|
IssuesEvent
|
2015-06-05 18:31:55
|
nim-lang/nimble
|
https://api.github.com/repos/nim-lang/nimble
|
closed
|
Empty uninstall
|
Bug High priority
|
You can pass an empty string to nimble uninstall. Might want to uninstall the current package if a .nimble is found, similar to nimble install (that's the assumption i had, and that's why i ran this command)? I got scared for a minute after actually pressing "Y" :fearful:.
```
$ nimble uninstall
The following packages will be removed:
Do you wish to continue? [y/N]
```
|
1.0
|
Empty uninstall - You can pass an empty string to nimble uninstall. Might want to uninstall the current package if a .nimble is found, similar to nimble install (that's the assumption i had, and that's why i ran this command)? I got scared for a minute after actually pressing "Y" :fearful:.
```
$ nimble uninstall
The following packages will be removed:
Do you wish to continue? [y/N]
```
|
priority
|
empty uninstall you can pass an empty string to nimble uninstall might want to uninstall the current package if a nimble is found similar to nimble install that s the assumption i had and that s why i ran this command i got scared for a minute after actually pressing y fearful nimble uninstall the following packages will be removed do you wish to continue
| 1
|
591,973
| 17,866,993,361
|
IssuesEvent
|
2021-09-06 10:39:51
|
AbsaOSS/enceladus
|
https://api.github.com/repos/AbsaOSS/enceladus
|
closed
|
Authenticated using keytab, configuring Kerberos on the EMR master node
|
feature security priority: high cloud
|
## Background
Keytab is used for authentication within Enceladus
## Feature
Set-up and verify keytab authentication in EMR
|
1.0
|
Authenticated using keytab, configuring Kerberos on the EMR master node - ## Background
Keytab is used for authentication within Enceladus
## Feature
Set-up and verify keytab authentication in EMR
|
priority
|
authenticated using keytab configuring kerberos on the emr master node background keytab is used for authentication within enceladus feature set up and verify keytab authentication in emr
| 1
|
588,976
| 17,687,070,986
|
IssuesEvent
|
2021-08-24 04:07:06
|
woowa-techcamp-2021/store-6
|
https://api.github.com/repos/woowa-techcamp-2021/store-6
|
closed
|
[FE] 헤더 컴포넌트 가장 기본 UI 구현
|
high priority feat
|
## :hammer: 기능 설명
드랍다운, 장바구니 API 연동 등을 제외한 가장 기본적인 헤더의 UI를 구현합니다.
## 📑 완료 조건
- [x] 각 버튼/링크를 클릭했을 때 해당 페이지로 이동해야 합니다.
- [x] 로그인이 되었을 경우 사용자의 프로필이미지를 기존 아이콘 대신 보여줘야 합니다.
## :thought_balloon: 관련 Backlog
> [대분류] - [중분류] - [Backlog 이름]
- [FE] 헤더 - 헤더 컴포넌트 - 리팩토링(공통 컴포넌트 분리 등)
- [FE] 헤더 - 헤더 컴포넌트 - 로고 클릭 시 홈으로 이동
- [FE] 헤더 - 헤더 컴포넌트 - 장바구니, 좋아요, 마이페이지 버튼 클릭 시 페이지 이동
- [FE] 기타 - 공통 컴포넌트 - SEO를 위한 로고 개선
|
1.0
|
[FE] 헤더 컴포넌트 가장 기본 UI 구현 - ## :hammer: 기능 설명
드랍다운, 장바구니 API 연동 등을 제외한 가장 기본적인 헤더의 UI를 구현합니다.
## 📑 완료 조건
- [x] 각 버튼/링크를 클릭했을 때 해당 페이지로 이동해야 합니다.
- [x] 로그인이 되었을 경우 사용자의 프로필이미지를 기존 아이콘 대신 보여줘야 합니다.
## :thought_balloon: 관련 Backlog
> [대분류] - [중분류] - [Backlog 이름]
- [FE] 헤더 - 헤더 컴포넌트 - 리팩토링(공통 컴포넌트 분리 등)
- [FE] 헤더 - 헤더 컴포넌트 - 로고 클릭 시 홈으로 이동
- [FE] 헤더 - 헤더 컴포넌트 - 장바구니, 좋아요, 마이페이지 버튼 클릭 시 페이지 이동
- [FE] 기타 - 공통 컴포넌트 - SEO를 위한 로고 개선
|
priority
|
헤더 컴포넌트 가장 기본 ui 구현 hammer 기능 설명 드랍다운 장바구니 api 연동 등을 제외한 가장 기본적인 헤더의 ui를 구현합니다 📑 완료 조건 각 버튼 링크를 클릭했을 때 해당 페이지로 이동해야 합니다 로그인이 되었을 경우 사용자의 프로필이미지를 기존 아이콘 대신 보여줘야 합니다 thought balloon 관련 backlog 헤더 헤더 컴포넌트 리팩토링 공통 컴포넌트 분리 등 헤더 헤더 컴포넌트 로고 클릭 시 홈으로 이동 헤더 헤더 컴포넌트 장바구니 좋아요 마이페이지 버튼 클릭 시 페이지 이동 기타 공통 컴포넌트 seo를 위한 로고 개선
| 1
|
789,048
| 27,776,805,147
|
IssuesEvent
|
2023-03-16 17:48:09
|
WFP-VAM/prism-app
|
https://api.github.com/repos/WFP-VAM/prism-app
|
closed
|
[Bug]: Kobo get dates request doesn't implement filters
|
bug priority:high triage
|
### What happened?
The timeline interface shows that there's data available on a given date, but when I change to that date, I get an error message saying there's no data.
I can see that the initial request for available dates adds the filter attribute as expected, but the results do not apply the filter, so dates are being returned when they should not. See this query:
https://prism-api.ovio.org/kobo/dates?formId=aMLBShjyyhYeS5ERuRy8s3&datetimeField=Date_report&measureField=NumPeoAff&koboUrl=https://kobo.humanitarianresponse.info/api/v2/assets/&filters=status=Approved
As an example, '2023-01-16' is returned as a valid date. If I select that date in the app, the request correctly adds the filter parameter and (correctly) returns an empty dataset. Here's the request:
https://prism-api.ovio.org/kobo/forms?beginDateTime=2023-01-16&endDateTime=2023-01-16&formId=aMLBShjyyhYeS5ERuRy8s3&datetimeField=Date_report&measureField=NumPeoAff&koboUrl=https://kobo.humanitarianresponse.info/api/v2/assets/&filters=status=Approved
I can see in [api/app/kobo.py ](https://github.com/WFP-VAM/prism-app/blob/master/api/app/kobo.py)get_form_dates() doesn't make use of the filter, but get_form_responses() does.
### Which country / deployment are you running?
Cambodia on PR https://github.com/WFP-VAM/prism-app/pull/688
### Add a screenshot (if relevant)
https://prism-khm-new-ux.surge.sh/?hazardLayerIds=flood_report&date=2023-01-16
<img width="1520" alt="Screenshot 2023-03-15 at 22 45 06" src="https://user-images.githubusercontent.com/3343536/225529229-c47bfbe8-8065-48c0-8bfb-6bf5fc50b145.png">
Response from the date request:
<img width="355" alt="Screenshot 2023-03-15 at 23 06 23" src="https://user-images.githubusercontent.com/3343536/225529617-fa3b1d5f-f08a-421d-9c6d-efb1c53ea0fc.png">
|
1.0
|
[Bug]: Kobo get dates request doesn't implement filters - ### What happened?
The timeline interface shows that there's data available on a given date, but when I change to that date, I get an error message saying there's no data.
I can see that the initial request for available dates adds the filter attribute as expected, but the results do not apply the filter, so dates are being returned when they should not. See this query:
https://prism-api.ovio.org/kobo/dates?formId=aMLBShjyyhYeS5ERuRy8s3&datetimeField=Date_report&measureField=NumPeoAff&koboUrl=https://kobo.humanitarianresponse.info/api/v2/assets/&filters=status=Approved
As an example, '2023-01-16' is returned as a valid date. If I select that date in the app, the request correctly adds the filter parameter and (correctly) returns an empty dataset. Here's the request:
https://prism-api.ovio.org/kobo/forms?beginDateTime=2023-01-16&endDateTime=2023-01-16&formId=aMLBShjyyhYeS5ERuRy8s3&datetimeField=Date_report&measureField=NumPeoAff&koboUrl=https://kobo.humanitarianresponse.info/api/v2/assets/&filters=status=Approved
I can see in [api/app/kobo.py ](https://github.com/WFP-VAM/prism-app/blob/master/api/app/kobo.py)get_form_dates() doesn't make use of the filter, but get_form_responses() does.
### Which country / deployment are you running?
Cambodia on PR https://github.com/WFP-VAM/prism-app/pull/688
### Add a screenshot (if relevant)
https://prism-khm-new-ux.surge.sh/?hazardLayerIds=flood_report&date=2023-01-16
<img width="1520" alt="Screenshot 2023-03-15 at 22 45 06" src="https://user-images.githubusercontent.com/3343536/225529229-c47bfbe8-8065-48c0-8bfb-6bf5fc50b145.png">
Response from the date request:
<img width="355" alt="Screenshot 2023-03-15 at 23 06 23" src="https://user-images.githubusercontent.com/3343536/225529617-fa3b1d5f-f08a-421d-9c6d-efb1c53ea0fc.png">
|
priority
|
kobo get dates request doesn t implement filters what happened the timeline interface shows that there s data available on a given date but when i change to that date i get an error message saying there s no data i can see that the initial request for available dates adds the filter attribute as expected but the results do not apply the filter so dates are being returned when they should not see this query as an example is returned as a valid date if i select that date in the app the request correctly adds the filter parameter and correctly returns an empty dataset here s the request i can see in doesn t make use of the filter but get form responses does which country deployment are you running cambodia on pr add a screenshot if relevant img width alt screenshot at src response from the date request img width alt screenshot at src
| 1
|
136,510
| 5,284,090,531
|
IssuesEvent
|
2017-02-07 23:09:28
|
pymedusa/Medusa
|
https://api.github.com/repos/pymedusa/Medusa
|
closed
|
[APP SUBMITTED]: TypeError: unhashable type: 'list'
|
Bug Confirmed Priority: 1. High
|
### INFO
**Python Version**: `2.7.12 (v2.7.12:d33e0cf91556, Jun 27 2016, 15:24:40) [MSC v.1500 64 bit (AMD64)]`
**Operating System**: `Windows-10-10.0.14393`
**Locale**: `cp1252`
**Branch**: [develop](../tree/develop)
**Database**: `44.4`
**Commit**: pymedusa/Medusa@63f66cdd827c391aa500986cd2b1e3b8276f858e
**Link to Log**: https://gist.github.com/3ac6567533f759fba5d4d3288d93eb69
### ERROR
<pre>
2017-02-07 15:57:36 ERROR Thread-37 :: [63f66cd] Exception generated: unhashable type: 'list'
Traceback (most recent call last):
File "C:\Medusa\medusa\server\web\core\base.py", line 259, in async_call
result = function(**kwargs)
File "C:\Medusa\medusa\server\web\home\handler.py", line 1431, in editShow
exceptions = {exceptions}
TypeError: unhashable type: 'list'
</pre>
---
_STAFF NOTIFIED_: @pymedusa/support @pymedusa/moderators
|
1.0
|
[APP SUBMITTED]: TypeError: unhashable type: 'list' - ### INFO
**Python Version**: `2.7.12 (v2.7.12:d33e0cf91556, Jun 27 2016, 15:24:40) [MSC v.1500 64 bit (AMD64)]`
**Operating System**: `Windows-10-10.0.14393`
**Locale**: `cp1252`
**Branch**: [develop](../tree/develop)
**Database**: `44.4`
**Commit**: pymedusa/Medusa@63f66cdd827c391aa500986cd2b1e3b8276f858e
**Link to Log**: https://gist.github.com/3ac6567533f759fba5d4d3288d93eb69
### ERROR
<pre>
2017-02-07 15:57:36 ERROR Thread-37 :: [63f66cd] Exception generated: unhashable type: 'list'
Traceback (most recent call last):
File "C:\Medusa\medusa\server\web\core\base.py", line 259, in async_call
result = function(**kwargs)
File "C:\Medusa\medusa\server\web\home\handler.py", line 1431, in editShow
exceptions = {exceptions}
TypeError: unhashable type: 'list'
</pre>
---
_STAFF NOTIFIED_: @pymedusa/support @pymedusa/moderators
|
priority
|
typeerror unhashable type list info python version jun operating system windows locale branch tree develop database commit pymedusa medusa link to log error error thread exception generated unhashable type list traceback most recent call last file c medusa medusa server web core base py line in async call result function kwargs file c medusa medusa server web home handler py line in editshow exceptions exceptions typeerror unhashable type list staff notified pymedusa support pymedusa moderators
| 1
|
769,894
| 27,021,546,756
|
IssuesEvent
|
2023-02-11 03:42:22
|
codersforcauses/poops
|
https://api.github.com/repos/codersforcauses/poops
|
opened
|
update admin email domain in firebase functions
|
backend bug enhancement difficulty::easy priority::high point::1
|
**Is your feature request related to a problem? Please describe.**
The current admin email domain is `poop.org` which is incorrect.
**Describe the solution you'd like**
Change the domain name in `addAdmin.ts` firebase functions to `poopswa.org.au` .
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
|
1.0
|
update admin email domain in firebase functions - **Is your feature request related to a problem? Please describe.**
The current admin email domain is `poop.org` which is incorrect.
**Describe the solution you'd like**
Change the domain name in `addAdmin.ts` firebase functions to `poopswa.org.au` .
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
|
priority
|
update admin email domain in firebase functions is your feature request related to a problem please describe the current admin email domain is poop org which is incorrect describe the solution you d like change the domain name in addadmin ts firebase functions to poopswa org au describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
| 1
|
133,559
| 5,205,368,774
|
IssuesEvent
|
2017-01-24 17:45:39
|
RobotLocomotion/drake
|
https://api.github.com/repos/RobotLocomotion/drake
|
opened
|
Need to Add Ground Truth Object ID to Depth Sensor
|
priority: high team: sensors type: feature request
|
For every depth measurement, it'll be nice to include "ground truth" in terms of identifying what object was measured. This ground truth should ideally take the form of an integer ID, which could then be translated into something meaningful like "car", "bicycle", and "pedestrian".
|
1.0
|
Need to Add Ground Truth Object ID to Depth Sensor - For every depth measurement, it'll be nice to include "ground truth" in terms of identifying what object was measured. This ground truth should ideally take the form of an integer ID, which could then be translated into something meaningful like "car", "bicycle", and "pedestrian".
|
priority
|
need to add ground truth object id to depth sensor for every depth measurement it ll be nice to include ground truth in terms of identifying what object was measured this ground truth should ideally take the form of an integer id which could then be translated into something meaningful like car bicycle and pedestrian
| 1
|
230,024
| 7,603,426,823
|
IssuesEvent
|
2018-04-29 14:27:39
|
LazyFoxStudio/Project-2
|
https://api.github.com/repos/LazyFoxStudio/Project-2
|
opened
|
Camera bugs (black screen)
|
Gameplay bug high priority
|
When you move the camera to the top or left of the map. The screen turns black
|
1.0
|
Camera bugs (black screen) - When you move the camera to the top or left of the map. The screen turns black
|
priority
|
camera bugs black screen when you move the camera to the top or left of the map the screen turns black
| 1
|
338,191
| 10,225,518,273
|
IssuesEvent
|
2019-08-16 15:20:37
|
GeekyAnts/NativeBase
|
https://api.github.com/repos/GeekyAnts/NativeBase
|
closed
|
v2.12.2 refers to FontAwesome5Pro
|
high priority
|
v2.12.2 imports react-native-vector-icons/FontAwesome5Pro which breaks on Expo (probably all Expo versions) where package @expo/vector-icons override react-native-vector-icons.
v2.12.1 which has no FontAwesome5Pro import (only FontAwesome5) works.
Comment in other issue: https://github.com/GeekyAnts/NativeBase/issues/2731#issuecomment-512796306_
Quick solution: downgrade to v2.12.1.
|
1.0
|
v2.12.2 refers to FontAwesome5Pro - v2.12.2 imports react-native-vector-icons/FontAwesome5Pro which breaks on Expo (probably all Expo versions) where package @expo/vector-icons override react-native-vector-icons.
v2.12.1 which has no FontAwesome5Pro import (only FontAwesome5) works.
Comment in other issue: https://github.com/GeekyAnts/NativeBase/issues/2731#issuecomment-512796306_
Quick solution: downgrade to v2.12.1.
|
priority
|
refers to imports react native vector icons which breaks on expo probably all expo versions where package expo vector icons override react native vector icons which has no import only works comment in other issue quick solution downgrade to
| 1
|
327,685
| 9,978,951,100
|
IssuesEvent
|
2019-07-09 21:15:22
|
Baystation12/Baystation12
|
https://api.github.com/repos/Baystation12/Baystation12
|
closed
|
RP-Rev recruiting is not working.
|
Priority: High ⚠
|
#### Description of issue
Recruiting a revolutioner is not working
#### Difference between expected and actual behavior
Being recruted and pushing "Yes" button should make you a rev. It doesn't.
#### Steps to reproduce
>play RP-Rev round
>accept the proposition to join revolution
#### Specific information for locating
It happened with all of headrevs and all of people who tried to accept the proposition.
#### Length of time in which bug has been known to occur
I've only spotted this during one round because it was first time me playing rev and taking part in conversion.
#### Client version, Server revision & Game ID
Client Version: 510
Server Revision: 1f54f8a2655ed900b64da48a0d24b1dea24d3129 - dev -
Game ID: bMS-cJu5
Current map: SEV Torch
#### Issue bingo
Please check whatever applies. More checkboxes checked increase your chances of the issue being looked at sooner.
- [X] Issue could be reproduced at least once
- [X] Issue could be reproduced by different players
- [?] Issue could be reproduced in multiple rounds
- [X] Issue happened in a recent (less than 7 days ago) round
- [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
|
1.0
|
RP-Rev recruiting is not working. - #### Description of issue
Recruiting a revolutioner is not working
#### Difference between expected and actual behavior
Being recruted and pushing "Yes" button should make you a rev. It doesn't.
#### Steps to reproduce
>play RP-Rev round
>accept the proposition to join revolution
#### Specific information for locating
It happened with all of headrevs and all of people who tried to accept the proposition.
#### Length of time in which bug has been known to occur
I've only spotted this during one round because it was first time me playing rev and taking part in conversion.
#### Client version, Server revision & Game ID
Client Version: 510
Server Revision: 1f54f8a2655ed900b64da48a0d24b1dea24d3129 - dev -
Game ID: bMS-cJu5
Current map: SEV Torch
#### Issue bingo
Please check whatever applies. More checkboxes checked increase your chances of the issue being looked at sooner.
- [X] Issue could be reproduced at least once
- [X] Issue could be reproduced by different players
- [?] Issue could be reproduced in multiple rounds
- [X] Issue happened in a recent (less than 7 days ago) round
- [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
|
priority
|
rp rev recruiting is not working description of issue recruiting a revolutioner is not working difference between expected and actual behavior being recruted and pushing yes button should make you a rev it doesn t steps to reproduce play rp rev round accept the proposition to join revolution specific information for locating it happened with all of headrevs and all of people who tried to accept the proposition length of time in which bug has been known to occur i ve only spotted this during one round because it was first time me playing rev and taking part in conversion client version server revision game id client version server revision dev game id bms current map sev torch issue bingo please check whatever applies more checkboxes checked increase your chances of the issue being looked at sooner issue could be reproduced at least once issue could be reproduced by different players issue could be reproduced in multiple rounds issue happened in a recent less than days ago round
| 1
|
282,213
| 8,704,534,204
|
IssuesEvent
|
2018-12-05 19:40:41
|
AICrowd/AIcrowd
|
https://api.github.com/repos/AICrowd/AIcrowd
|
closed
|
Add GDPR button
|
feature high priority
|
_From @seanfcarroll on May 14, 2018 07:45_
_Copied from original issue: crowdAI/crowdai#784_
|
1.0
|
Add GDPR button - _From @seanfcarroll on May 14, 2018 07:45_
_Copied from original issue: crowdAI/crowdai#784_
|
priority
|
add gdpr button from seanfcarroll on may copied from original issue crowdai crowdai
| 1
|
436,957
| 12,556,700,456
|
IssuesEvent
|
2020-06-07 10:36:57
|
modernweb-dev/web
|
https://api.github.com/repos/modernweb-dev/web
|
opened
|
[test-runner-*] dedupe syntax errors
|
enhancement priority: high
|
Syntax errors are now logged by es-dev-server, so they are not deduped cross browsers. We should pull this logic into WTR and dedupe them.
|
1.0
|
[test-runner-*] dedupe syntax errors - Syntax errors are now logged by es-dev-server, so they are not deduped cross browsers. We should pull this logic into WTR and dedupe them.
|
priority
|
dedupe syntax errors syntax errors are now logged by es dev server so they are not deduped cross browsers we should pull this logic into wtr and dedupe them
| 1
|
608,657
| 18,845,243,241
|
IssuesEvent
|
2021-11-11 14:17:43
|
boostcampwm-2021/iOS05-Escaper
|
https://api.github.com/repos/boostcampwm-2021/iOS05-Escaper
|
closed
|
탭바, 네비게이션바 세팅 및 iOS13 위치 정보 인증 개선
|
feature High Priority
|
- 탭바, 네비게이션바 iOS 13,14 와 15가 backgroud.color가 다름.
- 네비게이션바 iOS15에서 스크롤 시 backgroud.color가 변경됨
- 유저 위치 정보 인증
- iOS13에서 첫 접속시 delegate 함수를 진행하지 않음
|
1.0
|
탭바, 네비게이션바 세팅 및 iOS13 위치 정보 인증 개선 - - 탭바, 네비게이션바 iOS 13,14 와 15가 backgroud.color가 다름.
- 네비게이션바 iOS15에서 스크롤 시 backgroud.color가 변경됨
- 유저 위치 정보 인증
- iOS13에서 첫 접속시 delegate 함수를 진행하지 않음
|
priority
|
탭바 네비게이션바 세팅 및 위치 정보 인증 개선 탭바 네비게이션바 ios 와 backgroud color가 다름 네비게이션바 스크롤 시 backgroud color가 변경됨 유저 위치 정보 인증 첫 접속시 delegate 함수를 진행하지 않음
| 1
|
375,947
| 11,136,336,383
|
IssuesEvent
|
2019-12-20 16:21:54
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
We should test __cuda_array_interface__ stride handling
|
high priority module: cuda triage review
|
https://github.com/pytorch/pytorch/pull/24947 fixed a bug around how PyTorch handles contiguous tensor strides in __cuda_array_interface__, but didn't add a test for it. We should add a test for it.
cc @ezyang @gchanan @zou3519 @ngimel
|
1.0
|
We should test __cuda_array_interface__ stride handling - https://github.com/pytorch/pytorch/pull/24947 fixed a bug around how PyTorch handles contiguous tensor strides in __cuda_array_interface__, but didn't add a test for it. We should add a test for it.
cc @ezyang @gchanan @zou3519 @ngimel
|
priority
|
we should test cuda array interface stride handling fixed a bug around how pytorch handles contiguous tensor strides in cuda array interface but didn t add a test for it we should add a test for it cc ezyang gchanan ngimel
| 1
|
302,046
| 9,254,537,614
|
IssuesEvent
|
2019-03-16 00:21:28
|
ESGF/esgf-compute-wps
|
https://api.github.com/repos/ESGF/esgf-compute-wps
|
closed
|
Need to fix the webapp wps execute
|
area/frontend kind/bug priority/high size/small
|
**Describe the bug**
The Execute request generated by the webapp is not properly formatted, it's using LiteralData when the service is expecting ComplexData.
**To Reproduce**
Submit a job using the webapp.
**Expected behavior**
Should successfully submit a properly formatted Execute POST request.
**Screenshots**
n/a
**Additional context**
n/a
|
1.0
|
Need to fix the webapp wps execute - **Describe the bug**
The Execute request generated by the webapp is not properly formatted, it's using LiteralData when the service is expecting ComplexData.
**To Reproduce**
Submit a job using the webapp.
**Expected behavior**
Should successfully submit a properly formatted Execute POST request.
**Screenshots**
n/a
**Additional context**
n/a
|
priority
|
need to fix the webapp wps execute describe the bug the execute request generated by the webapp is not properly formatted it s using literaldata when the service is expecting complexdata to reproduce submit a job using the webapp expected behavior should successfully submit a properly formatted execute post request screenshots n a additional context n a
| 1
|
240,930
| 7,807,424,065
|
IssuesEvent
|
2018-06-11 16:48:36
|
ministryofjustice/cloud-platform-roadmap
|
https://api.github.com/repos/ministryofjustice/cloud-platform-roadmap
|
opened
|
Monitoring capacity of cluster
|
high-priority
|
## Background
As we add more things to the cluster we need to know when to add more nodes. The first metric to do this on the back of will be the amount of memory/cpu available on the cluster overall. We need to be able to:
- measure that metric,
- alert when it is nearing capacity and
- have a process for adding more resources when we are near capacity
## Proposed user journey
- Cloud platform team member gets alert that cluster is at e.g. 80% capacity for CPU and Memory
- Team member goes to a run book
- Team member provisions a new node
- Alert stops firing as the new node comes into service and overall capacity increases
## Approach
## Questions / Assumptions
<!-- Additional information to explain approach taken -->
## Definition of done
<!-- Checklist for definition of done and acceptance criteria, for example: -->
- [ ] must compile
- [ ] must pass tests
- [ ] must address all the steps of user journey
## Reference
[How to write good user stories](https://www.gov.uk/service-manual/agile-delivery/writing-user-stories)
|
1.0
|
Monitoring capacity of cluster - ## Background
As we add more things to the cluster we need to know when to add more nodes. The first metric to do this on the back of will be the amount of memory/cpu available on the cluster overall. We need to be able to:
- measure that metric,
- alert when it is nearing capacity and
- have a process for adding more resources when we are near capacity
## Proposed user journey
- Cloud platform team member gets alert that cluster is at e.g. 80% capacity for CPU and Memory
- Team member goes to a run book
- Team member provisions a new node
- Alert stops firing as the new node comes into service and overall capacity increases
## Approach
## Questions / Assumptions
<!-- Additional information to explain approach taken -->
## Definition of done
<!-- Checklist for definition of done and acceptance criteria, for example: -->
- [ ] must compile
- [ ] must pass tests
- [ ] must address all the steps of user journey
## Reference
[How to write good user stories](https://www.gov.uk/service-manual/agile-delivery/writing-user-stories)
|
priority
|
monitoring capacity of cluster background as we add more things to the cluster we need to know when to add more nodes the first metric to do this on the back of will be the amount of memory cpu available on the cluster overall we need to be able to measure that metric alert when it is nearing capacity and have a process for adding more resources when we are near capacity proposed user journey cloud platform team member gets alert that cluster is at e g capacity for cpu and memory team member goes to a run book team member provisions a new node alert stops firing as the new node comes into service and overall capacity increases approach questions assumptions definition of done must compile must pass tests must address all the steps of user journey reference
| 1
|
179,250
| 6,622,878,820
|
IssuesEvent
|
2017-09-22 03:02:10
|
minio/mint
|
https://api.github.com/repos/minio/mint
|
closed
|
failure in aws-sdk-ruby
|
fixed priority: high
|
aws-sdk-ruby fails with the log
```
{"name":"aws-sdk-ruby","function":"listBuckets()","args":{},"duration":31.82,"status":"PASS"}
{"name":"aws-sdk-ruby","function":"removeObjects(bucket_name)","args":{"bucket_name":"4ed81b3f25bc"},"duration":173.16,"error":"Failed to clean-up bucket '4ed81b3f25bc': undefined method `merge' for \"datafile-1-MB\":String","status":"FAIL"}
```
|
1.0
|
failure in aws-sdk-ruby - aws-sdk-ruby fails with the log
```
{"name":"aws-sdk-ruby","function":"listBuckets()","args":{},"duration":31.82,"status":"PASS"}
{"name":"aws-sdk-ruby","function":"removeObjects(bucket_name)","args":{"bucket_name":"4ed81b3f25bc"},"duration":173.16,"error":"Failed to clean-up bucket '4ed81b3f25bc': undefined method `merge' for \"datafile-1-MB\":String","status":"FAIL"}
```
|
priority
|
failure in aws sdk ruby aws sdk ruby fails with the log name aws sdk ruby function listbuckets args duration status pass name aws sdk ruby function removeobjects bucket name args bucket name duration error failed to clean up bucket undefined method merge for datafile mb string status fail
| 1
|
274,002
| 8,555,996,603
|
IssuesEvent
|
2018-11-08 11:44:51
|
BeamMW/beam
|
https://api.github.com/repos/BeamMW/beam
|
opened
|
Failed transactions on local node
|
High priority bug
|
Precon: wallet 1 is connected to local node
wallet 2 is connected to testnet node
Steps:
1. Run wallet 1
2. While it's syncronyzing send beams from wallet 2 to wallet 1
3. Wait for some minutes
Actual: transaction is changed from Pending to Failed
4. Try to send one more transaction
Actual: transaction is failed
[wallet_18_11_08_13_30_59.log](https://github.com/BeamMW/beam/files/2561547/wallet_18_11_08_13_30_59.log)
Transactions are also failed when node is synchronized
|
1.0
|
Failed transactions on local node - Precon: wallet 1 is connected to local node
wallet 2 is connected to testnet node
Steps:
1. Run wallet 1
2. While it's syncronyzing send beams from wallet 2 to wallet 1
3. Wait for some minutes
Actual: transaction is changed from Pending to Failed
4. Try to send one more transaction
Actual: transaction is failed
[wallet_18_11_08_13_30_59.log](https://github.com/BeamMW/beam/files/2561547/wallet_18_11_08_13_30_59.log)
Transactions are also failed when node is synchronized
|
priority
|
failed transactions on local node precon wallet is connected to local node wallet is connected to testnet node steps run wallet while it s syncronyzing send beams from wallet to wallet wait for some minutes actual transaction is changed from pending to failed try to send one more transaction actual transaction is failed transactions are also failed when node is synchronized
| 1
|
810,810
| 30,262,060,188
|
IssuesEvent
|
2023-07-07 08:57:54
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
[Bug]: Goto Definition not working for regexp lang lib
|
Type/Bug Priority/High Status/Blocked Team/LanguageServer Area/Definition
|
### Description

### Steps to Reproduce
```ballerina
import ballerina/lang.regexp;
public function main() {
regexp:RegExp reg = re ``;
}
```
### Affected Version(s)
2201.4.0-SNAPSHOT
### OS, DB, other environment details and versions
_No response_
### Related area
-> Editor
### Related issue(s) (optional)
_No response_
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_
|
1.0
|
[Bug]: Goto Definition not working for regexp lang lib - ### Description

### Steps to Reproduce
```ballerina
import ballerina/lang.regexp;
public function main() {
regexp:RegExp reg = re ``;
}
```
### Affected Version(s)
2201.4.0-SNAPSHOT
### OS, DB, other environment details and versions
_No response_
### Related area
-> Editor
### Related issue(s) (optional)
_No response_
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_
|
priority
|
goto definition not working for regexp lang lib description steps to reproduce ballerina import ballerina lang regexp public function main regexp regexp reg re affected version s snapshot os db other environment details and versions no response related area editor related issue s optional no response suggested label s optional no response suggested assignee s optional no response
| 1
|
296,180
| 9,105,227,529
|
IssuesEvent
|
2019-02-20 20:13:00
|
ganeti/ganeti
|
https://api.github.com/repos/ganeti/ganeti
|
closed
|
QEMU 3.1 compatibility
|
Compatibility Priority:High
|
Debian Buster will be shipping with QEMU 3.1. Currently KVM instances fail to start because we're using `-balloon virtio`, which was removed in 3.1:
```
$ qemu-system-x86_64 -balloon virtio
qemu-system-x86_64: -balloon: invalid option
$ qemu-system-x86_64 --version
QEMU emulator version 3.1.0 (Debian 1:3.1+dfsg-2+b1)
Copyright (c) 2003-2018 Fabrice Bellard and the QEMU Project developers
```
We need to make sure Ganeti works with QEMU 3.1, if possible without using any deprecated features.
|
1.0
|
QEMU 3.1 compatibility - Debian Buster will be shipping with QEMU 3.1. Currently KVM instances fail to start because we're using `-balloon virtio`, which was removed in 3.1:
```
$ qemu-system-x86_64 -balloon virtio
qemu-system-x86_64: -balloon: invalid option
$ qemu-system-x86_64 --version
QEMU emulator version 3.1.0 (Debian 1:3.1+dfsg-2+b1)
Copyright (c) 2003-2018 Fabrice Bellard and the QEMU Project developers
```
We need to make sure Ganeti works with QEMU 3.1, if possible without using any deprecated features.
|
priority
|
qemu compatibility debian buster will be shipping with qemu currently kvm instances fail to start because we re using balloon virtio which was removed in qemu system balloon virtio qemu system balloon invalid option qemu system version qemu emulator version debian dfsg copyright c fabrice bellard and the qemu project developers we need to make sure ganeti works with qemu if possible without using any deprecated features
| 1
|
404,008
| 11,850,872,683
|
IssuesEvent
|
2020-03-24 17:15:56
|
BlocKoopa/KnightsCapstone
|
https://api.github.com/repos/BlocKoopa/KnightsCapstone
|
opened
|
Programming - Switches & Gates
|
Point: 5 Priority: Highest
|
2 parts:
1. Program Gates to block player. (Program Tiles to recognize if a gate is in the way, and block player).
2. Program Switches to remove corresponding Gates when landed on.
|
1.0
|
Programming - Switches & Gates - 2 parts:
1. Program Gates to block player. (Program Tiles to recognize if a gate is in the way, and block player).
2. Program Switches to remove corresponding Gates when landed on.
|
priority
|
programming switches gates parts program gates to block player program tiles to recognize if a gate is in the way and block player program switches to remove corresponding gates when landed on
| 1
|
614,822
| 19,190,335,074
|
IssuesEvent
|
2021-12-05 22:01:05
|
RE-SS3D/SS3D
|
https://api.github.com/repos/RE-SS3D/SS3D
|
opened
|
Implement Wire Adjacency Connections
|
Type: Feature (Addition) Asset: Script Coding: C# Priority: 2 - High Difficulty: 2 - Medium System: Tilemaps
|
<!-- The notes within these arrows are for you but can be deleted. -->
## Summary
Implement a new tilemap adjacency connection script (similar to the others) for "wire connections". This should follow the design located in the link below.
## Goal
This will allow for wires to be added to the map via the editor and perform intended connections.
https://docs.google.com/document/d/1ful7_gIJo7e74i9LMQuYMpjZMH2V1aT90mwlrQ0hcgE/#heading=h.15jb46xsi8l2
|
1.0
|
Implement Wire Adjacency Connections - <!-- The notes within these arrows are for you but can be deleted. -->
## Summary
Implement a new tilemap adjacency connection script (similar to the others) for "wire connections". This should follow the design located in the link below.
## Goal
This will allow for wires to be added to the map via the editor and perform intended connections.
https://docs.google.com/document/d/1ful7_gIJo7e74i9LMQuYMpjZMH2V1aT90mwlrQ0hcgE/#heading=h.15jb46xsi8l2
|
priority
|
implement wire adjacency connections summary implement a new tilemap adjacency connection script similar to the others for wire connections this should follow the design located in the link below goal this will allow for wires to be added to the map via the editor and perform intended connections
| 1
|
424,250
| 12,308,298,950
|
IssuesEvent
|
2020-05-12 06:55:51
|
GrassrootsEconomics/CIC-Docs
|
https://api.github.com/repos/GrassrootsEconomics/CIC-Docs
|
opened
|
From and To on reciept messages
|
enhancement priority:high
|
**Is your feature request related to a problem? Please describe.**
we have old ladies swapping sim cards and they are VERY confused .... because on the same phone from 'Sarafu' they can't tell whom sent what to whom and they are swapping sim cards
not all users have phones but they do have a sim
**Describe the solution you'd like**
we need on the receipt message to say both the sender and receiver...
Sent receipt
```Successfully sent a payment of 1000.00 SARAFU to Lydia Wanjiku +254720305812 on 11/05/2020 at 09:45 PM from Jim Wanger +254720303811. New SARAFU balance is 1236.37
```
received receipt
```Successfully received a payment of 1000.00 SARAFU from Wanjiru Watheka +254702875427 on 11/05/2020 at 09:45 PM to Sayy Hornsby +254720303822. New SARAFU balance is 1809.01```
|
1.0
|
From and To on reciept messages - **Is your feature request related to a problem? Please describe.**
we have old ladies swapping sim cards and they are VERY confused .... because on the same phone from 'Sarafu' they can't tell whom sent what to whom and they are swapping sim cards
not all users have phones but they do have a sim
**Describe the solution you'd like**
we need on the receipt message to say both the sender and receiver...
Sent receipt
```Successfully sent a payment of 1000.00 SARAFU to Lydia Wanjiku +254720305812 on 11/05/2020 at 09:45 PM from Jim Wanger +254720303811. New SARAFU balance is 1236.37
```
received receipt
```Successfully received a payment of 1000.00 SARAFU from Wanjiru Watheka +254702875427 on 11/05/2020 at 09:45 PM to Sayy Hornsby +254720303822. New SARAFU balance is 1809.01```
|
priority
|
from and to on reciept messages is your feature request related to a problem please describe we have old ladies swapping sim cards and they are very confused because on the same phone from sarafu they can t tell whom sent what to whom and they are swapping sim cards not all users have phones but they do have a sim describe the solution you d like we need on the receipt message to say both the sender and receiver sent receipt successfully sent a payment of sarafu to lydia wanjiku on at pm from jim wanger new sarafu balance is received receipt successfully received a payment of sarafu from wanjiru watheka on at pm to sayy hornsby new sarafu balance is
| 1
|
54,117
| 3,060,050,758
|
IssuesEvent
|
2015-08-14 18:24:55
|
loklak/loklak_webclient
|
https://api.github.com/repos/loklak/loklak_webclient
|
closed
|
wall not working on loklak.net
|
bug Priority 1 - High Twitter Wall - Aneesh
|
After I moved test.loklak.net to loklak.net, the wall is not working any more. The javascript terminal shows the following error:
```
"Error: [$rootScope:inprog] $apply already in progress
http://errors.angularjs.org/1.3.16/$rootScope/inprog?p0=%24apply
i/<@http://loklak.net/js/main.js:30:1177
f@http://loklak.net/js/main.js:31:30933
Ht/this.$get</h.prototype.$apply@http://loklak.net/js/main.js:33:2968
["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</<@http://loklak.net/js/main.js:35:18841
jQuery.event.dispatch@http://code.jquery.com/jquery-2.1.4.js:4434:15
jQuery.event.add/elemData.handle@http://code.jquery.com/jquery-2.1.4.js:4121:6
jQuery.event.trigger@http://code.jquery.com/jquery-2.1.4.js:4350:5
.trigger/<@http://code.jquery.com/jquery-2.1.4.js:4901:4
.each@http://code.jquery.com/jquery-2.1.4.js:374:14
jQuery.prototype.each@http://code.jquery.com/jquery-2.1.4.js:139:10
.trigger@http://code.jquery.com/jquery-2.1.4.js:4900:1
i/e.proceed@http://loklak.net/js/main.js:22:16928
["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</va.prototype.functionCall/<@http://loklak.net/js/main.js:35:12265
["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</</a@http://loklak.net/js/main.js:35:18791
Ht/this.$get</h.prototype.$eval@http://loklak.net/js/main.js:33:2755
Ht/this.$get</h.prototype.$apply@http://loklak.net/js/main.js:33:2980
["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</<@http://loklak.net/js/main.js:35:18841
jQuery.event.dispatch@http://code.jquery.com/jquery-2.1.4.js:4434:15
jQuery.event.add/elemData.handle@http://code.jquery.com/jquery-2.1.4.js:4121:6
"
```
|
1.0
|
wall not working on loklak.net - After I moved test.loklak.net to loklak.net, the wall is not working any more. The javascript terminal shows the following error:
```
"Error: [$rootScope:inprog] $apply already in progress
http://errors.angularjs.org/1.3.16/$rootScope/inprog?p0=%24apply
i/<@http://loklak.net/js/main.js:30:1177
f@http://loklak.net/js/main.js:31:30933
Ht/this.$get</h.prototype.$apply@http://loklak.net/js/main.js:33:2968
["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</<@http://loklak.net/js/main.js:35:18841
jQuery.event.dispatch@http://code.jquery.com/jquery-2.1.4.js:4434:15
jQuery.event.add/elemData.handle@http://code.jquery.com/jquery-2.1.4.js:4121:6
jQuery.event.trigger@http://code.jquery.com/jquery-2.1.4.js:4350:5
.trigger/<@http://code.jquery.com/jquery-2.1.4.js:4901:4
.each@http://code.jquery.com/jquery-2.1.4.js:374:14
jQuery.prototype.each@http://code.jquery.com/jquery-2.1.4.js:139:10
.trigger@http://code.jquery.com/jquery-2.1.4.js:4900:1
i/e.proceed@http://loklak.net/js/main.js:22:16928
["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</va.prototype.functionCall/<@http://loklak.net/js/main.js:35:12265
["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</</a@http://loklak.net/js/main.js:35:18791
Ht/this.$get</h.prototype.$eval@http://loklak.net/js/main.js:33:2755
Ht/this.$get</h.prototype.$apply@http://loklak.net/js/main.js:33:2980
["/home/loklaktest/loklak_webclient/node_modules/angular/angular.js"]</</</or[t]</<.compile/</<@http://loklak.net/js/main.js:35:18841
jQuery.event.dispatch@http://code.jquery.com/jquery-2.1.4.js:4434:15
jQuery.event.add/elemData.handle@http://code.jquery.com/jquery-2.1.4.js:4121:6
"
```
|
priority
|
wall not working on loklak net after i moved test loklak net to loklak net the wall is not working any more the javascript terminal shows the following error error apply already in progress i f ht this get h prototype apply or compile jquery event dispatch jquery event add elemdata handle jquery event trigger trigger each jquery prototype each trigger i e proceed va prototype functioncall or compile a ht this get h prototype eval ht this get h prototype apply or compile jquery event dispatch jquery event add elemdata handle
| 1
|
637,329
| 20,625,531,265
|
IssuesEvent
|
2022-03-07 22:04:33
|
microsoft/fluentui
|
https://api.github.com/repos/microsoft/fluentui
|
closed
|
Project: Manual accessibility review for vNext Beta/RC
|
Area: Accessibility Priority 1: High Fluent UI vNext
|
## Problem
Fluent vNext components could use an accessibility review before we release them in Beta, just to catch any potential issues while they're still easy to change.
## Appetite
2 weeks
## Solution
Review components that have UI and are included in the initial beta release, and raise any necessary issues
- [x] Accordion #21101
- [x] Avatar #21102
- [x] Badge (includes CounterBadge and PresenceBadge) #21103
- [x] Button (includes CompoundButton, MenuButton, SplitButton, and ToggleButton) #20917
- [ ] Card #21104
- [x] Divider #21105
- [x] Link #19926
- [x] Menu #21106
- [x] Popover #21107
- [x] Tooltip #21108
## Risks (Rabbit holes)
N/A
## Out of scope (No-gos)
N/A
|
1.0
|
Project: Manual accessibility review for vNext Beta/RC - ## Problem
Fluent vNext components could use an accessibility review before we release them in Beta, just to catch any potential issues while they're still easy to change.
## Appetite
2 weeks
## Solution
Review components that have UI and are included in the initial beta release, and raise any necessary issues
- [x] Accordion #21101
- [x] Avatar #21102
- [x] Badge (includes CounterBadge and PresenceBadge) #21103
- [x] Button (includes CompoundButton, MenuButton, SplitButton, and ToggleButton) #20917
- [ ] Card #21104
- [x] Divider #21105
- [x] Link #19926
- [x] Menu #21106
- [x] Popover #21107
- [x] Tooltip #21108
## Risks (Rabbit holes)
N/A
## Out of scope (No-gos)
N/A
|
priority
|
project manual accessibility review for vnext beta rc problem fluent vnext components could use an accessibility review before we release them in beta just to catch any potential issues while they re still easy to change appetite weeks solution review components that have ui and are included in the initial beta release and raise any necessary issues accordion avatar badge includes counterbadge and presencebadge button includes compoundbutton menubutton splitbutton and togglebutton card divider link menu popover tooltip risks rabbit holes n a out of scope no gos n a
| 1
|
314,430
| 9,596,903,720
|
IssuesEvent
|
2019-05-09 19:51:53
|
fgpv-vpgf/fgpv-vpgf
|
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
|
closed
|
Design Options for JSON File Data Table
|
addition: feature feedback: discussion priority: high
|
Original request can be found in https://github.com/fgpv-vpgf/fgpv-vpgf/issues/1415
Things have changed a bit since that request, and we are now looking at supporting layers that traditionally had no attribute support at all (i.e. adding a table feature instead of just making existing table features faster).
Basically, the concept is for a layer we can target a json file to use as a source to populate the data table for that layer. There are a number of considerations for how fancy we make this.
## Level of Integration
### Unlinked
In this case, the data table is essentially independent of the layer, with the exception that clicking the layer in the legend will open the table. Any additional functions (e.g. zoom to feature, details, icons) will need to be addressed for the unlinked case (details in sections below).
This case is a :bear: minimum to implement.
### Linked
In this case, the data table (and all other functions that run off attributes) remain connected to the map layer. The only difference is when the app goes to download the attribute set, it will source it from the json file instead of querying the map service. ObjectIds in the json file must align to the Ids on the map service.
Why use a json file when the service already exists? Some benefits include faster load time, and possibly supporting Dynamic Layers on servers that have the `/query` endpoint disabled.
The linked case will only work on layers that already support ESRI attributes (i.e. anything that inherits from `AttribRecord` -- mainly Feature Layers and Dynamic Layers).
Supporting the linked mode is more work, but lets us hit additional use cases. If we decide not to implement in the first cut of file based tables, we should at least attempt to design the new code to allow support in the future.
## File Schema
### ESRI-like Dictionary
This approach stores attributes in the JSON file using a similar/identical structure to how ArcGIS Server provides attributes. The pro of this approach is our current parsing routines can eat it raw. We can also create content from existing ESRI sources with relative ease. The con is the file size will be larger as attribute keys are repeated for every entry.
```text
{
fields: [
{
name: "Field1",
type: "esriFieldTypeString",
alias: "Field One",
length: 200
},
{
name: "Field2",
type: "esriFieldTypeDouble",
alias: "Field Two"
}
],
features: [
{
attributes: {
Field1: "Big Duck Pond",
Field2: 12.82
}
},
{
attributes: {
Field1: "Germy Lake",
Field2: 4.98
}
}
]
}
```
### Array Based Structure
This approach stores data in an array form (making it look similar to a table). The pro of this approach is the file size is minimized. The con is we need to write parsing code, and possibly it is more effort to translate date into this structure when generating the files.
```text
{
fields: [
{
name: "Field1",
type: "esriFieldTypeString",
alias: "Field One",
length: 200
},
{
name: "Field2",
type: "esriFieldTypeDouble",
alias: "Field Two"
}
],
features: [
[ "Big Duck Pond", 12.82 ],
[ "Germy Lake", 4.98 ]
]
}
```
As it stands, this is the recommended approach for speed gains.
### Split Structure
This approach separates the bulk data and the metadata into two files. This lets us quickly download table metadata (e.g. field names) up front if we need them. Whether this is necessary depends on how much support we give to non-attribute layer types (e.g. tabe on WMS, table on ESRI Tile). If we essentially disable everything but "table view" then it may be unnecessary.
## Config Structure
### Core Elements
Need an optional property on a `layerNode` element to allow a url to a json file to be defined. The presence of a value in that property will trigger the appropriate setup for a file based grid (see sections below).
### Optional Elements
Potential for additional properties to define enhanced things.
* Marking a lat & long field in the data to enable a zoom to point function
* An indicator if the file is linked to the layer, or is independent (e.g. object ids in file align to object ids on service). Should default to unlinked for safety.
* Ability to provide an icon for each row. This could be enhanced to fancier structures (essentially providing a class break or unique value renderer type thing); not recommended at this point.
* Ability to specify a name field? Would be needed if we want to support a details pane view with a reliable title value (See `Details Pane` section below for more analysis).
If we are considering adding options beyond the url, we may want to create a sub-object to contain all the json table properties.
E.g. (field names not final):
```text
{
id: "myLayer",
url: "http://myserver.com/fancymap/",
jsonTable: {
url: "http://myserver.com/fancytable.json",
linked: false,
icon: "http://myserver.com/happyDuck.png"
}
}
```
## Loading Mechanism
### Considerations
* Files must be loaded in the Angular code (using `$http.get()`). The request library in geoApi fails with files due to the proxy being cooked into `esriRequest`. (See https://github.com/fgpv-vpgf/fgpv-vpgf/issues/2061#issuecomment-313451883 )
* The results can be provided from the same function within geoApi as normal attributes. See `attribute.js - newLayerPackage()`.
* The loading from the server should not begin until `layerPackage.getAttribs()` is first called.
* This approach allows for existing workflows regarding the table / attributes to be re-used with minimal impact
* The results can be provided via a separate mechanism
* Would only support unlinked mode
* The "load the table" step would recognize its the json file case and call the alternate method to get data
* Allows us to ignore compatibility issues of trying to have traditional attributes and json file attributes existing harmoniously in the same workflows
### Implementations
If attempting to integrate with the standard `getAttribs()` approach:
* We want the actual web request in angular, so that function will exist in the viewer. However, we trigger our attribute load from the geoApi.
* Suggest our new geoApi function to setup the file table mode should accept a loader function. This way, when the geoApi is ready to load the table (usually first time opening the table), it can just execute the registered loader function.
* Makes geoApi blissfully unaware that it's using angular.
* Allows the option to switch to a different load approach later, as long as the function interface remains the same.
* Likely need an alternate form of `newLayerPackage`, or need to overwrite it's `getAttribs` function.
* The new function needs to see if `_attribData` exists, and if not, it needs to make a new Promise
* Inside new Promise: execute the registered function to trigger the $http.get(), then massage result from json file format into esri format
* Resolve Promise with proper object (see `createAttribSet`)
* Likely need a third function similar to `loadFileAttribsBuilder` and `loadServerAttribsBuilder` (probably closer to `loadFileAttribsBuilder` in content).
* It will set the various metadata properties correctly.
* Need to be aware of sources, particularly if we are dealing with a layer that doesn't natively have attributes (e.g. fields would have to come from file, so they may not be set upfront)
If taking the separate mechanism route -- we can pretty much do whatever we please.
## Layer Objects
This new feature has the potential to complicate our layer hierarchy in geoApi. Attribute support is currently confined to the `AttribRecord` class and all inheritors. File based tables means classes not derived from `AttribRecord` now require some level of support.
### Minimal Approach - Integrated
This approach essentially fakes the attribute loader on layers not related to `AttribRecord`. It would not use the class inheritance, and would instead inject appropriate functions onto Record/FC objects to make things look compatible.
We can also entertain doing this on `AttribRecord` based classes, where we overwrite existing attribute loader logic with file loader logic on a case-by-case basis. This is a bit risky in that there are quite a few related functions that depend on attributes and may break / act weird.
### Minimal Approach - Separate Mechanism
This ties into the separate mechanism approach described in the `Loading Mechanism` section above. Basically the json file to grid flow is separate from all existing structure.
This approach will only support the Unlinked mode. So it causes the least amount of chaos, but offers the least amount of value.
### Reeefactor
This involves significant changes to the Layer Record / FC object hierarchies. While it would need further analysis if we decide to persue this, at a high level:
* Attribute support would need to exist at a low level in the hierarchy. The `LayerRecord` class would make sense.
* The `LayerRecord` implementation could have the file based load support. If no file has been defined in the config, it can return an empty array or an error (to be decided which is more appropriate).
* The `AttribRecord` class would retain it's code for supporting ArcGIS Server based attribute loading. However, it would need to allow for recognizing the file based case, and let the logic drop through to `LayerRecord`. In other words, it can't just steamroll the attribute logic; it needs to allow both branches of code to execute depending on the configuration.
This approach offers good support for all layer types, and best supports the Linked mode (having file that aligns with server data).
## Details Pane
In unlinked mode, we have a few choices
* Disable the details pane. The data is in the grid anyways so its just showing the same thing but in vertical layout.
* Lazy!
* Should remove the detail button from the grid.
* Push data to the standard zebra-stripe attribute view.
* May need to disable things like zoom to feature and icon, depending on what configuration options we've added.
* May also have to guess or suppress the "name field" if we have decided to not make that a config option (or it was not provided)
* Could make the detail pane a configuration option.
In linked mode, everything should work as normal as the data will be aligned to the layer metadata.
## Zoom To Feature
In unlinked mode:
* Default behavior would be no zoom. Button should not be in the grid.
* Can have optional lat-long columns defined in config to zoom to point. Would involve a projection and then the zoom to point pipeline.
* If geometry is a line/poly, only real option is to use the lat-long as centroid and accept that zoom level may be a poor choice for the shape.
In linked mode, everything should work as normal as the data will be aligned to the layer metadata.
## Hovertips
In the case of having a layer that natively supports attributes, and we are using a json sourced table, and the table is not aligned to the service, we would need to disable hovertips, or force them to always use the one-shot quickloader (essentially ignoring the attribute store even though it exists).
|
1.0
|
Design Options for JSON File Data Table - Original request can be found in https://github.com/fgpv-vpgf/fgpv-vpgf/issues/1415
Things have changed a bit since that request, and we are now looking at supporting layers that traditionally had no attribute support at all (i.e. adding a table feature instead of just making existing table features faster).
Basically, the concept is for a layer we can target a json file to use as a source to populate the data table for that layer. There are a number of considerations for how fancy we make this.
## Level of Integration
### Unlinked
In this case, the data table is essentially independent of the layer, with the exception that clicking the layer in the legend will open the table. Any additional functions (e.g. zoom to feature, details, icons) will need to be addressed for the unlinked case (details in sections below).
This case is a :bear: minimum to implement.
### Linked
In this case, the data table (and all other functions that run off attributes) remain connected to the map layer. The only difference is when the app goes to download the attribute set, it will source it from the json file instead of querying the map service. ObjectIds in the json file must align to the Ids on the map service.
Why use a json file when the service already exists? Some benefits include faster load time, and possibly supporting Dynamic Layers on servers that have the `/query` endpoint disabled.
The linked case will only work on layers that already support ESRI attributes (i.e. anything that inherits from `AttribRecord` -- mainly Feature Layers and Dynamic Layers).
Supporting the linked mode is more work, but lets us hit additional use cases. If we decide not to implement in the first cut of file based tables, we should at least attempt to design the new code to allow support in the future.
## File Schema
### ESRI-like Dictionary
This approach stores attributes in the JSON file using a similar/identical structure to how ArcGIS Server provides attributes. The pro of this approach is our current parsing routines can eat it raw. We can also create content from existing ESRI sources with relative ease. The con is the file size will be larger as attribute keys are repeated for every entry.
```text
{
fields: [
{
name: "Field1",
type: "esriFieldTypeString",
alias: "Field One",
length: 200
},
{
name: "Field2",
type: "esriFieldTypeDouble",
alias: "Field Two"
}
],
features: [
{
attributes: {
Field1: "Big Duck Pond",
Field2: 12.82
}
},
{
attributes: {
Field1: "Germy Lake",
Field2: 4.98
}
}
]
}
```
### Array Based Structure
This approach stores data in an array form (making it look similar to a table). The pro of this approach is the file size is minimized. The con is we need to write parsing code, and possibly it is more effort to translate date into this structure when generating the files.
```text
{
fields: [
{
name: "Field1",
type: "esriFieldTypeString",
alias: "Field One",
length: 200
},
{
name: "Field2",
type: "esriFieldTypeDouble",
alias: "Field Two"
}
],
features: [
[ "Big Duck Pond", 12.82 ],
[ "Germy Lake", 4.98 ]
]
}
```
As it stands, this is the recommended approach for speed gains.
### Split Structure
This approach separates the bulk data and the metadata into two files. This lets us quickly download table metadata (e.g. field names) up front if we need them. Whether this is necessary depends on how much support we give to non-attribute layer types (e.g. tabe on WMS, table on ESRI Tile). If we essentially disable everything but "table view" then it may be unnecessary.
## Config Structure
### Core Elements
Need an optional property on a `layerNode` element to allow a url to a json file to be defined. The presence of a value in that property will trigger the appropriate setup for a file based grid (see sections below).
### Optional Elements
Potential for additional properties to define enhanced things.
* Marking a lat & long field in the data to enable a zoom to point function
* An indicator if the file is linked to the layer, or is independent (e.g. object ids in file align to object ids on service). Should default to unlinked for safety.
* Ability to provide an icon for each row. This could be enhanced to fancier structures (essentially providing a class break or unique value renderer type thing); not recommended at this point.
* Ability to specify a name field? Would be needed if we want to support a details pane view with a reliable title value (See `Details Pane` section below for more analysis).
If we are considering adding options beyond the url, we may want to create a sub-object to contain all the json table properties.
E.g. (field names not final):
```text
{
id: "myLayer",
url: "http://myserver.com/fancymap/",
jsonTable: {
url: "http://myserver.com/fancytable.json",
linked: false,
icon: "http://myserver.com/happyDuck.png"
}
}
```
## Loading Mechanism
### Considerations
* Files must be loaded in the Angular code (using `$http.get()`). The request library in geoApi fails with files due to the proxy being cooked into `esriRequest`. (See https://github.com/fgpv-vpgf/fgpv-vpgf/issues/2061#issuecomment-313451883 )
* The results can be provided from the same function within geoApi as normal attributes. See `attribute.js - newLayerPackage()`.
* The loading from the server should not begin until `layerPackage.getAttribs()` is first called.
* This approach allows for existing workflows regarding the table / attributes to be re-used with minimal impact
* The results can be provided via a separate mechanism
* Would only support unlinked mode
* The "load the table" step would recognize its the json file case and call the alternate method to get data
* Allows us to ignore compatibility issues of trying to have traditional attributes and json file attributes existing harmoniously in the same workflows
### Implementations
If attempting to integrate with the standard `getAttribs()` approach:
* We want the actual web request in angular, so that function will exist in the viewer. However, we trigger our attribute load from the geoApi.
* Suggest our new geoApi function to setup the file table mode should accept a loader function. This way, when the geoApi is ready to load the table (usually first time opening the table), it can just execute the registered loader function.
* Makes geoApi blissfully unaware that it's using angular.
* Allows the option to switch to a different load approach later, as long as the function interface remains the same.
* Likely need an alternate form of `newLayerPackage`, or need to overwrite it's `getAttribs` function.
* The new function needs to see if `_attribData` exists, and if not, it needs to make a new Promise
* Inside new Promise: execute the registered function to trigger the $http.get(), then massage result from json file format into esri format
* Resolve Promise with proper object (see `createAttribSet`)
* Likely need a third function similar to `loadFileAttribsBuilder` and `loadServerAttribsBuilder` (probably closer to `loadFileAttribsBuilder` in content).
* It will set the various metadata properties correctly.
* Need to be aware of sources, particularly if we are dealing with a layer that doesn't natively have attributes (e.g. fields would have to come from file, so they may not be set upfront)
If taking the separate mechanism route -- we can pretty much do whatever we please.
## Layer Objects
This new feature has the potential to complicate our layer hierarchy in geoApi. Attribute support is currently confined to the `AttribRecord` class and all inheritors. File based tables means classes not derived from `AttribRecord` now require some level of support.
### Minimal Approach - Integrated
This approach essentially fakes the attribute loader on layers not related to `AttribRecord`. It would not use the class inheritance, and would instead inject appropriate functions onto Record/FC objects to make things look compatible.
We can also entertain doing this on `AttribRecord` based classes, where we overwrite existing attribute loader logic with file loader logic on a case-by-case basis. This is a bit risky in that there are quite a few related functions that depend on attributes and may break / act weird.
### Minimal Approach - Separate Mechanism
This ties into the separate mechanism approach described in the `Loading Mechanism` section above. Basically the json file to grid flow is separate from all existing structure.
This approach will only support the Unlinked mode. So it causes the least amount of chaos, but offers the least amount of value.
### Reeefactor
This involves significant changes to the Layer Record / FC object hierarchies. While it would need further analysis if we decide to persue this, at a high level:
* Attribute support would need to exist at a low level in the hierarchy. The `LayerRecord` class would make sense.
* The `LayerRecord` implementation could have the file based load support. If no file has been defined in the config, it can return an empty array or an error (to be decided which is more appropriate).
* The `AttribRecord` class would retain it's code for supporting ArcGIS Server based attribute loading. However, it would need to allow for recognizing the file based case, and let the logic drop through to `LayerRecord`. In other words, it can't just steamroll the attribute logic; it needs to allow both branches of code to execute depending on the configuration.
This approach offers good support for all layer types, and best supports the Linked mode (having file that aligns with server data).
## Details Pane
In unlinked mode, we have a few choices
* Disable the details pane. The data is in the grid anyways so its just showing the same thing but in vertical layout.
* Lazy!
* Should remove the detail button from the grid.
* Push data to the standard zebra-stripe attribute view.
* May need to disable things like zoom to feature and icon, depending on what configuration options we've added.
* May also have to guess or suppress the "name field" if we have decided to not make that a config option (or it was not provided)
* Could make the detail pane a configuration option.
In linked mode, everything should work as normal as the data will be aligned to the layer metadata.
## Zoom To Feature
In unlinked mode:
* Default behavior would be no zoom. Button should not be in the grid.
* Can have optional lat-long columns defined in config to zoom to point. Would involve a projection and then the zoom to point pipeline.
* If geometry is a line/poly, only real option is to use the lat-long as centroid and accept that zoom level may be a poor choice for the shape.
In linked mode, everything should work as normal as the data will be aligned to the layer metadata.
## Hovertips
In the case of having a layer that natively supports attributes, and we are using a json sourced table, and the table is not aligned to the service, we would need to disable hovertips, or force them to always use the one-shot quickloader (essentially ignoring the attribute store even though it exists).
|
priority
|
design options for json file data table original request can be found in things have changed a bit since that request and we are now looking at supporting layers that traditionally had no attribute support at all i e adding a table feature instead of just making existing table features faster basically the concept is for a layer we can target a json file to use as a source to populate the data table for that layer there are a number of considerations for how fancy we make this level of integration unlinked in this case the data table is essentially independent of the layer with the exception that clicking the layer in the legend will open the table any additional functions e g zoom to feature details icons will need to be addressed for the unlinked case details in sections below this case is a bear minimum to implement linked in this case the data table and all other functions that run off attributes remain connected to the map layer the only difference is when the app goes to download the attribute set it will source it from the json file instead of querying the map service objectids in the json file must align to the ids on the map service why use a json file when the service already exists some benefits include faster load time and possibly supporting dynamic layers on servers that have the query endpoint disabled the linked case will only work on layers that already support esri attributes i e anything that inherits from attribrecord mainly feature layers and dynamic layers supporting the linked mode is more work but lets us hit additional use cases if we decide not to implement in the first cut of file based tables we should at least attempt to design the new code to allow support in the future file schema esri like dictionary this approach stores attributes in the json file using a similar identical structure to how arcgis server provides attributes the pro of this approach is our current parsing routines can eat it raw we can also create content from existing esri sources with relative ease the con is the file size will be larger as attribute keys are repeated for every entry text fields name type esrifieldtypestring alias field one length name type esrifieldtypedouble alias field two features attributes big duck pond attributes germy lake array based structure this approach stores data in an array form making it look similar to a table the pro of this approach is the file size is minimized the con is we need to write parsing code and possibly it is more effort to translate date into this structure when generating the files text fields name type esrifieldtypestring alias field one length name type esrifieldtypedouble alias field two features as it stands this is the recommended approach for speed gains split structure this approach separates the bulk data and the metadata into two files this lets us quickly download table metadata e g field names up front if we need them whether this is necessary depends on how much support we give to non attribute layer types e g tabe on wms table on esri tile if we essentially disable everything but table view then it may be unnecessary config structure core elements need an optional property on a layernode element to allow a url to a json file to be defined the presence of a value in that property will trigger the appropriate setup for a file based grid see sections below optional elements potential for additional properties to define enhanced things marking a lat long field in the data to enable a zoom to point function an indicator if the file is linked to the layer or is independent e g object ids in file align to object ids on service should default to unlinked for safety ability to provide an icon for each row this could be enhanced to fancier structures essentially providing a class break or unique value renderer type thing not recommended at this point ability to specify a name field would be needed if we want to support a details pane view with a reliable title value see details pane section below for more analysis if we are considering adding options beyond the url we may want to create a sub object to contain all the json table properties e g field names not final text id mylayer url jsontable url linked false icon loading mechanism considerations files must be loaded in the angular code using http get the request library in geoapi fails with files due to the proxy being cooked into esrirequest see the results can be provided from the same function within geoapi as normal attributes see attribute js newlayerpackage the loading from the server should not begin until layerpackage getattribs is first called this approach allows for existing workflows regarding the table attributes to be re used with minimal impact the results can be provided via a separate mechanism would only support unlinked mode the load the table step would recognize its the json file case and call the alternate method to get data allows us to ignore compatibility issues of trying to have traditional attributes and json file attributes existing harmoniously in the same workflows implementations if attempting to integrate with the standard getattribs approach we want the actual web request in angular so that function will exist in the viewer however we trigger our attribute load from the geoapi suggest our new geoapi function to setup the file table mode should accept a loader function this way when the geoapi is ready to load the table usually first time opening the table it can just execute the registered loader function makes geoapi blissfully unaware that it s using angular allows the option to switch to a different load approach later as long as the function interface remains the same likely need an alternate form of newlayerpackage or need to overwrite it s getattribs function the new function needs to see if attribdata exists and if not it needs to make a new promise inside new promise execute the registered function to trigger the http get then massage result from json file format into esri format resolve promise with proper object see createattribset likely need a third function similar to loadfileattribsbuilder and loadserverattribsbuilder probably closer to loadfileattribsbuilder in content it will set the various metadata properties correctly need to be aware of sources particularly if we are dealing with a layer that doesn t natively have attributes e g fields would have to come from file so they may not be set upfront if taking the separate mechanism route we can pretty much do whatever we please layer objects this new feature has the potential to complicate our layer hierarchy in geoapi attribute support is currently confined to the attribrecord class and all inheritors file based tables means classes not derived from attribrecord now require some level of support minimal approach integrated this approach essentially fakes the attribute loader on layers not related to attribrecord it would not use the class inheritance and would instead inject appropriate functions onto record fc objects to make things look compatible we can also entertain doing this on attribrecord based classes where we overwrite existing attribute loader logic with file loader logic on a case by case basis this is a bit risky in that there are quite a few related functions that depend on attributes and may break act weird minimal approach separate mechanism this ties into the separate mechanism approach described in the loading mechanism section above basically the json file to grid flow is separate from all existing structure this approach will only support the unlinked mode so it causes the least amount of chaos but offers the least amount of value reeefactor this involves significant changes to the layer record fc object hierarchies while it would need further analysis if we decide to persue this at a high level attribute support would need to exist at a low level in the hierarchy the layerrecord class would make sense the layerrecord implementation could have the file based load support if no file has been defined in the config it can return an empty array or an error to be decided which is more appropriate the attribrecord class would retain it s code for supporting arcgis server based attribute loading however it would need to allow for recognizing the file based case and let the logic drop through to layerrecord in other words it can t just steamroll the attribute logic it needs to allow both branches of code to execute depending on the configuration this approach offers good support for all layer types and best supports the linked mode having file that aligns with server data details pane in unlinked mode we have a few choices disable the details pane the data is in the grid anyways so its just showing the same thing but in vertical layout lazy should remove the detail button from the grid push data to the standard zebra stripe attribute view may need to disable things like zoom to feature and icon depending on what configuration options we ve added may also have to guess or suppress the name field if we have decided to not make that a config option or it was not provided could make the detail pane a configuration option in linked mode everything should work as normal as the data will be aligned to the layer metadata zoom to feature in unlinked mode default behavior would be no zoom button should not be in the grid can have optional lat long columns defined in config to zoom to point would involve a projection and then the zoom to point pipeline if geometry is a line poly only real option is to use the lat long as centroid and accept that zoom level may be a poor choice for the shape in linked mode everything should work as normal as the data will be aligned to the layer metadata hovertips in the case of having a layer that natively supports attributes and we are using a json sourced table and the table is not aligned to the service we would need to disable hovertips or force them to always use the one shot quickloader essentially ignoring the attribute store even though it exists
| 1
|
501,443
| 14,528,424,543
|
IssuesEvent
|
2020-12-14 16:31:09
|
E3SM-Project/scorpio
|
https://api.github.com/repos/E3SM-Project/scorpio
|
opened
|
Scorpio asserts for non-NULL starts/counts when writing scalar variables
|
High Priority Next Release
|
Scorpio asserts when the starts/counts are valid when writing out scalar variables. Although passing starts/counts don't make sense when writing out scalar variables libraries like NetCDF allow (and silently ignore) it.
From pio_getput_int.c
```
...
if (ndims == 0)
{
/* This is a scalar var. */
LOG((2, "pnetcdf writing scalar with ncmpi_put_vars_*() file->fh = %d varid = %d",
file->fh, varid));
pioassert(!start && !count && !stride, "expected NULLs", __FILE__, __LINE__);
...
```
We will need to remove this assert and include a tighter check and warning to the user.
Also see E3SM-Project/E3SM#3988
|
1.0
|
Scorpio asserts for non-NULL starts/counts when writing scalar variables - Scorpio asserts when the starts/counts are valid when writing out scalar variables. Although passing starts/counts don't make sense when writing out scalar variables libraries like NetCDF allow (and silently ignore) it.
From pio_getput_int.c
```
...
if (ndims == 0)
{
/* This is a scalar var. */
LOG((2, "pnetcdf writing scalar with ncmpi_put_vars_*() file->fh = %d varid = %d",
file->fh, varid));
pioassert(!start && !count && !stride, "expected NULLs", __FILE__, __LINE__);
...
```
We will need to remove this assert and include a tighter check and warning to the user.
Also see E3SM-Project/E3SM#3988
|
priority
|
scorpio asserts for non null starts counts when writing scalar variables scorpio asserts when the starts counts are valid when writing out scalar variables although passing starts counts don t make sense when writing out scalar variables libraries like netcdf allow and silently ignore it from pio getput int c if ndims this is a scalar var log pnetcdf writing scalar with ncmpi put vars file fh d varid d file fh varid pioassert start count stride expected nulls file line we will need to remove this assert and include a tighter check and warning to the user also see project
| 1
|
381,825
| 11,296,020,360
|
IssuesEvent
|
2020-01-17 00:14:02
|
lokka30/PhantomCombat
|
https://api.github.com/repos/lokka30/PhantomCombat
|
closed
|
plural feature for combat mode boss bar
|
enhancement priority: high
|
at the moment it says '1 seconds' but there should be a placeholder so it says '1 second'
e.g. '%time% second%s%'
|
1.0
|
plural feature for combat mode boss bar - at the moment it says '1 seconds' but there should be a placeholder so it says '1 second'
e.g. '%time% second%s%'
|
priority
|
plural feature for combat mode boss bar at the moment it says seconds but there should be a placeholder so it says second e g time second s
| 1
|
372,448
| 11,014,791,415
|
IssuesEvent
|
2019-12-04 23:41:11
|
data2health/maturity-model
|
https://api.github.com/repos/data2health/maturity-model
|
closed
|
Develop app login page
|
Priority: High enhancement
|
Should
- [x] Authenticate with Python Flask API
- [x] Check user email and survey entry code
|
1.0
|
Develop app login page - Should
- [x] Authenticate with Python Flask API
- [x] Check user email and survey entry code
|
priority
|
develop app login page should authenticate with python flask api check user email and survey entry code
| 1
|
193,313
| 6,883,852,647
|
IssuesEvent
|
2017-11-21 10:50:17
|
wordpress-mobile/AztecEditor-Android
|
https://api.github.com/repos/wordpress-mobile/AztecEditor-Android
|
closed
|
List items broken for RTL languages
|
bug high priority
|
The list items (both ordered and unordered) are not rendered properly for RTL languages. The bullets and numbers are not displayed.

|
1.0
|
List items broken for RTL languages - The list items (both ordered and unordered) are not rendered properly for RTL languages. The bullets and numbers are not displayed.

|
priority
|
list items broken for rtl languages the list items both ordered and unordered are not rendered properly for rtl languages the bullets and numbers are not displayed
| 1
|
474,653
| 13,673,223,327
|
IssuesEvent
|
2020-09-29 09:33:00
|
swan-cern/jupyter-extensions
|
https://api.github.com/repos/swan-cern/jupyter-extensions
|
closed
|
Renew of oAuth tokens inside the user session
|
priority:high type:feature request
|
We need to support oAuth token for accessing EOS and, in the future, to access the CS3APIs #26.
The token can, and will, be pushed while Spawning, but we also need to ensure that is renewed inside the user session. We should be able able to retrieve that information directly from the authenticator renewal process.
The best way to retrieve this information is to try to get the user auth_state directly from the JH API. If this information is not available, we create a Authenticated Service that will return this information.
|
1.0
|
Renew of oAuth tokens inside the user session - We need to support oAuth token for accessing EOS and, in the future, to access the CS3APIs #26.
The token can, and will, be pushed while Spawning, but we also need to ensure that is renewed inside the user session. We should be able able to retrieve that information directly from the authenticator renewal process.
The best way to retrieve this information is to try to get the user auth_state directly from the JH API. If this information is not available, we create a Authenticated Service that will return this information.
|
priority
|
renew of oauth tokens inside the user session we need to support oauth token for accessing eos and in the future to access the the token can and will be pushed while spawning but we also need to ensure that is renewed inside the user session we should be able able to retrieve that information directly from the authenticator renewal process the best way to retrieve this information is to try to get the user auth state directly from the jh api if this information is not available we create a authenticated service that will return this information
| 1
|
495,108
| 14,272,193,420
|
IssuesEvent
|
2020-11-21 15:58:36
|
UC-Davis-molecular-computing/scadnano
|
https://api.github.com/repos/UC-Davis-molecular-computing/scadnano
|
opened
|
Fix deprecation warning on Release action
|
bug super high priority
|
In [this run](https://github.com/UC-Davis-molecular-computing/scadnano/runs/1435916555?check_suite_focus=true), the action "failed" because of the use of deprecated `set-env` variable:
```
The `set-env` command is disabled. Please upgrade to using Environment Files or opt into unsecure command execution by setting the `ACTIONS_ALLOW_UNSECURE_COMMANDS` environment variable to `true`. For more information see: https://github.blog/changelog/2020-10-01-github-actions-deprecating-set-env-and-add-path-commands/
```
The action was still able to successfully generate the release log, but the workflow is considered a failure.
|
1.0
|
Fix deprecation warning on Release action - In [this run](https://github.com/UC-Davis-molecular-computing/scadnano/runs/1435916555?check_suite_focus=true), the action "failed" because of the use of deprecated `set-env` variable:
```
The `set-env` command is disabled. Please upgrade to using Environment Files or opt into unsecure command execution by setting the `ACTIONS_ALLOW_UNSECURE_COMMANDS` environment variable to `true`. For more information see: https://github.blog/changelog/2020-10-01-github-actions-deprecating-set-env-and-add-path-commands/
```
The action was still able to successfully generate the release log, but the workflow is considered a failure.
|
priority
|
fix deprecation warning on release action in the action failed because of the use of deprecated set env variable the set env command is disabled please upgrade to using environment files or opt into unsecure command execution by setting the actions allow unsecure commands environment variable to true for more information see the action was still able to successfully generate the release log but the workflow is considered a failure
| 1
|
499,273
| 14,443,858,154
|
IssuesEvent
|
2020-12-07 20:18:43
|
TheIOFoundation/ProjectLockdown
|
https://api.github.com/repos/TheIOFoundation/ProjectLockdown
|
opened
|
[ADM] Position: Crowdfunding assistant
|
Level: Member Module: ADM Need: Position Priority: High Project: PLD Stage: Ready Team: RA
|

# Open Position
## Position title
Crowdfunding Assistant
## Position type
Volunteer
## Position description
The Crowdfunding assistant will be a member of Team Resource Allocation and will be responsible for the following:
- Creating a comprehensive crowdfunding strategic plan
- Setting up and maintaining accounts
- Working with Team Communications to write the listing content and produce other necessary materials
(such as brochures or pitch videos)
- Working with Team UXUI to create graphics and other necessary materials
- Developing and producing incentives
- Updating the crowdfunding platform with latest developments
- Building email lists
- Report to the Team Resource Allocation Coordinator
## Skills & Qualifications
- Requirements:
-- Excellent communication skills
-- Creative thinker
-- Familiarity with platforms for crowdfunding
(such as Patreon, GitHub Sponsors, Open Collective GoFundMe, Indiegogo and others)
-- Able to participate in a highly dynamic environment as a part of a team
-- Able to learn independently and dedicate at least 6 hours a week and attend regular team meetings
-- Able to participate in a highly dynamic environment as a part of a team
- Desirable
-- Previous experience with Crowdfunding
- Optional
-- Previous experience with an NGO
## Disclaimer
This is a completely volunteer, non remunerated position. Please ensure that there are no legal impediments that may enter in conflict with the position and tasks assigned.
If interested, please fill the following form: https://TIOF.Click/PLDEoI
|
1.0
|
[ADM] Position: Crowdfunding assistant - 
# Open Position
## Position title
Crowdfunding Assistant
## Position type
Volunteer
## Position description
The Crowdfunding assistant will be a member of Team Resource Allocation and will be responsible for the following:
- Creating a comprehensive crowdfunding strategic plan
- Setting up and maintaining accounts
- Working with Team Communications to write the listing content and produce other necessary materials
(such as brochures or pitch videos)
- Working with Team UXUI to create graphics and other necessary materials
- Developing and producing incentives
- Updating the crowdfunding platform with latest developments
- Building email lists
- Report to the Team Resource Allocation Coordinator
## Skills & Qualifications
- Requirements:
-- Excellent communication skills
-- Creative thinker
-- Familiarity with platforms for crowdfunding
(such as Patreon, GitHub Sponsors, Open Collective GoFundMe, Indiegogo and others)
-- Able to participate in a highly dynamic environment as a part of a team
-- Able to learn independently and dedicate at least 6 hours a week and attend regular team meetings
-- Able to participate in a highly dynamic environment as a part of a team
- Desirable
-- Previous experience with Crowdfunding
- Optional
-- Previous experience with an NGO
## Disclaimer
This is a completely volunteer, non remunerated position. Please ensure that there are no legal impediments that may enter in conflict with the position and tasks assigned.
If interested, please fill the following form: https://TIOF.Click/PLDEoI
|
priority
|
position crowdfunding assistant open position position title crowdfunding assistant position type volunteer position description the crowdfunding assistant will be a member of team resource allocation and will be responsible for the following creating a comprehensive crowdfunding strategic plan setting up and maintaining accounts working with team communications to write the listing content and produce other necessary materials such as brochures or pitch videos working with team uxui to create graphics and other necessary materials developing and producing incentives updating the crowdfunding platform with latest developments building email lists report to the team resource allocation coordinator skills qualifications requirements excellent communication skills creative thinker familiarity with platforms for crowdfunding such as patreon github sponsors open collective gofundme indiegogo and others able to participate in a highly dynamic environment as a part of a team able to learn independently and dedicate at least hours a week and attend regular team meetings able to participate in a highly dynamic environment as a part of a team desirable previous experience with crowdfunding optional previous experience with an ngo disclaimer this is a completely volunteer non remunerated position please ensure that there are no legal impediments that may enter in conflict with the position and tasks assigned if interested please fill the following form
| 1
|
586,100
| 17,569,660,273
|
IssuesEvent
|
2021-08-14 12:14:18
|
umple/umple
|
https://api.github.com/repos/umple/umple
|
closed
|
No option to cancel the Task Create Panel after a task is created in UmpleOnline
|
Component-UmpleOnline Priority-High Diffic-Easy tasks
|
## Summary
Within the UmpleOnline, the users can create a task and save it. Once the task a saved, the system has no option to cancel this panel. Moreover, it does not show the "Load a task" option under the "TASKS" section. Therefore, the user is forced to reload the home page in browser to create a new task.
## Steps to Reproduce
1. Go to the "TASKS", then click on the "Create a Task".

2. Enter a sample value at task name and instructions. Click "Submit Task".

3. Now, there is no option to hide this task creation panel. This panel is visible at every action that the user does.

## Expected Feature
The system should have a option to close this panel after a task is saved. A similar workflow is implemented when clicked on "Cancel this task response".
|
1.0
|
No option to cancel the Task Create Panel after a task is created in UmpleOnline - ## Summary
Within the UmpleOnline, the users can create a task and save it. Once the task a saved, the system has no option to cancel this panel. Moreover, it does not show the "Load a task" option under the "TASKS" section. Therefore, the user is forced to reload the home page in browser to create a new task.
## Steps to Reproduce
1. Go to the "TASKS", then click on the "Create a Task".

2. Enter a sample value at task name and instructions. Click "Submit Task".

3. Now, there is no option to hide this task creation panel. This panel is visible at every action that the user does.

## Expected Feature
The system should have a option to close this panel after a task is saved. A similar workflow is implemented when clicked on "Cancel this task response".
|
priority
|
no option to cancel the task create panel after a task is created in umpleonline summary within the umpleonline the users can create a task and save it once the task a saved the system has no option to cancel this panel moreover it does not show the load a task option under the tasks section therefore the user is forced to reload the home page in browser to create a new task steps to reproduce go to the tasks then click on the create a task enter a sample value at task name and instructions click submit task now there is no option to hide this task creation panel this panel is visible at every action that the user does expected feature the system should have a option to close this panel after a task is saved a similar workflow is implemented when clicked on cancel this task response
| 1
|
130,678
| 5,119,295,075
|
IssuesEvent
|
2017-01-08 16:31:55
|
bmoreonrails/rails_tutorial
|
https://api.github.com/repos/bmoreonrails/rails_tutorial
|
closed
|
Make the syntax colors less awful
|
help wanted high-priority
|
- [ ] get rid of ugly default colors like the red in the html blocks
- [ ] We also need the contrast to work for anyone who might be colorblind. Here's just one tool for checking it - you can take a screen shot and upload it to here: http://www.color-blindness.com/coblis-color-blindness-simulator/
There is a syntax.scss file that I copied from GH in there that you can start with. The markup might be a little inaccurate in spots with regard to the language in the block - so be sure to check which language is indicated and change that if you need to.
|
1.0
|
Make the syntax colors less awful - - [ ] get rid of ugly default colors like the red in the html blocks
- [ ] We also need the contrast to work for anyone who might be colorblind. Here's just one tool for checking it - you can take a screen shot and upload it to here: http://www.color-blindness.com/coblis-color-blindness-simulator/
There is a syntax.scss file that I copied from GH in there that you can start with. The markup might be a little inaccurate in spots with regard to the language in the block - so be sure to check which language is indicated and change that if you need to.
|
priority
|
make the syntax colors less awful get rid of ugly default colors like the red in the html blocks we also need the contrast to work for anyone who might be colorblind here s just one tool for checking it you can take a screen shot and upload it to here there is a syntax scss file that i copied from gh in there that you can start with the markup might be a little inaccurate in spots with regard to the language in the block so be sure to check which language is indicated and change that if you need to
| 1
|
471,267
| 13,563,633,771
|
IssuesEvent
|
2020-09-18 08:50:42
|
decentraland/explorer
|
https://api.github.com/repos/decentraland/explorer
|
closed
|
transparencyMode issues in materials
|
SDK bug high priority
|
Here we see a planeshape that appears to have z-fighting issues with what's behind it, even though there's a significant distance in between

[https://app.zenhub.com/files/220994476/b53ddd64-9638-41d5-ba3a-979778dc26fc/download](https://app.zenhub.com/files/220994476/b53ddd64-9638-41d5-ba3a-979778dc26fc/download)
Here the material is using a same image as the alphaTexture and the albedoTexture and transparencyMode is not defined
spriteMaterial.albedoTexture = spriteTexture;
spriteMaterial.alphaTexture = spriteTexture;
If I add an extra line to this scene with a transparencyMode = 1, it fixes it
spriteMaterial.transparencyMode = 1
But if I set that to 2 or 3, if fails again. It fails if undefined because 2 is the default value of transparencyMode.
* | Value | Type |
* | ----- | ---------------------------------------------- |
* | 0 | OPAQUE (default) |
* | 1 | ALPHATEST |
* | 2 | ALPHABLEND |
* | 3 | ALPHATESTANDBLEND |
* | 4 | AUTO (ALPHABLEND if alpha OPAQUE otherwise |
*/
So ALPHATEST works, ALPHABLEND and ALPHATESTANDBLEND fail.
Here's a second example where an alphaTexture is not working.
https://github.com/MetaZoneio/AlphaDemo
In this case, we're applying a different texture in alphaTexture from the one in albedoTexture. We should be seeing that a couple of these plains have a round shape being cut out, but the alphaTexture is not affecting what we see in any of the cases.

We can edit the transparencyMode property for all of them, and set any value we want there, but nothing changes with any value we set.
|
1.0
|
transparencyMode issues in materials -
Here we see a planeshape that appears to have z-fighting issues with what's behind it, even though there's a significant distance in between

[https://app.zenhub.com/files/220994476/b53ddd64-9638-41d5-ba3a-979778dc26fc/download](https://app.zenhub.com/files/220994476/b53ddd64-9638-41d5-ba3a-979778dc26fc/download)
Here the material is using a same image as the alphaTexture and the albedoTexture and transparencyMode is not defined
spriteMaterial.albedoTexture = spriteTexture;
spriteMaterial.alphaTexture = spriteTexture;
If I add an extra line to this scene with a transparencyMode = 1, it fixes it
spriteMaterial.transparencyMode = 1
But if I set that to 2 or 3, if fails again. It fails if undefined because 2 is the default value of transparencyMode.
* | Value | Type |
* | ----- | ---------------------------------------------- |
* | 0 | OPAQUE (default) |
* | 1 | ALPHATEST |
* | 2 | ALPHABLEND |
* | 3 | ALPHATESTANDBLEND |
* | 4 | AUTO (ALPHABLEND if alpha OPAQUE otherwise |
*/
So ALPHATEST works, ALPHABLEND and ALPHATESTANDBLEND fail.
Here's a second example where an alphaTexture is not working.
https://github.com/MetaZoneio/AlphaDemo
In this case, we're applying a different texture in alphaTexture from the one in albedoTexture. We should be seeing that a couple of these plains have a round shape being cut out, but the alphaTexture is not affecting what we see in any of the cases.

We can edit the transparencyMode property for all of them, and set any value we want there, but nothing changes with any value we set.
|
priority
|
transparencymode issues in materials here we see a planeshape that appears to have z fighting issues with what s behind it even though there s a significant distance in between here the material is using a same image as the alphatexture and the albedotexture and transparencymode is not defined spritematerial albedotexture spritetexture spritematerial alphatexture spritetexture if i add an extra line to this scene with a transparencymode it fixes it spritematerial transparencymode but if i set that to or if fails again it fails if undefined because is the default value of transparencymode value type opaque default alphatest alphablend alphatestandblend auto alphablend if alpha opaque otherwise so alphatest works alphablend and alphatestandblend fail here s a second example where an alphatexture is not working in this case we re applying a different texture in alphatexture from the one in albedotexture we should be seeing that a couple of these plains have a round shape being cut out but the alphatexture is not affecting what we see in any of the cases we can edit the transparencymode property for all of them and set any value we want there but nothing changes with any value we set
| 1
|
434,504
| 12,519,339,817
|
IssuesEvent
|
2020-06-03 14:18:17
|
luna/ide
|
https://api.github.com/repos/luna/ide
|
closed
|
Nodes and edges view
|
Category: IDE Change: Non-Breaking Difficulty: Core Contributor Epic Priority: Highest Type: Enhancement
|
- Collapsed and expanded (no widgets yet)
- With labels above edges
- With text above nodes
- Ability to change the expression
- flags
|
1.0
|
Nodes and edges view - - Collapsed and expanded (no widgets yet)
- With labels above edges
- With text above nodes
- Ability to change the expression
- flags
|
priority
|
nodes and edges view collapsed and expanded no widgets yet with labels above edges with text above nodes ability to change the expression flags
| 1
|
401,906
| 11,799,849,195
|
IssuesEvent
|
2020-03-18 16:32:40
|
geneontology/minerva
|
https://api.github.com/repos/geneontology/minerva
|
opened
|
API - add search models by ontology terms other than GO
|
high priority
|
Task requirement from Noctua Landing Page Project
- [ ] include ontologies such as Uberon, CL in blazegraph
- [ ] make sure one can perform a closure search (eg retrieve models also containing the more specific/children terms of the specified term)
- [ ] Provide the API route for NLP UI
|
1.0
|
API - add search models by ontology terms other than GO - Task requirement from Noctua Landing Page Project
- [ ] include ontologies such as Uberon, CL in blazegraph
- [ ] make sure one can perform a closure search (eg retrieve models also containing the more specific/children terms of the specified term)
- [ ] Provide the API route for NLP UI
|
priority
|
api add search models by ontology terms other than go task requirement from noctua landing page project include ontologies such as uberon cl in blazegraph make sure one can perform a closure search eg retrieve models also containing the more specific children terms of the specified term provide the api route for nlp ui
| 1
|
614,397
| 19,181,887,324
|
IssuesEvent
|
2021-12-04 14:50:25
|
BlueBubblesApp/bluebubbles-app
|
https://api.github.com/repos/BlueBubblesApp/bluebubbles-app
|
opened
|
Fix issue where image disappears after sending. Comes back after leave and re enter
|
Bug priority: high Alpha Difficulty: Medium
|
Not sure if this will work for you, but here is what I did:
1. Send an image with text
2. Before the image fully sends, leave the app
3. Wait a sec
4. Re enter the app
5. Image seems to disappear or flicker
6. Send a message and it comes back
|
1.0
|
Fix issue where image disappears after sending. Comes back after leave and re enter - Not sure if this will work for you, but here is what I did:
1. Send an image with text
2. Before the image fully sends, leave the app
3. Wait a sec
4. Re enter the app
5. Image seems to disappear or flicker
6. Send a message and it comes back
|
priority
|
fix issue where image disappears after sending comes back after leave and re enter not sure if this will work for you but here is what i did send an image with text before the image fully sends leave the app wait a sec re enter the app image seems to disappear or flicker send a message and it comes back
| 1
|
307,961
| 9,424,405,039
|
IssuesEvent
|
2019-04-11 14:02:52
|
level73/membernet
|
https://api.github.com/repos/level73/membernet
|
opened
|
Separate EMENA from Global results in membership fees
|
Priority: High Type: Development
|
Currently there is no option for viewing EMENA membership fees only (i presume they are grouped under global?). We'd like to separate them.
|
1.0
|
Separate EMENA from Global results in membership fees - Currently there is no option for viewing EMENA membership fees only (i presume they are grouped under global?). We'd like to separate them.
|
priority
|
separate emena from global results in membership fees currently there is no option for viewing emena membership fees only i presume they are grouped under global we d like to separate them
| 1
|
239,790
| 7,800,024,243
|
IssuesEvent
|
2018-06-09 03:37:37
|
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
|
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
|
closed
|
0006802:
Error on setting a Flags message filter.
|
Bug Felamimail Mantis high priority
|
**Reported by esakol on 23 Jul 2012 21:05**
**Version:** Milan (2012.03.5)
When selecting a Flags filter in Felamimail's messages filter interface, the value ComboBox is empty. And when clicking at the value combo box, the following error is reported on firebug's console.
this.selectionWidget is null
this.selectionWidget.on('select', this.onRecordSelect, this);
on Tinebase/js/widgets/grid/PickerFilter.js (line 303)
|
1.0
|
0006802:
Error on setting a Flags message filter. - **Reported by esakol on 23 Jul 2012 21:05**
**Version:** Milan (2012.03.5)
When selecting a Flags filter in Felamimail's messages filter interface, the value ComboBox is empty. And when clicking at the value combo box, the following error is reported on firebug's console.
this.selectionWidget is null
this.selectionWidget.on('select', this.onRecordSelect, this);
on Tinebase/js/widgets/grid/PickerFilter.js (line 303)
|
priority
|
error on setting a flags message filter reported by esakol on jul version milan when selecting a flags filter in felamimail s messages filter interface the value combobox is empty and when clicking at the value combo box the following error is reported on firebug s console this selectionwidget is null this selectionwidget on select this onrecordselect this on tinebase js widgets grid pickerfilter js line
| 1
|
340,562
| 10,273,310,344
|
IssuesEvent
|
2019-08-23 18:54:56
|
X-Plane/XPlane2Blender
|
https://api.github.com/repos/X-Plane/XPlane2Blender
|
closed
|
249: Convert Aircraft and Cockpit properties
|
2.49 Converter priority high
|
The first alpha of the converter should be Airplane only (with anything else a plus)
What's left to implement:
### ATTR
- [x] ATTR_draw_disable
- [x] ATTR_solid_camera (game prop only, not TexFace version)
- [x] ATTR_light_level (and _v1, and _v2 and lit_level)
### GLOBAL
- [x] NORMAL_METALNESS
- [x] BLEND_GLASS
~- GLOBAL_cockpit_lit (Isn't this simply always on now?)~ This was not a global prop, is true by default, and cockpit_lit only comes into play if the export version is less than `1100`
- [x] GLOBAL_no_blend
- [x] GLOBAL_specular
- [x] GLOBAL_shadow_blend
- [x] GLOBAL_tint
Plus, of course, unit tests
- [x] ATTR properties only
- [x] GLOBAL_properties only
These get applied before the mesh splits, since it previous applied to the whole Object anyway. It is just easier that way.
|
1.0
|
249: Convert Aircraft and Cockpit properties - The first alpha of the converter should be Airplane only (with anything else a plus)
What's left to implement:
### ATTR
- [x] ATTR_draw_disable
- [x] ATTR_solid_camera (game prop only, not TexFace version)
- [x] ATTR_light_level (and _v1, and _v2 and lit_level)
### GLOBAL
- [x] NORMAL_METALNESS
- [x] BLEND_GLASS
~- GLOBAL_cockpit_lit (Isn't this simply always on now?)~ This was not a global prop, is true by default, and cockpit_lit only comes into play if the export version is less than `1100`
- [x] GLOBAL_no_blend
- [x] GLOBAL_specular
- [x] GLOBAL_shadow_blend
- [x] GLOBAL_tint
Plus, of course, unit tests
- [x] ATTR properties only
- [x] GLOBAL_properties only
These get applied before the mesh splits, since it previous applied to the whole Object anyway. It is just easier that way.
|
priority
|
convert aircraft and cockpit properties the first alpha of the converter should be airplane only with anything else a plus what s left to implement attr attr draw disable attr solid camera game prop only not texface version attr light level and and and lit level global normal metalness blend glass global cockpit lit isn t this simply always on now this was not a global prop is true by default and cockpit lit only comes into play if the export version is less than global no blend global specular global shadow blend global tint plus of course unit tests attr properties only global properties only these get applied before the mesh splits since it previous applied to the whole object anyway it is just easier that way
| 1
|
461,111
| 13,223,867,159
|
IssuesEvent
|
2020-08-17 18:02:08
|
woocommerce/woocommerce
|
https://api.github.com/repos/woocommerce/woocommerce
|
closed
|
Order details page not displaying variation/attribute name
|
bug component: variation has pull request priority: high
|
The other bug report for this was closed incorrectly and is being ignored, but the bug is very much present.
On numerous of my client's sites, the order details do not say the variation's name in the title. I've tested this on WC 3.2.6 and 3.8, they both have the same issue. Occasionally it'll display the variation's name, and occasionally not. I can't work out why this happens sporadically.
**To Reproduce**
Check your order details. If it's happening to several of our sites it'd presumably appear on yours too.
The below screenshots are 2 separate stores. You can clearly see there's a variation ID, but no variation name in the title. I'm getting angry emails from my clients and I haven't a clue what to suggest:
[https://i.imgur.com/oqfZ13r.png](https://i.imgur.com/oqfZ13r.png)
[https://i.imgur.com/5OlvvgQ.png](https://i.imgur.com/5OlvvgQ.png)
**Expected behavior**
...that it'd say what the variation is in the title.
**Isolating the problem (mark completed items with an [x]):**
- [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active.
- [x] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/).
- [x] I can reproduce this bug consistently using the steps above.
I'm leaving out the status report as this is affecting numerous WC instances on different versions/themes/etc. It's everywhere.
|
1.0
|
Order details page not displaying variation/attribute name - The other bug report for this was closed incorrectly and is being ignored, but the bug is very much present.
On numerous of my client's sites, the order details do not say the variation's name in the title. I've tested this on WC 3.2.6 and 3.8, they both have the same issue. Occasionally it'll display the variation's name, and occasionally not. I can't work out why this happens sporadically.
**To Reproduce**
Check your order details. If it's happening to several of our sites it'd presumably appear on yours too.
The below screenshots are 2 separate stores. You can clearly see there's a variation ID, but no variation name in the title. I'm getting angry emails from my clients and I haven't a clue what to suggest:
[https://i.imgur.com/oqfZ13r.png](https://i.imgur.com/oqfZ13r.png)
[https://i.imgur.com/5OlvvgQ.png](https://i.imgur.com/5OlvvgQ.png)
**Expected behavior**
...that it'd say what the variation is in the title.
**Isolating the problem (mark completed items with an [x]):**
- [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active.
- [x] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/).
- [x] I can reproduce this bug consistently using the steps above.
I'm leaving out the status report as this is affecting numerous WC instances on different versions/themes/etc. It's everywhere.
|
priority
|
order details page not displaying variation attribute name the other bug report for this was closed incorrectly and is being ignored but the bug is very much present on numerous of my client s sites the order details do not say the variation s name in the title i ve tested this on wc and they both have the same issue occasionally it ll display the variation s name and occasionally not i can t work out why this happens sporadically to reproduce check your order details if it s happening to several of our sites it d presumably appear on yours too the below screenshots are separate stores you can clearly see there s a variation id but no variation name in the title i m getting angry emails from my clients and i haven t a clue what to suggest expected behavior that it d say what the variation is in the title isolating the problem mark completed items with an i have deactivated other plugins and confirmed this bug occurs when only woocommerce plugin is active this bug happens with a default wordpress theme active or i can reproduce this bug consistently using the steps above i m leaving out the status report as this is affecting numerous wc instances on different versions themes etc it s everywhere
| 1
|
770,837
| 27,058,552,301
|
IssuesEvent
|
2023-02-13 17:53:17
|
union-platform/union-mobile-app
|
https://api.github.com/repos/union-platform/union-mobile-app
|
opened
|
User wants to apply to the team, because he believes that it is suitable for him
|
priority: high type: feature
|
**Scope of action:** Team Profile Screen
**Precondition:**
**Design:** https://www.figma.com/file/2St3zSul4fHnLffqy3WK7P/union-mobile?node-id=1481%3A777
## Use cases:
1. The user clicks on the "submit an application" button
2. The system sends a request to the team
-----
**Область действия:** Экран профиля команды
**Предусловие:**
**Дизайн:** https://www.figma.com/file/2St3zSul4fHnLffqy3WK7P/union-mobile?node-id=1481%3A777
## Сценарий:
1. Пользователь нажимает на кнопку "подать заявку"
2. Система отправляет заявку в команду
|
1.0
|
User wants to apply to the team, because he believes that it is suitable for him - **Scope of action:** Team Profile Screen
**Precondition:**
**Design:** https://www.figma.com/file/2St3zSul4fHnLffqy3WK7P/union-mobile?node-id=1481%3A777
## Use cases:
1. The user clicks on the "submit an application" button
2. The system sends a request to the team
-----
**Область действия:** Экран профиля команды
**Предусловие:**
**Дизайн:** https://www.figma.com/file/2St3zSul4fHnLffqy3WK7P/union-mobile?node-id=1481%3A777
## Сценарий:
1. Пользователь нажимает на кнопку "подать заявку"
2. Система отправляет заявку в команду
|
priority
|
user wants to apply to the team because he believes that it is suitable for him scope of action team profile screen precondition design use cases the user clicks on the submit an application button the system sends a request to the team область действия экран профиля команды предусловие дизайн сценарий пользователь нажимает на кнопку подать заявку система отправляет заявку в команду
| 1
|
594,395
| 18,044,731,285
|
IssuesEvent
|
2021-09-18 17:44:26
|
AY2122S1-CS2103T-W16-2/tp
|
https://api.github.com/repos/AY2122S1-CS2103T-W16-2/tp
|
closed
|
Update about us page in project website
|
type.Task priority.High
|
The about us needs to be updated with the correct github link, photos, etc for every team member.
Update Information for the following team members
- [x] Ajay
- [x] Eunice
- [x] Geng Ning
- [x] Imran
- [x] Siva
|
1.0
|
Update about us page in project website - The about us needs to be updated with the correct github link, photos, etc for every team member.
Update Information for the following team members
- [x] Ajay
- [x] Eunice
- [x] Geng Ning
- [x] Imran
- [x] Siva
|
priority
|
update about us page in project website the about us needs to be updated with the correct github link photos etc for every team member update information for the following team members ajay eunice geng ning imran siva
| 1
|
486,355
| 14,007,856,620
|
IssuesEvent
|
2020-10-28 22:22:28
|
fossasia/open-event-frontend
|
https://api.github.com/repos/fossasia/open-event-frontend
|
opened
|
Sign up process tells user "reset your password" and "forgot your password"
|
Priority: High Priority: Urgent bug
|
Something changed at one point in the sign up process. Instead of welcoming users who have just signed up, we tell them "reset your password" and direct them to a page "forgot your password".
Expected:
1. Users who sign up should receive an email: "Hello, thank you for signing up on eventyay.com. Now please verify your email address and set a password for your account. Please click here [link]."

2. On the page that users are directed to then please show the following:
* Thank you for verifying your email! Please create a password for your account.
* Instead of "reset password" show "Set Password".
* Do not show "new to the site? create an account..."

3. Ensure the changes do not impact the way passwords are reset for existing accounts. For existing verified accounts, the password reset process should stay as is.
|
2.0
|
Sign up process tells user "reset your password" and "forgot your password" - Something changed at one point in the sign up process. Instead of welcoming users who have just signed up, we tell them "reset your password" and direct them to a page "forgot your password".
Expected:
1. Users who sign up should receive an email: "Hello, thank you for signing up on eventyay.com. Now please verify your email address and set a password for your account. Please click here [link]."

2. On the page that users are directed to then please show the following:
* Thank you for verifying your email! Please create a password for your account.
* Instead of "reset password" show "Set Password".
* Do not show "new to the site? create an account..."

3. Ensure the changes do not impact the way passwords are reset for existing accounts. For existing verified accounts, the password reset process should stay as is.
|
priority
|
sign up process tells user reset your password and forgot your password something changed at one point in the sign up process instead of welcoming users who have just signed up we tell them reset your password and direct them to a page forgot your password expected users who sign up should receive an email hello thank you for signing up on eventyay com now please verify your email address and set a password for your account please click here on the page that users are directed to then please show the following thank you for verifying your email please create a password for your account instead of reset password show set password do not show new to the site create an account ensure the changes do not impact the way passwords are reset for existing accounts for existing verified accounts the password reset process should stay as is
| 1
|
765,677
| 26,856,985,808
|
IssuesEvent
|
2023-02-03 15:19:25
|
seddonym/grimp
|
https://api.github.com/repos/seddonym/grimp
|
closed
|
Better handling of modules that are not in the graph
|
high-priority
|
For example:
```
File ".../grimp/adaptors/graph.py", line 201, in find_shortest_chain
target=imported))
File ".../networkx/algorithms/shortest_paths/generic.py", line 170, in shortest_path
paths = nx.bidirectional_shortest_path(G, source, target)
File ".../networkx/algorithms/shortest_paths/unweighted.py", line 223, in bidirectional_shortest_path
raise nx.NodeNotFound(msg.format(source, target))
networkx.exception.NodeNotFound: Either source myproject.foo or target myproject.bar is not in G
```
This should raise a specific grimp exception class `ModuleNotInGraph` with a nicer error message.
|
1.0
|
Better handling of modules that are not in the graph - For example:
```
File ".../grimp/adaptors/graph.py", line 201, in find_shortest_chain
target=imported))
File ".../networkx/algorithms/shortest_paths/generic.py", line 170, in shortest_path
paths = nx.bidirectional_shortest_path(G, source, target)
File ".../networkx/algorithms/shortest_paths/unweighted.py", line 223, in bidirectional_shortest_path
raise nx.NodeNotFound(msg.format(source, target))
networkx.exception.NodeNotFound: Either source myproject.foo or target myproject.bar is not in G
```
This should raise a specific grimp exception class `ModuleNotInGraph` with a nicer error message.
|
priority
|
better handling of modules that are not in the graph for example file grimp adaptors graph py line in find shortest chain target imported file networkx algorithms shortest paths generic py line in shortest path paths nx bidirectional shortest path g source target file networkx algorithms shortest paths unweighted py line in bidirectional shortest path raise nx nodenotfound msg format source target networkx exception nodenotfound either source myproject foo or target myproject bar is not in g this should raise a specific grimp exception class modulenotingraph with a nicer error message
| 1
|
247,410
| 7,918,423,435
|
IssuesEvent
|
2018-07-04 13:16:24
|
openaq/openaq-fetch
|
https://api.github.com/repos/openaq/openaq-fetch
|
closed
|
China - Sources (not useable yet)
|
help wanted high priority new data question
|
There are a ton of China air quality data sources. Here are some:
aepb.gov.cn
bjmemc.com.cn
cdemc.cn
cfhb.gov.cn
cepb.gov.cn
dl.gov.cn
dyhb.gov.cn
nbemc.gov.cn
sdein.gov.cn
fjepb.gov.cn
gsep.gansu.gov.cn
qhepb.gov.cn
gdep.gov.cn
gxepb.gov.cn
ghb.gov.cn
dloer.gov.cn
hebei.gov.cn
hljdep.gov.cn
hnep.gov.cn
hbepb.gov.cn
hbt.hunan.gov.cn
nmgepb.gov.cn
jshb.gov.cn
jxepb.gov.cn
shbj.klmy.gov.cn
lnemc.cn
lzhb.gov.cn
nnems.gov.cn
nxep.gov.cn
ordoshb.gov.cn
semc.gov.cn
sxhjjcz.com.cn
szhec.gov.cn
tjemc.org.cn
xzep.gov.cn
wlmqhb.gov.cn
whepb.gov.cn
xianemc.gov.cn
xnepb.gov.cn
xjepb.gov.cn
ynepb.gov.cn
zjepb.gov.cn
Sources from: aqicn.org/sources
There is also a site that provides an API to Chinese AQ data:
http://pm25.in/
However the issue is that from what I can tell, these data are not shared in their raw format only AQI (I have not clicked on each and everyone however).
Anyone know differently or find ones above that are shared our in raw format?
|
1.0
|
China - Sources (not useable yet) - There are a ton of China air quality data sources. Here are some:
aepb.gov.cn
bjmemc.com.cn
cdemc.cn
cfhb.gov.cn
cepb.gov.cn
dl.gov.cn
dyhb.gov.cn
nbemc.gov.cn
sdein.gov.cn
fjepb.gov.cn
gsep.gansu.gov.cn
qhepb.gov.cn
gdep.gov.cn
gxepb.gov.cn
ghb.gov.cn
dloer.gov.cn
hebei.gov.cn
hljdep.gov.cn
hnep.gov.cn
hbepb.gov.cn
hbt.hunan.gov.cn
nmgepb.gov.cn
jshb.gov.cn
jxepb.gov.cn
shbj.klmy.gov.cn
lnemc.cn
lzhb.gov.cn
nnems.gov.cn
nxep.gov.cn
ordoshb.gov.cn
semc.gov.cn
sxhjjcz.com.cn
szhec.gov.cn
tjemc.org.cn
xzep.gov.cn
wlmqhb.gov.cn
whepb.gov.cn
xianemc.gov.cn
xnepb.gov.cn
xjepb.gov.cn
ynepb.gov.cn
zjepb.gov.cn
Sources from: aqicn.org/sources
There is also a site that provides an API to Chinese AQ data:
http://pm25.in/
However the issue is that from what I can tell, these data are not shared in their raw format only AQI (I have not clicked on each and everyone however).
Anyone know differently or find ones above that are shared our in raw format?
|
priority
|
china sources not useable yet there are a ton of china air quality data sources here are some aepb gov cn bjmemc com cn cdemc cn cfhb gov cn cepb gov cn dl gov cn dyhb gov cn nbemc gov cn sdein gov cn fjepb gov cn gsep gansu gov cn qhepb gov cn gdep gov cn gxepb gov cn ghb gov cn dloer gov cn hebei gov cn hljdep gov cn hnep gov cn hbepb gov cn hbt hunan gov cn nmgepb gov cn jshb gov cn jxepb gov cn shbj klmy gov cn lnemc cn lzhb gov cn nnems gov cn nxep gov cn ordoshb gov cn semc gov cn sxhjjcz com cn szhec gov cn tjemc org cn xzep gov cn wlmqhb gov cn whepb gov cn xianemc gov cn xnepb gov cn xjepb gov cn ynepb gov cn zjepb gov cn sources from aqicn org sources there is also a site that provides an api to chinese aq data however the issue is that from what i can tell these data are not shared in their raw format only aqi i have not clicked on each and everyone however anyone know differently or find ones above that are shared our in raw format
| 1
|
106,487
| 4,273,247,792
|
IssuesEvent
|
2016-07-13 16:43:31
|
TerraTex-Community/TerraTex-Reallife-Reloaded
|
https://api.github.com/repos/TerraTex-Community/TerraTex-Reallife-Reloaded
|
closed
|
Bug mit Drogen und Materialsverkäufer
|
bug EXTREMLY HIGH PRIORITY / EMERGENCY FIX NEEDED
|
Wenn man jeweils die Fenster auf hat, bei denen man sich Drogen oder Mats kaufen kann, und dabei die Läden verlässt, bleiben diese auf.
Man kann also auf der gesamten Karte Materials und Drogen kaufen. Ebenfalls steigt dadurch der Preis der Gegenstände für einen selbst nicht an. Jedoch für alle anderen.
Auch braucht man keine Masse an Drogen mehr mit sich herum tragen. und könte damit Kontrollen umgehen.
ich schreibe es hier herein da ich irgendwie keinen Bugtracker gefunden habe und da Stalker mich dann auf das Github hingewiesen hat
|
1.0
|
Bug mit Drogen und Materialsverkäufer - Wenn man jeweils die Fenster auf hat, bei denen man sich Drogen oder Mats kaufen kann, und dabei die Läden verlässt, bleiben diese auf.
Man kann also auf der gesamten Karte Materials und Drogen kaufen. Ebenfalls steigt dadurch der Preis der Gegenstände für einen selbst nicht an. Jedoch für alle anderen.
Auch braucht man keine Masse an Drogen mehr mit sich herum tragen. und könte damit Kontrollen umgehen.
ich schreibe es hier herein da ich irgendwie keinen Bugtracker gefunden habe und da Stalker mich dann auf das Github hingewiesen hat
|
priority
|
bug mit drogen und materialsverkäufer wenn man jeweils die fenster auf hat bei denen man sich drogen oder mats kaufen kann und dabei die läden verlässt bleiben diese auf man kann also auf der gesamten karte materials und drogen kaufen ebenfalls steigt dadurch der preis der gegenstände für einen selbst nicht an jedoch für alle anderen auch braucht man keine masse an drogen mehr mit sich herum tragen und könte damit kontrollen umgehen ich schreibe es hier herein da ich irgendwie keinen bugtracker gefunden habe und da stalker mich dann auf das github hingewiesen hat
| 1
|
468,748
| 13,489,753,229
|
IssuesEvent
|
2020-09-11 14:16:09
|
pxblue/angular-component-library
|
https://api.github.com/repos/pxblue/angular-component-library
|
closed
|
Components should set display rule for first rendered elements
|
bug high-priority info-list-item
|
#### Describe the bug
The first element in our component DOM trees do not always set a display type. This can cause unexpected behavior when trying to style certain components. This is a known issue in the InfoListItem, but we should verify the other components as well.
#### What is the expected behavior?
First element in component trees should explicitly set the desired display type.
#### Anything else to add?
When the browser doesn't recognize a HTML tag (such as for custom components), it renders them with display:inline by default (with height/width of 0 and certain CSS rules having no effect). When the first element in the DOM of a component is also an unrecognized browser tag (such as `<mat-list-item>`), that seems to be when the issues arise - we just need to make sure these root elements have an explicit display type.
Do this with #111
|
1.0
|
Components should set display rule for first rendered elements - #### Describe the bug
The first element in our component DOM trees do not always set a display type. This can cause unexpected behavior when trying to style certain components. This is a known issue in the InfoListItem, but we should verify the other components as well.
#### What is the expected behavior?
First element in component trees should explicitly set the desired display type.
#### Anything else to add?
When the browser doesn't recognize a HTML tag (such as for custom components), it renders them with display:inline by default (with height/width of 0 and certain CSS rules having no effect). When the first element in the DOM of a component is also an unrecognized browser tag (such as `<mat-list-item>`), that seems to be when the issues arise - we just need to make sure these root elements have an explicit display type.
Do this with #111
|
priority
|
components should set display rule for first rendered elements describe the bug the first element in our component dom trees do not always set a display type this can cause unexpected behavior when trying to style certain components this is a known issue in the infolistitem but we should verify the other components as well what is the expected behavior first element in component trees should explicitly set the desired display type anything else to add when the browser doesn t recognize a html tag such as for custom components it renders them with display inline by default with height width of and certain css rules having no effect when the first element in the dom of a component is also an unrecognized browser tag such as that seems to be when the issues arise we just need to make sure these root elements have an explicit display type do this with
| 1
|
371,674
| 10,979,908,571
|
IssuesEvent
|
2019-11-30 09:55:43
|
OperationCode/front-end
|
https://api.github.com/repos/OperationCode/front-end
|
closed
|
Cannot register with apostrophies in name
|
Priority: High Type: Bug beginner friendly good first issue
|
# Bug Report
## What is the current behavior?
Doing names like: N'golo or O'Hara are impossible due to incorrect validation of the name inputs.
## What is the expected behavior?
The name fields should have no validation.
## What steps did you take to get this behavior?
Noticed a Cypress test failing with randomly generated data
### Screenshots

|
1.0
|
Cannot register with apostrophies in name - # Bug Report
## What is the current behavior?
Doing names like: N'golo or O'Hara are impossible due to incorrect validation of the name inputs.
## What is the expected behavior?
The name fields should have no validation.
## What steps did you take to get this behavior?
Noticed a Cypress test failing with randomly generated data
### Screenshots

|
priority
|
cannot register with apostrophies in name bug report what is the current behavior doing names like n golo or o hara are impossible due to incorrect validation of the name inputs what is the expected behavior the name fields should have no validation what steps did you take to get this behavior noticed a cypress test failing with randomly generated data screenshots
| 1
|
82,592
| 3,615,800,127
|
IssuesEvent
|
2016-02-07 00:46:04
|
haskell/cabal
|
https://api.github.com/repos/haskell/cabal
|
closed
|
manual cabal flag toggled by modular solver
|
bug high-priority solver
|
With GHC 7.9/7.10, only `deepseq>=1.4.0.0` is installable.
`semigroups-0.16` has the following flag defined:
```
flag deepseq
description:
You can disable the use of the `deepseq` package using `-f-deepseq`.
.
Disabing this is an unsupported configuration, but it may be useful for accelerating builds in sandboxes for expert users.
default: True
manual: True
```
and it's used for toggling a build-dep:
```
if flag(deepseq)
build-depends: deepseq >= 1.1 && < 1.4
```
So it's expected that `semigroups-0.16` isn't installable with GHC 7.9/7.10. However, Cabal decides to toggle the `deepseq` flag, even though it isn't allowed to:
```
$ cabal install semigroups-0.16 -v2
Reading available packages...
Choosing modular solver.
Resolving dependencies...
Ready to install semigroups-0.16
Waiting for install task to finish...
Extracting
/home/hvr/.cabal/packages/hackage.haskell.org/semigroups/0.16/semigroups-0.16.tar.gz
to /tmp/semigroups-0.16-25781...
Updating semigroups.cabal with the latest revision from the index.
Configuring semigroups-0.16...
Flags chosen: unordered-containers=True, text=True, deepseq=False,
containers=True, bytestring=True, hashable=True
Dependency base ==4.8.0.0: using base-4.8.0.0
Dependency bytestring ==0.10.5.0: using bytestring-0.10.5.0
Dependency containers ==0.5.6.1: using containers-0.5.6.1
Dependency hashable ==1.2.3.0: using hashable-1.2.3.0
Dependency nats ==1: using nats-1
Dependency text ==1.2.0.3: using text-1.2.0.3
Dependency unordered-containers ==0.2.5.1: using unordered-containers-0.2.5.1
Using Cabal-1.21.1.0 compiled by ghc-7.4
Using compiler: ghc-7.9.20141216
Using install prefix: /home/hvr/.cabal
Binaries installed in: /home/hvr/.cabal/bin
Libraries installed in:
...
```
as you can see in the logging output above, surprisingly `deepseq=False` is set.
|
1.0
|
manual cabal flag toggled by modular solver - With GHC 7.9/7.10, only `deepseq>=1.4.0.0` is installable.
`semigroups-0.16` has the following flag defined:
```
flag deepseq
description:
You can disable the use of the `deepseq` package using `-f-deepseq`.
.
Disabing this is an unsupported configuration, but it may be useful for accelerating builds in sandboxes for expert users.
default: True
manual: True
```
and it's used for toggling a build-dep:
```
if flag(deepseq)
build-depends: deepseq >= 1.1 && < 1.4
```
So it's expected that `semigroups-0.16` isn't installable with GHC 7.9/7.10. However, Cabal decides to toggle the `deepseq` flag, even though it isn't allowed to:
```
$ cabal install semigroups-0.16 -v2
Reading available packages...
Choosing modular solver.
Resolving dependencies...
Ready to install semigroups-0.16
Waiting for install task to finish...
Extracting
/home/hvr/.cabal/packages/hackage.haskell.org/semigroups/0.16/semigroups-0.16.tar.gz
to /tmp/semigroups-0.16-25781...
Updating semigroups.cabal with the latest revision from the index.
Configuring semigroups-0.16...
Flags chosen: unordered-containers=True, text=True, deepseq=False,
containers=True, bytestring=True, hashable=True
Dependency base ==4.8.0.0: using base-4.8.0.0
Dependency bytestring ==0.10.5.0: using bytestring-0.10.5.0
Dependency containers ==0.5.6.1: using containers-0.5.6.1
Dependency hashable ==1.2.3.0: using hashable-1.2.3.0
Dependency nats ==1: using nats-1
Dependency text ==1.2.0.3: using text-1.2.0.3
Dependency unordered-containers ==0.2.5.1: using unordered-containers-0.2.5.1
Using Cabal-1.21.1.0 compiled by ghc-7.4
Using compiler: ghc-7.9.20141216
Using install prefix: /home/hvr/.cabal
Binaries installed in: /home/hvr/.cabal/bin
Libraries installed in:
...
```
as you can see in the logging output above, surprisingly `deepseq=False` is set.
|
priority
|
manual cabal flag toggled by modular solver with ghc only deepseq is installable semigroups has the following flag defined flag deepseq description you can disable the use of the deepseq package using f deepseq disabing this is an unsupported configuration but it may be useful for accelerating builds in sandboxes for expert users default true manual true and it s used for toggling a build dep if flag deepseq build depends deepseq so it s expected that semigroups isn t installable with ghc however cabal decides to toggle the deepseq flag even though it isn t allowed to cabal install semigroups reading available packages choosing modular solver resolving dependencies ready to install semigroups waiting for install task to finish extracting home hvr cabal packages hackage haskell org semigroups semigroups tar gz to tmp semigroups updating semigroups cabal with the latest revision from the index configuring semigroups flags chosen unordered containers true text true deepseq false containers true bytestring true hashable true dependency base using base dependency bytestring using bytestring dependency containers using containers dependency hashable using hashable dependency nats using nats dependency text using text dependency unordered containers using unordered containers using cabal compiled by ghc using compiler ghc using install prefix home hvr cabal binaries installed in home hvr cabal bin libraries installed in as you can see in the logging output above surprisingly deepseq false is set
| 1
|
354,546
| 10,569,210,203
|
IssuesEvent
|
2019-10-06 18:05:26
|
AY1920S1-CS2113-T14-1/main
|
https://api.github.com/repos/AY1920S1-CS2113-T14-1/main
|
opened
|
Generate final UML diagrams
|
component.Model priority.High type.Task
|
When preparing for final release, use the [Sketch It plugin for Intellij] (https://plugins.jetbrains.com/plugin/10387-sketch-it-) to generate final UML class diagrams for the Developer Guide.
|
1.0
|
Generate final UML diagrams - When preparing for final release, use the [Sketch It plugin for Intellij] (https://plugins.jetbrains.com/plugin/10387-sketch-it-) to generate final UML class diagrams for the Developer Guide.
|
priority
|
generate final uml diagrams when preparing for final release use the to generate final uml class diagrams for the developer guide
| 1
|
131,012
| 5,141,073,820
|
IssuesEvent
|
2017-01-12 08:39:50
|
hugopeek/romanesco-patterns
|
https://api.github.com/repos/hugopeek/romanesco-patterns
|
closed
|
setUserPlaceholders throws fatal error if user ID doesn't exist
|
Confirmed bug High priority
|
Fatal error: Call to a member function getOne() on null in /var/www/ddgc-merge/core/cache/includes/elements/modsnippet/83.include.cache.php on line 10
|
1.0
|
setUserPlaceholders throws fatal error if user ID doesn't exist - Fatal error: Call to a member function getOne() on null in /var/www/ddgc-merge/core/cache/includes/elements/modsnippet/83.include.cache.php on line 10
|
priority
|
setuserplaceholders throws fatal error if user id doesn t exist fatal error call to a member function getone on null in var www ddgc merge core cache includes elements modsnippet include cache php on line
| 1
|
815,616
| 30,564,530,419
|
IssuesEvent
|
2023-07-20 16:44:23
|
gamefreedomgit/Maelstrom
|
https://api.github.com/repos/gamefreedomgit/Maelstrom
|
closed
|
Valor points are all sorts of fucked up.
|
Core Status: Need Info Priority: High
|
[//]: # (REMBEMBER! Add links to things related to the bug using for example:)
[//]: # (http://wowhead.com/)
[//]: # (cata-twinhead.twinstar.cz)
**Description:**
Raids are giving inconsistent amounts of valor points. They're all over the place.
(no chars have done any dungeons this reset or anything else to gain valor + chars have been logged out and in to ensure its not a visual bug)
Char 1: Completed BH10/BH25 = 175/1250 Valor
Char 2: Completed BH10/BH25/Tot4W10/Tot4W25 = 280/1250 Valor
Char 3: Completed BH25 = 70/1250 Valor
Char 4: Completed BH10 = 70/1250 Valor
Char 4 went into a BH25 and it's total went to 140/1250 (+70), completed the same raids as Char 1 and has 35 Valor less.
Char 1 then went into a Tot4W10.
After the first boss the chat window said +70 and the Valor points went to 210/1250 (an actual increase of 35)
After the second boss the chat window said +70 and the Valor points went to 210/1250
**How to reproduce:**
Go into raids.
**How it should work:**
From what I can see online it should be 75 per 10 man boss and 105 per 25 man boss (links below) although I'm not sure if this was changed at a later date, regardless of this the points given are inconsistent.
**Database links:**
https://www.wowhead.com/news/cataclysm-justice-valor-point-drop-quantities-170048
https://www.engadget.com/2010-10-06-cataclysm-justice-and-valor-point-drop-rates-revealed.html
|
1.0
|
Valor points are all sorts of fucked up. - [//]: # (REMBEMBER! Add links to things related to the bug using for example:)
[//]: # (http://wowhead.com/)
[//]: # (cata-twinhead.twinstar.cz)
**Description:**
Raids are giving inconsistent amounts of valor points. They're all over the place.
(no chars have done any dungeons this reset or anything else to gain valor + chars have been logged out and in to ensure its not a visual bug)
Char 1: Completed BH10/BH25 = 175/1250 Valor
Char 2: Completed BH10/BH25/Tot4W10/Tot4W25 = 280/1250 Valor
Char 3: Completed BH25 = 70/1250 Valor
Char 4: Completed BH10 = 70/1250 Valor
Char 4 went into a BH25 and it's total went to 140/1250 (+70), completed the same raids as Char 1 and has 35 Valor less.
Char 1 then went into a Tot4W10.
After the first boss the chat window said +70 and the Valor points went to 210/1250 (an actual increase of 35)
After the second boss the chat window said +70 and the Valor points went to 210/1250
**How to reproduce:**
Go into raids.
**How it should work:**
From what I can see online it should be 75 per 10 man boss and 105 per 25 man boss (links below) although I'm not sure if this was changed at a later date, regardless of this the points given are inconsistent.
**Database links:**
https://www.wowhead.com/news/cataclysm-justice-valor-point-drop-quantities-170048
https://www.engadget.com/2010-10-06-cataclysm-justice-and-valor-point-drop-rates-revealed.html
|
priority
|
valor points are all sorts of fucked up rembember add links to things related to the bug using for example cata twinhead twinstar cz description raids are giving inconsistent amounts of valor points they re all over the place no chars have done any dungeons this reset or anything else to gain valor chars have been logged out and in to ensure its not a visual bug char completed valor char completed valor char completed valor char completed valor char went into a and it s total went to completed the same raids as char and has valor less char then went into a after the first boss the chat window said and the valor points went to an actual increase of after the second boss the chat window said and the valor points went to how to reproduce go into raids how it should work from what i can see online it should be per man boss and per man boss links below although i m not sure if this was changed at a later date regardless of this the points given are inconsistent database links
| 1
|
343,573
| 10,332,845,312
|
IssuesEvent
|
2019-09-03 02:25:32
|
mateowatson/timetracker
|
https://api.github.com/repos/mateowatson/timetracker
|
closed
|
JavaScript implementation doesn’t support the back button
|
bug high-priority
|
I uncovered this issue while testing Turbolinks. Turbolinks _does_ solve this, but my inclination is to try to fix this in the current system because I don’t want to rush a wholesale update to Turbolinks. And this bug should be fixed ASAP.
|
1.0
|
JavaScript implementation doesn’t support the back button - I uncovered this issue while testing Turbolinks. Turbolinks _does_ solve this, but my inclination is to try to fix this in the current system because I don’t want to rush a wholesale update to Turbolinks. And this bug should be fixed ASAP.
|
priority
|
javascript implementation doesn’t support the back button i uncovered this issue while testing turbolinks turbolinks does solve this but my inclination is to try to fix this in the current system because i don’t want to rush a wholesale update to turbolinks and this bug should be fixed asap
| 1
|
412,993
| 12,059,067,581
|
IssuesEvent
|
2020-04-15 18:36:01
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
[TASK] Pre-loading scripts for tenant
|
Developer-Portal Priority/High Type/New Feature Type/Task
|
**Description:**
The objective of this task is to write scripts to load templates such as IDP_TEMPLATE, APPLICATION_TEMPLATE, etc to the cloud by default.
|
1.0
|
[TASK] Pre-loading scripts for tenant - **Description:**
The objective of this task is to write scripts to load templates such as IDP_TEMPLATE, APPLICATION_TEMPLATE, etc to the cloud by default.
|
priority
|
pre loading scripts for tenant description the objective of this task is to write scripts to load templates such as idp template application template etc to the cloud by default
| 1
|
512,763
| 14,908,934,274
|
IssuesEvent
|
2021-01-22 07:00:27
|
ckiplab/ckip-transformers
|
https://api.github.com/repos/ckiplab/ckip-transformers
|
opened
|
Implement custom Chinese tokenizer.
|
Priority: High
|
We may implement our own tokenizer rather than using BertTokenizerFast.
Our own tokenizer should have the following features:
- Disable word piece. Convert text to token ID character by character (e.g. `tokenizer.convert_tokens_to_ids(list(input_text))`)
- Reimplement `clean_up_tokenization` method. The default method is implemented for English only. Our method may remove whitespaces and convert half-width punctuations to full-width ones.
|
1.0
|
Implement custom Chinese tokenizer. - We may implement our own tokenizer rather than using BertTokenizerFast.
Our own tokenizer should have the following features:
- Disable word piece. Convert text to token ID character by character (e.g. `tokenizer.convert_tokens_to_ids(list(input_text))`)
- Reimplement `clean_up_tokenization` method. The default method is implemented for English only. Our method may remove whitespaces and convert half-width punctuations to full-width ones.
|
priority
|
implement custom chinese tokenizer we may implement our own tokenizer rather than using berttokenizerfast our own tokenizer should have the following features disable word piece convert text to token id character by character e g tokenizer convert tokens to ids list input text reimplement clean up tokenization method the default method is implemented for english only our method may remove whitespaces and convert half width punctuations to full width ones
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.