Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
735,798 | 25,414,725,724 | IssuesEvent | 2022-11-22 22:30:34 | VEuPathDB/web-eda | https://api.github.com/repos/VEuPathDB/web-eda | closed | Bug: Scatterplot has legend issue on QA site | bug high priority | There is now a bug with the legend when an overlay variable is added to the plot. See this GEMS1 example.
**On the live site it looks fine**
<img width="944" alt="image" src="https://user-images.githubusercontent.com/25450900/202033718-9af0b401-1d74-41ad-82d5-ef6390709e6d.png">
**On the QA site we are getting a gradient color map but the colors don't match what is on the plot and this is a categorical variable**
<img width="980" alt="image" src="https://user-images.githubusercontent.com/25450900/202033790-ad59b01e-2166-493b-9b2f-9466253286d5.png">
| 1.0 | Bug: Scatterplot has legend issue on QA site - There is now a bug with the legend when an overlay variable is added to the plot. See this GEMS1 example.
**On the live site it looks fine**
<img width="944" alt="image" src="https://user-images.githubusercontent.com/25450900/202033718-9af0b401-1d74-41ad-82d5-ef6390709e6d.png">
**On the QA site we are getting a gradient color map but the colors don't match what is on the plot and this is a categorical variable**
<img width="980" alt="image" src="https://user-images.githubusercontent.com/25450900/202033790-ad59b01e-2166-493b-9b2f-9466253286d5.png">
| priority | bug scatterplot has legend issue on qa site there is now a bug with the legend when an overlay variable is added to the plot see this example on the live site it looks fine img width alt image src on the qa site we are getting a gradient color map but the colors don t match what is on the plot and this is a categorical variable img width alt image src | 1 |
816,855 | 30,614,577,860 | IssuesEvent | 2023-07-24 01:06:21 | steedos/steedos-platform | https://api.github.com/repos/steedos/steedos-platform | closed | [Bug]: 应用程序选项卡未能显示 | bug done priority: High | ### Description
<img width="1083" alt="image" src="https://github.com/steedos/steedos-platform/assets/6194462/d1fc0a59-b658-4d67-abd9-fd3384ce93a9">
### Steps To Reproduce 重现步骤
2.5.8之前新建的应用程序-选项卡. 在2.5.8版本中未能正常显示.
### Version 版本
2.5.8 | 1.0 | [Bug]: 应用程序选项卡未能显示 - ### Description
<img width="1083" alt="image" src="https://github.com/steedos/steedos-platform/assets/6194462/d1fc0a59-b658-4d67-abd9-fd3384ce93a9">
### Steps To Reproduce 重现步骤
2.5.8之前新建的应用程序-选项卡. 在2.5.8版本中未能正常显示.
### Version 版本
2.5.8 | priority | 应用程序选项卡未能显示 description img width alt image src steps to reproduce 重现步骤 选项卡 version 版本 | 1 |
787,957 | 27,737,077,438 | IssuesEvent | 2023-03-15 11:59:44 | SzFMV2023-Tavasz/AutomatedCar-A | https://api.github.com/repos/SzFMV2023-Tavasz/AutomatedCar-A | opened | Vészfékező | effort: high priority: critical type: user story | A modul felelőssége a radar szenzorra épülő [automata vészfékező rendszer](https://szfmv2023-tavasz.github.io/handout/functions.html#autonóm-vészfékező-rendszer-automatic-emergency-brake---aeb) megvalósítása. A vészfékező kritikus biztonsági funkció, így nem kapcsolható ki manuálisan, de maximum 70 km/h sebességig működik. A működése két esetre bontható: ütközés statikus vagy dinamikus objektummal.
Az előbbi az egyszerűbb eset, mivel a veszélyt jelentő objektum pozíciója változatlan.

El kell dönteni, hogy az autó az aktuális irányvektort figyelembe véve ütközni fog-e az objektummal. Ha igen, az autó ismert sebességét figyelembe véve kiszámolható, hogy ehhez mennyi időre van szükség és, hogy mekkora mértékű lassulás kell ehhez.
A radar visszaadja az autó előtt levő legközelebbi releváns objektum adatait (táv, sebesség), ezekkel lehet számolni. A távolságból és az autó sebességéből meghatározható, hogy milyen lassulást kell adni az autónak, hogy még megálljon, de ne lépje túl a \\( 9 m/s^2 \\)-et.
Ha az ütközés elkerülhető, vizuális figyelmeztetést kell elhelyezni a vezetőnek, hogy fékezzen. Ha nem reagál, azaz továbbra is ütközési pályán vagyunk és már csak vészfékezéssel kerülhető el az ütközés, akkor a hajtásláncnak vészfékezési inputot kell adni. Ez a maximálisan megengedett, \\( 9 m/s^2 \\)-es lassulást (ennél nagyobb lassulás veszélyes az utasokra), akkor
Ha más nem próbálgatással meg kell határozni, hogy adott sebességről egy maximális fékezési input (100% pedál állás) mennyi idő alatt fékezi állóra az autót.
A modul olyan triggerekkel vezérli az autót mint amilyenek a billentyűlenyomás kezelőtől jönnek (fékpedál állás).
Dinamikus objektumok esetében a vészfékezés elve azonos, de az ütközési pálya meghatározása összetettebb.

Másik sávban szembe jövő autóra nem kell vészfékezést kiváltani, tehát el kell tudni dönteni, hogy abban az esetben nincs ütközési pálya.

### Definition of Done
- [ ] Elkerülhető ütközés esetén vizuális figyelmeztetés a sofőrnek
- [ ] Ha a sofőr nem avatkozik közbe, automatikus fékezés (az utolsó pillanatban, ahol az ütközés még elkerülhető)
- [ ] Az automatikus fékezés mértéke a sebességgel arányos, de nem lehet \\( 9 m/s^2 \\)-nél nagyobb
- [ ] 70 km/h felett figyelmeztetés, hogy az AEB nem tud minden helyzetet kezelni
- [ ] A vezérelt autó nem üt el gyalogost, nem megy neki fának
- [ ] Nem releváns objektumok esetében (fals pozitív) mint a szembejövő autó nem történik vészfékezés | 1.0 | Vészfékező - A modul felelőssége a radar szenzorra épülő [automata vészfékező rendszer](https://szfmv2023-tavasz.github.io/handout/functions.html#autonóm-vészfékező-rendszer-automatic-emergency-brake---aeb) megvalósítása. A vészfékező kritikus biztonsági funkció, így nem kapcsolható ki manuálisan, de maximum 70 km/h sebességig működik. A működése két esetre bontható: ütközés statikus vagy dinamikus objektummal.
Az előbbi az egyszerűbb eset, mivel a veszélyt jelentő objektum pozíciója változatlan.

El kell dönteni, hogy az autó az aktuális irányvektort figyelembe véve ütközni fog-e az objektummal. Ha igen, az autó ismert sebességét figyelembe véve kiszámolható, hogy ehhez mennyi időre van szükség és, hogy mekkora mértékű lassulás kell ehhez.
A radar visszaadja az autó előtt levő legközelebbi releváns objektum adatait (táv, sebesség), ezekkel lehet számolni. A távolságból és az autó sebességéből meghatározható, hogy milyen lassulást kell adni az autónak, hogy még megálljon, de ne lépje túl a \\( 9 m/s^2 \\)-et.
Ha az ütközés elkerülhető, vizuális figyelmeztetést kell elhelyezni a vezetőnek, hogy fékezzen. Ha nem reagál, azaz továbbra is ütközési pályán vagyunk és már csak vészfékezéssel kerülhető el az ütközés, akkor a hajtásláncnak vészfékezési inputot kell adni. Ez a maximálisan megengedett, \\( 9 m/s^2 \\)-es lassulást (ennél nagyobb lassulás veszélyes az utasokra), akkor
Ha más nem próbálgatással meg kell határozni, hogy adott sebességről egy maximális fékezési input (100% pedál állás) mennyi idő alatt fékezi állóra az autót.
A modul olyan triggerekkel vezérli az autót mint amilyenek a billentyűlenyomás kezelőtől jönnek (fékpedál állás).
Dinamikus objektumok esetében a vészfékezés elve azonos, de az ütközési pálya meghatározása összetettebb.

Másik sávban szembe jövő autóra nem kell vészfékezést kiváltani, tehát el kell tudni dönteni, hogy abban az esetben nincs ütközési pálya.

### Definition of Done
- [ ] Elkerülhető ütközés esetén vizuális figyelmeztetés a sofőrnek
- [ ] Ha a sofőr nem avatkozik közbe, automatikus fékezés (az utolsó pillanatban, ahol az ütközés még elkerülhető)
- [ ] Az automatikus fékezés mértéke a sebességgel arányos, de nem lehet \\( 9 m/s^2 \\)-nél nagyobb
- [ ] 70 km/h felett figyelmeztetés, hogy az AEB nem tud minden helyzetet kezelni
- [ ] A vezérelt autó nem üt el gyalogost, nem megy neki fának
- [ ] Nem releváns objektumok esetében (fals pozitív) mint a szembejövő autó nem történik vészfékezés | priority | vészfékező a modul felelőssége a radar szenzorra épülő megvalósítása a vészfékező kritikus biztonsági funkció így nem kapcsolható ki manuálisan de maximum km h sebességig működik a működése két esetre bontható ütközés statikus vagy dinamikus objektummal az előbbi az egyszerűbb eset mivel a veszélyt jelentő objektum pozíciója változatlan el kell dönteni hogy az autó az aktuális irányvektort figyelembe véve ütközni fog e az objektummal ha igen az autó ismert sebességét figyelembe véve kiszámolható hogy ehhez mennyi időre van szükség és hogy mekkora mértékű lassulás kell ehhez a radar visszaadja az autó előtt levő legközelebbi releváns objektum adatait táv sebesség ezekkel lehet számolni a távolságból és az autó sebességéből meghatározható hogy milyen lassulást kell adni az autónak hogy még megálljon de ne lépje túl a m s et ha az ütközés elkerülhető vizuális figyelmeztetést kell elhelyezni a vezetőnek hogy fékezzen ha nem reagál azaz továbbra is ütközési pályán vagyunk és már csak vészfékezéssel kerülhető el az ütközés akkor a hajtásláncnak vészfékezési inputot kell adni ez a maximálisan megengedett m s es lassulást ennél nagyobb lassulás veszélyes az utasokra akkor ha más nem próbálgatással meg kell határozni hogy adott sebességről egy maximális fékezési input pedál állás mennyi idő alatt fékezi állóra az autót a modul olyan triggerekkel vezérli az autót mint amilyenek a billentyűlenyomás kezelőtől jönnek fékpedál állás dinamikus objektumok esetében a vészfékezés elve azonos de az ütközési pálya meghatározása összetettebb másik sávban szembe jövő autóra nem kell vészfékezést kiváltani tehát el kell tudni dönteni hogy abban az esetben nincs ütközési pálya definition of done elkerülhető ütközés esetén vizuális figyelmeztetés a sofőrnek ha a sofőr nem avatkozik közbe automatikus fékezés az utolsó pillanatban ahol az ütközés még elkerülhető az automatikus fékezés mértéke a sebességgel arányos de nem lehet m s nél nagyobb km h felett figyelmeztetés hogy az aeb nem tud minden helyzetet kezelni a vezérelt autó nem üt el gyalogost nem megy neki fának nem releváns objektumok esetében fals pozitív mint a szembejövő autó nem történik vészfékezés | 1 |
350,374 | 10,483,067,677 | IssuesEvent | 2019-09-24 13:16:31 | medic/medic | https://api.github.com/repos/medic/medic | closed | Guided Tour does not show control buttons | Priority: 1 - High Type: Bug | **Describe the bug**
There are no control buttons (previous, next, end tour).
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Menu >>
2. Click on 'Guided Tour'
3. Pick any option
**Observed behavior**
There are no control buttons (previous, next, end tour). No way to move between tabs or to exit the tour.
**Expected behavior**
The dialog box should show 3 buttons (`previous`, `next`, `end tour`).
**Logs**
No errors in console.
**Screenshots**

**Environment**
- Instance: local
- Browser: Firefox, Chrome
- Client platform: MacOS
- App: webapp, admin
- Version: 3.6.*
| 1.0 | Guided Tour does not show control buttons - **Describe the bug**
There are no control buttons (previous, next, end tour).
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Menu >>
2. Click on 'Guided Tour'
3. Pick any option
**Observed behavior**
There are no control buttons (previous, next, end tour). No way to move between tabs or to exit the tour.
**Expected behavior**
The dialog box should show 3 buttons (`previous`, `next`, `end tour`).
**Logs**
No errors in console.
**Screenshots**

**Environment**
- Instance: local
- Browser: Firefox, Chrome
- Client platform: MacOS
- App: webapp, admin
- Version: 3.6.*
| priority | guided tour does not show control buttons describe the bug there are no control buttons previous next end tour to reproduce steps to reproduce the behavior go to menu click on guided tour pick any option observed behavior there are no control buttons previous next end tour no way to move between tabs or to exit the tour expected behavior the dialog box should show buttons previous next end tour logs no errors in console screenshots environment instance local browser firefox chrome client platform macos app webapp admin version | 1 |
773,225 | 27,150,638,167 | IssuesEvent | 2023-02-17 00:45:19 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | [Moved from Discord] Pets are not stacking Focus fire | Class: Hunter Pet Talent Priority: High Status: Confirmed | Pet IDs below are shown not to stack focus fire when attacking, talents have been checked






| 1.0 | [Moved from Discord] Pets are not stacking Focus fire - Pet IDs below are shown not to stack focus fire when attacking, talents have been checked






| priority | pets are not stacking focus fire pet ids below are shown not to stack focus fire when attacking talents have been checked | 1 |
318,485 | 9,693,278,828 | IssuesEvent | 2019-05-24 15:41:07 | geosolutions-it/MapStore2-C027 | https://api.github.com/repos/geosolutions-it/MapStore2-C027 | closed | Supporto LDAP - MS2 | Priority: High Project: C027 | At the moment the MapStore's LDAP support does not allow to read users within nested groups. By implementing this feature in MapStore, GS, and GF there will be the possibility to set auth rules with greater granularity in GF; however the possibility to authenticate users within nested domain groups in LDAP should be allowed in MS.
The current configuration deployed in the client's infrastructure is composed by MapStore, GeoServer and GeoFence independently connected to the same LDAP path with the same configuration. The authentications GeoServer side for OGC requests are managed through Authkey generated by MapStore.
The client's MS version is the stable 2018.01.xx of 13 Feb 2018, the proposal [here](https://docs.google.com/document/d/1IbKi3dWXvxzVf_sR3iv3HwoJdGJGbrNwqoFKIyz5t0E/edit#heading=h.hpvkr3wxmvs0) | 1.0 | Supporto LDAP - MS2 - At the moment the MapStore's LDAP support does not allow to read users within nested groups. By implementing this feature in MapStore, GS, and GF there will be the possibility to set auth rules with greater granularity in GF; however the possibility to authenticate users within nested domain groups in LDAP should be allowed in MS.
The current configuration deployed in the client's infrastructure is composed by MapStore, GeoServer and GeoFence independently connected to the same LDAP path with the same configuration. The authentications GeoServer side for OGC requests are managed through Authkey generated by MapStore.
The client's MS version is the stable 2018.01.xx of 13 Feb 2018, the proposal [here](https://docs.google.com/document/d/1IbKi3dWXvxzVf_sR3iv3HwoJdGJGbrNwqoFKIyz5t0E/edit#heading=h.hpvkr3wxmvs0) | priority | supporto ldap at the moment the mapstore s ldap support does not allow to read users within nested groups by implementing this feature in mapstore gs and gf there will be the possibility to set auth rules with greater granularity in gf however the possibility to authenticate users within nested domain groups in ldap should be allowed in ms the current configuration deployed in the client s infrastructure is composed by mapstore geoserver and geofence independently connected to the same ldap path with the same configuration the authentications geoserver side for ogc requests are managed through authkey generated by mapstore the client s ms version is the stable xx of feb the proposal | 1 |
33,988 | 2,774,188,062 | IssuesEvent | 2015-05-04 06:14:02 | punongbayan-araullo/tickets | https://api.github.com/repos/punongbayan-araullo/tickets | opened | Move ePayroll website from Diadem 3 to Diadem 6 | priority - high status - accepted system - paysql | Move ePayroll website from Diadem 3 to Diadem 6 | 1.0 | Move ePayroll website from Diadem 3 to Diadem 6 - Move ePayroll website from Diadem 3 to Diadem 6 | priority | move epayroll website from diadem to diadem move epayroll website from diadem to diadem | 1 |
174,457 | 6,540,249,251 | IssuesEvent | 2017-09-01 14:45:13 | envistaInteractive/itagroup-ecommerce-template | https://api.github.com/repos/envistaInteractive/itagroup-ecommerce-template | opened | scss template variable customization | CSS Module enhancement High Priority | Restructure scss scaffolding to allow for customizable scss. | 1.0 | scss template variable customization - Restructure scss scaffolding to allow for customizable scss. | priority | scss template variable customization restructure scss scaffolding to allow for customizable scss | 1 |
24,634 | 2,671,258,850 | IssuesEvent | 2015-03-24 04:02:15 | thedouglenz/ape | https://api.github.com/repos/thedouglenz/ape | closed | Controllerize the outer page stuffs | High Priority | The outer stuff shows up on all pages and needs the nav bar to be inside it but the nav bar needs some program logic to hade and show the login/logout/register buttons when necessary | 1.0 | Controllerize the outer page stuffs - The outer stuff shows up on all pages and needs the nav bar to be inside it but the nav bar needs some program logic to hade and show the login/logout/register buttons when necessary | priority | controllerize the outer page stuffs the outer stuff shows up on all pages and needs the nav bar to be inside it but the nav bar needs some program logic to hade and show the login logout register buttons when necessary | 1 |
274,470 | 8,561,732,492 | IssuesEvent | 2018-11-09 08:16:57 | CS2113-AY1819S1-W13-2/main | https://api.github.com/repos/CS2113-AY1819S1-W13-2/main | reopened | "edit" command allows ending time to be earlier than starting time | duplicate priority.high type.bug | **Describe the bug**
When I entered the command "edit 1 te/0670" (a nonsense time format which is also earlier than the starting time), the update is successful.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to input textbox.
2. Type "edit 1 te/0670".
3. Press enter.
4. Click on entry #1.
5. See error.
**Expected behavior**
Program should prevent nonsense updates/entries.
**Screenshots**

<hr>
**Reported by:** @ongweekeong
**Severity:** `Low`
<sub>[original: nusCS2113-AY1819S1/pe-1#610]</sub> | 1.0 | "edit" command allows ending time to be earlier than starting time - **Describe the bug**
When I entered the command "edit 1 te/0670" (a nonsense time format which is also earlier than the starting time), the update is successful.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to input textbox.
2. Type "edit 1 te/0670".
3. Press enter.
4. Click on entry #1.
5. See error.
**Expected behavior**
Program should prevent nonsense updates/entries.
**Screenshots**

<hr>
**Reported by:** @ongweekeong
**Severity:** `Low`
<sub>[original: nusCS2113-AY1819S1/pe-1#610]</sub> | priority | edit command allows ending time to be earlier than starting time describe the bug when i entered the command edit te a nonsense time format which is also earlier than the starting time the update is successful to reproduce steps to reproduce the behavior go to input textbox type edit te press enter click on entry see error expected behavior program should prevent nonsense updates entries screenshots reported by ongweekeong severity low | 1 |
376,923 | 11,158,178,858 | IssuesEvent | 2019-12-25 18:22:52 | clabe45/vidar | https://api.github.com/repos/clabe45/vidar | closed | Cache properties | priority:high type:bug type:enhancement | ~~**Blocked by #15**~~
Because function properties don't have to be deterministic (for a given time in the movie), yet they should be constant for each frame, they should be cached per-frame. This way, multiple references to one property won't yield different results in one frame. | 1.0 | Cache properties - ~~**Blocked by #15**~~
Because function properties don't have to be deterministic (for a given time in the movie), yet they should be constant for each frame, they should be cached per-frame. This way, multiple references to one property won't yield different results in one frame. | priority | cache properties blocked by because function properties don t have to be deterministic for a given time in the movie yet they should be constant for each frame they should be cached per frame this way multiple references to one property won t yield different results in one frame | 1 |
580,158 | 17,211,143,157 | IssuesEvent | 2021-07-19 04:44:59 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | opened | [Linter] Detect possible usage of check | Area/CompilerLinter Priority/High Team/CompilerFE Type/Improvement | Rule: Whenever possible we should use check to handle errors. Give a `hint` for those occurrences.
Anti-pattern
```
As much as possible avoid use of T|error in variable declarations. Instead use check and add “error?” to the return type,
For example:
Do NOT do;
public main() {
string|error res = doSomething();
if res is error {
io:println(res.toString());
}
else {
io:println("The result is ", res);
}
}
Instead do:
public main() returns error? {
string res = check doSomething();
io:println("The result is ", res);
}
``` | 1.0 | [Linter] Detect possible usage of check - Rule: Whenever possible we should use check to handle errors. Give a `hint` for those occurrences.
Anti-pattern
```
As much as possible avoid use of T|error in variable declarations. Instead use check and add “error?” to the return type,
For example:
Do NOT do;
public main() {
string|error res = doSomething();
if res is error {
io:println(res.toString());
}
else {
io:println("The result is ", res);
}
}
Instead do:
public main() returns error? {
string res = check doSomething();
io:println("The result is ", res);
}
``` | priority | detect possible usage of check rule whenever possible we should use check to handle errors give a hint for those occurrences anti pattern as much as possible avoid use of t error in variable declarations instead use check and add “error ” to the return type for example do not do public main string error res dosomething if res is error io println res tostring else io println the result is res instead do public main returns error string res check dosomething io println the result is res | 1 |
65,686 | 3,237,609,649 | IssuesEvent | 2015-10-14 12:51:52 | nextgis/ngm_clink_monitoring | https://api.github.com/repos/nextgis/ngm_clink_monitoring | closed | Проблема со статусом | High Priority question | Несмотря на то, что уже есть данные с мобильного приложения, статус объекта почему-то не переходит в статус «Идет строительство». | 1.0 | Проблема со статусом - Несмотря на то, что уже есть данные с мобильного приложения, статус объекта почему-то не переходит в статус «Идет строительство». | priority | проблема со статусом несмотря на то что уже есть данные с мобильного приложения статус объекта почему то не переходит в статус «идет строительство» | 1 |
97,132 | 3,985,387,707 | IssuesEvent | 2016-05-07 21:01:27 | emencia/emencia-django-bazar | https://api.github.com/repos/emencia/emencia-django-bazar | closed | Missing package requirement | bug high-priority | At least setup lack requirement of "django-localflavor".
Also there is some additional package that may be optional but actually allways used but not mandated as required:
* rstview and djangocodemirror (related to formatting);
* sendfile (related to note attachments);
| 1.0 | Missing package requirement - At least setup lack requirement of "django-localflavor".
Also there is some additional package that may be optional but actually allways used but not mandated as required:
* rstview and djangocodemirror (related to formatting);
* sendfile (related to note attachments);
| priority | missing package requirement at least setup lack requirement of django localflavor also there is some additional package that may be optional but actually allways used but not mandated as required rstview and djangocodemirror related to formatting sendfile related to note attachments | 1 |
20,807 | 2,631,260,588 | IssuesEvent | 2015-03-07 00:08:26 | okTurtles/okturtles.com | https://api.github.com/repos/okTurtles/okturtles.com | closed | Get rid of TypeKit | high priority | People don't like it for privacy reasons (valid concern), and many of our visitors block TypeKit (breaking the site):
https://www.reddit.com/r/privacy/comments/2xjy13/love_this_idea_for_fixing_https_with_a/cp12d63
Need to find a good alternative font that looks similar to Proxima Nova.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/9093083-get-rid-of-typekit?utm_campaign=plugin&utm_content=tracker%2F835003&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F835003&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Get rid of TypeKit - People don't like it for privacy reasons (valid concern), and many of our visitors block TypeKit (breaking the site):
https://www.reddit.com/r/privacy/comments/2xjy13/love_this_idea_for_fixing_https_with_a/cp12d63
Need to find a good alternative font that looks similar to Proxima Nova.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/9093083-get-rid-of-typekit?utm_campaign=plugin&utm_content=tracker%2F835003&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F835003&utm_medium=issues&utm_source=github).
</bountysource-plugin> | priority | get rid of typekit people don t like it for privacy reasons valid concern and many of our visitors block typekit breaking the site need to find a good alternative font that looks similar to proxima nova want to back this issue we accept bounties via | 1 |
474,398 | 13,668,920,471 | IssuesEvent | 2020-09-29 00:24:10 | canonical-web-and-design/maas-ui | https://api.github.com/repos/canonical-web-and-design/maas-ui | closed | Main navigation loses active indicator in some subpages | Bug 🐛 Priority: High | Going to `/subnets` highlights the nav, but going to `/fabric` does not. | 1.0 | Main navigation loses active indicator in some subpages - Going to `/subnets` highlights the nav, but going to `/fabric` does not. | priority | main navigation loses active indicator in some subpages going to subnets highlights the nav but going to fabric does not | 1 |
176,891 | 6,568,951,158 | IssuesEvent | 2017-09-09 00:20:47 | open-austin/iced-coffee | https://api.github.com/repos/open-austin/iced-coffee | closed | Membership -> Bylaws | high priority | @seandellis and I talked about this today!
Review our agenda and notes here:
https://docs.google.com/document/d/1P8wbVHkb8Df1LKXiiJjD1nwg1axKNqvhJ9noxe4_4OA/edit?usp=sharing
Bylaws here:
https://docs.google.com/document/d/1D6mhqrxokBCXG8g9I5XXKtBvUyjUCJS4_njqc1N9DVs/edit?usp=sharing
Vickies notes here:
https://docs.google.com/document/d/1P8wbVHkb8Df1LKXiiJjD1nwg1axKNqvhJ9noxe4_4OA/edit?usp=sharing
To-Do
- [x] Determine necessary information to be an affiliate member #257 // @seandellis + Me to create google doc
- [x] Translate Vickie's Doc to bylaws and have it reviewed by Jon + Leadership at next monthly meeting // Me
- [x] Create affiliate form #257 // @seandellis
- [ ] Determine best/easiest way to create and update a "profile". Thinking Slack // Me
| 1.0 | Membership -> Bylaws - @seandellis and I talked about this today!
Review our agenda and notes here:
https://docs.google.com/document/d/1P8wbVHkb8Df1LKXiiJjD1nwg1axKNqvhJ9noxe4_4OA/edit?usp=sharing
Bylaws here:
https://docs.google.com/document/d/1D6mhqrxokBCXG8g9I5XXKtBvUyjUCJS4_njqc1N9DVs/edit?usp=sharing
Vickies notes here:
https://docs.google.com/document/d/1P8wbVHkb8Df1LKXiiJjD1nwg1axKNqvhJ9noxe4_4OA/edit?usp=sharing
To-Do
- [x] Determine necessary information to be an affiliate member #257 // @seandellis + Me to create google doc
- [x] Translate Vickie's Doc to bylaws and have it reviewed by Jon + Leadership at next monthly meeting // Me
- [x] Create affiliate form #257 // @seandellis
- [ ] Determine best/easiest way to create and update a "profile". Thinking Slack // Me
| priority | membership bylaws seandellis and i talked about this today review our agenda and notes here bylaws here vickies notes here to do determine necessary information to be an affiliate member seandellis me to create google doc translate vickie s doc to bylaws and have it reviewed by jon leadership at next monthly meeting me create affiliate form seandellis determine best easiest way to create and update a profile thinking slack me | 1 |
200,226 | 7,001,602,813 | IssuesEvent | 2017-12-18 10:50:53 | meumobi/AMS.Connect | https://api.github.com/repos/meumobi/AMS.Connect | closed | revenu net is not computed from unplugged records of the day | bug high-priority | ### Expected behaviour
Tell us what should happen
### Actual behaviour
margin is setted on Admargin daily file
but when get row margin value is empty
```
[2017-12-16 13:57:26] lumen.DEBUG: Get 'revenu net' row {"margin":"","revenu":0,"Margin is numeric: ":false,"Revenue is numeric: ":true} []
[2017-12-16 13:57:26] lumen.DEBUG: Get 'revenu net' row {"revenu net":"Unknown"} []
```
### Steps to reproduce
1.
2.
3.
### Expected responses
- Why it happens
- How to fix it
- How to test
| 1.0 | revenu net is not computed from unplugged records of the day - ### Expected behaviour
Tell us what should happen
### Actual behaviour
margin is setted on Admargin daily file
but when get row margin value is empty
```
[2017-12-16 13:57:26] lumen.DEBUG: Get 'revenu net' row {"margin":"","revenu":0,"Margin is numeric: ":false,"Revenue is numeric: ":true} []
[2017-12-16 13:57:26] lumen.DEBUG: Get 'revenu net' row {"revenu net":"Unknown"} []
```
### Steps to reproduce
1.
2.
3.
### Expected responses
- Why it happens
- How to fix it
- How to test
| priority | revenu net is not computed from unplugged records of the day expected behaviour tell us what should happen actual behaviour margin is setted on admargin daily file but when get row margin value is empty lumen debug get revenu net row margin revenu margin is numeric false revenue is numeric true lumen debug get revenu net row revenu net unknown steps to reproduce expected responses why it happens how to fix it how to test | 1 |
765,669 | 26,856,781,148 | IssuesEvent | 2023-02-03 15:11:10 | unlock-protocol/unlock | https://api.github.com/repos/unlock-protocol/unlock | closed | Enable claim for Goerli locks | 🚨 High Priority | It looks like we have not enabled free claims on Goerli? [Try here](https://app.unlock-protocol.com/checkout?paywallConfig=%7B%22locks%22%3A%7B%220x24a61c3afe7cdd992eed28ffbe62a7a262496928%22%3A%7B%22network%22%3A5%2C%22skipRecipient%22%3Atrue%7D%7D%2C%22pessimistic%22%3Atrue%2C%22skipRecipient%22%3Atrue%7D)
We should add that as it is a useful flow to "debug" things!
<img width="535" alt="Screenshot 2023-02-02 at 3 58 55 PM" src="https://user-images.githubusercontent.com/17735/216448161-9bcfe65e-5424-4574-bde7-135c5dd952f9.png">
| 1.0 | Enable claim for Goerli locks - It looks like we have not enabled free claims on Goerli? [Try here](https://app.unlock-protocol.com/checkout?paywallConfig=%7B%22locks%22%3A%7B%220x24a61c3afe7cdd992eed28ffbe62a7a262496928%22%3A%7B%22network%22%3A5%2C%22skipRecipient%22%3Atrue%7D%7D%2C%22pessimistic%22%3Atrue%2C%22skipRecipient%22%3Atrue%7D)
We should add that as it is a useful flow to "debug" things!
<img width="535" alt="Screenshot 2023-02-02 at 3 58 55 PM" src="https://user-images.githubusercontent.com/17735/216448161-9bcfe65e-5424-4574-bde7-135c5dd952f9.png">
| priority | enable claim for goerli locks it looks like we have not enabled free claims on goerli we should add that as it is a useful flow to debug things img width alt screenshot at pm src | 1 |
123,650 | 4,866,419,567 | IssuesEvent | 2016-11-14 23:43:51 | phetsims/tasks | https://api.github.com/repos/phetsims/tasks | opened | A better way to pass through configuration | priority:2-high | While working on adding nested options to NumberControl, @pixelzoom said:
> I still think there must be a better approach (than options) to customizing UI components.
I want to take a look at an idea I had for this. | 1.0 | A better way to pass through configuration - While working on adding nested options to NumberControl, @pixelzoom said:
> I still think there must be a better approach (than options) to customizing UI components.
I want to take a look at an idea I had for this. | priority | a better way to pass through configuration while working on adding nested options to numbercontrol pixelzoom said i still think there must be a better approach than options to customizing ui components i want to take a look at an idea i had for this | 1 |
442,115 | 12,739,792,465 | IssuesEvent | 2020-06-26 00:15:02 | ubclaunchpad/rocket2 | https://api.github.com/repos/ubclaunchpad/rocket2 | closed | `/rocket help' logs 'app command triggered incorrectly' | bug high priority | **Describe the bug**
Running `/rocket help` gives the help text correctly, but later also logs that an `app command triggered incorrectly`, which is obviously not the case.
**To Reproduce**
Steps to reproduce the behavior:
1. `/rocket help`
**Expected behavior**
No errors should be logged.
**Screenshots**
```console
2020-06-20T00:37:35.404290984Z {Time: 2020-06-20 00:37:35, Level: [INFO], function: server.handle_commands():104, message: Slash command received}
2020-06-20T00:37:35.405382493Z {Time: 2020-06-20 00:37:35, Level: [INFO], function: server.handle_commands():110, message: Slack signature verified}
2020-06-20T00:37:35.406558906Z {Time: 2020-06-20 00:37:35, Level: [INFO], function: server.handle_commands():113, message: @UCXF0357H: /rocket help}
2020-06-20T00:37:35.409130177Z {Time: 2020-06-20 00:37:35, Level: [INFO], function: parser.handle_app_command():62, message: Help command was called}
2020-06-20T00:37:35.410432442Z {Time: 2020-06-20 00:37:35, Level: [ERROR], function: parser.handle_app_command():70, message: app command triggered incorrectly}
``` | 1.0 | `/rocket help' logs 'app command triggered incorrectly' - **Describe the bug**
Running `/rocket help` gives the help text correctly, but later also logs that an `app command triggered incorrectly`, which is obviously not the case.
**To Reproduce**
Steps to reproduce the behavior:
1. `/rocket help`
**Expected behavior**
No errors should be logged.
**Screenshots**
```console
2020-06-20T00:37:35.404290984Z {Time: 2020-06-20 00:37:35, Level: [INFO], function: server.handle_commands():104, message: Slash command received}
2020-06-20T00:37:35.405382493Z {Time: 2020-06-20 00:37:35, Level: [INFO], function: server.handle_commands():110, message: Slack signature verified}
2020-06-20T00:37:35.406558906Z {Time: 2020-06-20 00:37:35, Level: [INFO], function: server.handle_commands():113, message: @UCXF0357H: /rocket help}
2020-06-20T00:37:35.409130177Z {Time: 2020-06-20 00:37:35, Level: [INFO], function: parser.handle_app_command():62, message: Help command was called}
2020-06-20T00:37:35.410432442Z {Time: 2020-06-20 00:37:35, Level: [ERROR], function: parser.handle_app_command():70, message: app command triggered incorrectly}
``` | priority | rocket help logs app command triggered incorrectly describe the bug running rocket help gives the help text correctly but later also logs that an app command triggered incorrectly which is obviously not the case to reproduce steps to reproduce the behavior rocket help expected behavior no errors should be logged screenshots console time level function server handle commands message slash command received time level function server handle commands message slack signature verified time level function server handle commands message rocket help time level function parser handle app command message help command was called time level function parser handle app command message app command triggered incorrectly | 1 |
240,138 | 7,800,472,774 | IssuesEvent | 2018-06-09 09:50:55 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0009464:
user grid does not refresh after ctx menu action | Admin Bug Mantis high priority | **Reported by pschuele on 27 Dec 2013 13:56**
**Version:** Kristina (2013.03.8)
user grid does not refresh after ctx menu action like enable/disable user
| 1.0 | 0009464:
user grid does not refresh after ctx menu action - **Reported by pschuele on 27 Dec 2013 13:56**
**Version:** Kristina (2013.03.8)
user grid does not refresh after ctx menu action like enable/disable user
| priority | user grid does not refresh after ctx menu action reported by pschuele on dec version kristina user grid does not refresh after ctx menu action like enable disable user | 1 |
802,853 | 29,047,588,390 | IssuesEvent | 2023-05-13 19:19:02 | exellian/rex | https://api.github.com/repos/exellian/rex | closed | DRAFT: Template rendering function design | draft priority: high | A really important problem which needs to be solved is the design of the rendering function of a template. This function should not only render an arbitrary string instead it should be able to be used also for hydration and rendering a template to different objects. Therefore the function must be generated in a way that is very generic.
idea:
A template is built out of tree structure called Nodes (Tags, VoidTags, Text and Expressions):
```html
<div>
{ 0 + 3 } // Expression
<div> //Tag with child nodes
...
</div>
this is the text of a text node
</div>
```
So each node is either a leaf (TextNode, VoidTag, Expressions) or a Node with children (TagNodes, Expressions).
The real problem is that an Expression can not only be a leaf but also a node in this tree. In the following the individual cases are shown were Expressions are nodes with children:
```html
<div>
{ <div> hallo </div> } // Expression which contains a tag with children. // Type: Node
{ // Expression which contains a for loop // Type: Array<Node>
for x in 0..10 {
<div></div>
}
}
{ // Expression which contains a for loop // Type: Array<Array<Node>>
for x in 0..10 {
for x in 0..10 {
<div></div>
}
}
}
{ // Expression which has an if block // Type: Node
if success {
<div>success!</div>
} else {
8 + 9 // Any other datatype which is not of type Node or Array<Node> gets automatically converted to a TextNode
}
}
{ (<div></div>) } // We could allow a Node inside a group, but this should probably discussed later on!
{ x.test } //
```
So basically an expression embedded directly into the template can only have the type Node, Array<Node>, Array<Array<Node>>, Array<Array<Array<Node>>>, ...
Because we don't want to have different possible types of expressions the idea is to transform expressions in a way so that simply every expression has type Array<Node>. To do this we have to generate for every non Node expression a make_text_node statement, for every simply Node a make_vector_from_one_element statement, for every Array<Node> statement nothing and if we have a for loop expression, we have to perform a flat_map from Array<Array<Node>> to Array<Node>
| 1.0 | DRAFT: Template rendering function design - A really important problem which needs to be solved is the design of the rendering function of a template. This function should not only render an arbitrary string instead it should be able to be used also for hydration and rendering a template to different objects. Therefore the function must be generated in a way that is very generic.
idea:
A template is built out of tree structure called Nodes (Tags, VoidTags, Text and Expressions):
```html
<div>
{ 0 + 3 } // Expression
<div> //Tag with child nodes
...
</div>
this is the text of a text node
</div>
```
So each node is either a leaf (TextNode, VoidTag, Expressions) or a Node with children (TagNodes, Expressions).
The real problem is that an Expression can not only be a leaf but also a node in this tree. In the following the individual cases are shown were Expressions are nodes with children:
```html
<div>
{ <div> hallo </div> } // Expression which contains a tag with children. // Type: Node
{ // Expression which contains a for loop // Type: Array<Node>
for x in 0..10 {
<div></div>
}
}
{ // Expression which contains a for loop // Type: Array<Array<Node>>
for x in 0..10 {
for x in 0..10 {
<div></div>
}
}
}
{ // Expression which has an if block // Type: Node
if success {
<div>success!</div>
} else {
8 + 9 // Any other datatype which is not of type Node or Array<Node> gets automatically converted to a TextNode
}
}
{ (<div></div>) } // We could allow a Node inside a group, but this should probably discussed later on!
{ x.test } //
```
So basically an expression embedded directly into the template can only have the type Node, Array<Node>, Array<Array<Node>>, Array<Array<Array<Node>>>, ...
Because we don't want to have different possible types of expressions the idea is to transform expressions in a way so that simply every expression has type Array<Node>. To do this we have to generate for every non Node expression a make_text_node statement, for every simply Node a make_vector_from_one_element statement, for every Array<Node> statement nothing and if we have a for loop expression, we have to perform a flat_map from Array<Array<Node>> to Array<Node>
| priority | draft template rendering function design a really important problem which needs to be solved is the design of the rendering function of a template this function should not only render an arbitrary string instead it should be able to be used also for hydration and rendering a template to different objects therefore the function must be generated in a way that is very generic idea a template is built out of tree structure called nodes tags voidtags text and expressions html expression tag with child nodes this is the text of a text node so each node is either a leaf textnode voidtag expressions or a node with children tagnodes expressions the real problem is that an expression can not only be a leaf but also a node in this tree in the following the individual cases are shown were expressions are nodes with children html hallo expression which contains a tag with children type node expression which contains a for loop type array for x in expression which contains a for loop type array for x in for x in expression which has an if block type node if success success else any other datatype which is not of type node or array gets automatically converted to a textnode we could allow a node inside a group but this should probably discussed later on x test so basically an expression embedded directly into the template can only have the type node array array array because we don t want to have different possible types of expressions the idea is to transform expressions in a way so that simply every expression has type array to do this we have to generate for every non node expression a make text node statement for every simply node a make vector from one element statement for every array statement nothing and if we have a for loop expression we have to perform a flat map from array to array | 1 |
419,434 | 12,223,588,717 | IssuesEvent | 2020-05-02 18:16:49 | Broken-Gem-Studio/Broken-Engine | https://api.github.com/repos/Broken-Gem-Studio/Broken-Engine | closed | Dt delay | Bug High Priority | ## Bug Description
dt has a delay of some kind, time is not advancing properly
## Type of Bug
Select the type of bug with and "x" ([x])
* [ ] Visual
* [ ] Physics
* [ ] Audio
* [ ] Particles
* [ ] Resource Management & Save/Load
* [ ] Materials
* [ ] Components
* [ ] Game Objects
* [ ] UI/UX
* [X] Scripting
* [X] Other
## Severity
Select the severity of bug affection and mark with "x" ([x])
- [ ] Crash
- [X] Game stopper/slower
- [ ] Cosmetic
## Reproduction
Steps to reproduce the behavior:
1. just enter play mode, or use time in a script
2.
3.
4.
## Frequency
Select the frequency with which the bug appears and mark it "x" ([x])
* [X] Always
* [ ] Very Often
* [ ] Usually
* [ ] Few Times
* [ ] Few Times under specific conditions
## Conduct
### Expected result:
time works correctly
### Actual result:
time advances slowly
## Screenshots and Illustrations:
## Build
- **Please specify the build:** ``Insert the build here``
latest dev commit 2225
## Observations and Additional Information
| 1.0 | Dt delay - ## Bug Description
dt has a delay of some kind, time is not advancing properly
## Type of Bug
Select the type of bug with and "x" ([x])
* [ ] Visual
* [ ] Physics
* [ ] Audio
* [ ] Particles
* [ ] Resource Management & Save/Load
* [ ] Materials
* [ ] Components
* [ ] Game Objects
* [ ] UI/UX
* [X] Scripting
* [X] Other
## Severity
Select the severity of bug affection and mark with "x" ([x])
- [ ] Crash
- [X] Game stopper/slower
- [ ] Cosmetic
## Reproduction
Steps to reproduce the behavior:
1. just enter play mode, or use time in a script
2.
3.
4.
## Frequency
Select the frequency with which the bug appears and mark it "x" ([x])
* [X] Always
* [ ] Very Often
* [ ] Usually
* [ ] Few Times
* [ ] Few Times under specific conditions
## Conduct
### Expected result:
time works correctly
### Actual result:
time advances slowly
## Screenshots and Illustrations:
## Build
- **Please specify the build:** ``Insert the build here``
latest dev commit 2225
## Observations and Additional Information
| priority | dt delay bug description dt has a delay of some kind time is not advancing properly type of bug select the type of bug with and x visual physics audio particles resource management save load materials components game objects ui ux scripting other severity select the severity of bug affection and mark with x crash game stopper slower cosmetic reproduction steps to reproduce the behavior just enter play mode or use time in a script frequency select the frequency with which the bug appears and mark it x always very often usually few times few times under specific conditions conduct expected result time works correctly actual result time advances slowly screenshots and illustrations build please specify the build insert the build here latest dev commit observations and additional information | 1 |
146,991 | 5,631,853,766 | IssuesEvent | 2017-04-05 15:22:27 | nus-mtp/nus-oracle | https://api.github.com/repos/nus-mtp/nus-oracle | closed | [SignIn Page] Clicking on any Modal will render wrong loading msg subsequently when logging in | bug high priority UI | * Because no setState for loadingMsgs.LOAD_LOGGING_IN was set when user hid any of the Modal windows for signing up/forget password | 1.0 | [SignIn Page] Clicking on any Modal will render wrong loading msg subsequently when logging in - * Because no setState for loadingMsgs.LOAD_LOGGING_IN was set when user hid any of the Modal windows for signing up/forget password | priority | clicking on any modal will render wrong loading msg subsequently when logging in because no setstate for loadingmsgs load logging in was set when user hid any of the modal windows for signing up forget password | 1 |
526,501 | 15,294,359,233 | IssuesEvent | 2021-02-24 02:18:27 | rdsaliba/notorious-eng | https://api.github.com/repos/rdsaliba/notorious-eng | closed | (F10) Evaluating RUL models for specific asset type | High Priority user story | As a user, I would like to see an RUL model performance for a specific asset type.
Description:
Given a static file filled with asset measurement data from the asset type that the user is consulting, the cbms should be able to train the different models for that asset type with the data provided and should be able to generate the evaluation performance of each model comparing to real RUL values. This is to calculate the RMSE.
Acceptance criteria:
- [x] The system reads data from archived assets for a specific asset type
- [x] The user can specify which ones are for testing and which ones are for training
- [x] Display the number of maximum assets available
- [x] Validate that the number of specified assets is under the maximum
- [x] The system has a generate models button to evaluate all models
- [x] The system has a generate model button for each individual model
- [x] The system generates the trained models and displays the evaluation of each model (RMSE) | 1.0 | (F10) Evaluating RUL models for specific asset type - As a user, I would like to see an RUL model performance for a specific asset type.
Description:
Given a static file filled with asset measurement data from the asset type that the user is consulting, the cbms should be able to train the different models for that asset type with the data provided and should be able to generate the evaluation performance of each model comparing to real RUL values. This is to calculate the RMSE.
Acceptance criteria:
- [x] The system reads data from archived assets for a specific asset type
- [x] The user can specify which ones are for testing and which ones are for training
- [x] Display the number of maximum assets available
- [x] Validate that the number of specified assets is under the maximum
- [x] The system has a generate models button to evaluate all models
- [x] The system has a generate model button for each individual model
- [x] The system generates the trained models and displays the evaluation of each model (RMSE) | priority | evaluating rul models for specific asset type as a user i would like to see an rul model performance for a specific asset type description given a static file filled with asset measurement data from the asset type that the user is consulting the cbms should be able to train the different models for that asset type with the data provided and should be able to generate the evaluation performance of each model comparing to real rul values this is to calculate the rmse acceptance criteria the system reads data from archived assets for a specific asset type the user can specify which ones are for testing and which ones are for training display the number of maximum assets available validate that the number of specified assets is under the maximum the system has a generate models button to evaluate all models the system has a generate model button for each individual model the system generates the trained models and displays the evaluation of each model rmse | 1 |
193,318 | 6,883,924,638 | IssuesEvent | 2017-11-21 11:05:34 | BaselLaserMouse/BakingTray | https://api.github.com/repos/BaselLaserMouse/BakingTray | closed | Extracting imaging FOV from ScanImage is slow. Should cache it. | high priority | It takes about 200 ms to read the imaging FOV each time `SIBT.returnScanSettings` is called and sometimes this is repeatedly. To minimise this reading, we should cache all scan settings in the scanner class and only re-read if any settings that might alter the imaging FOV have changed. i.e. The scan angle, scan angle multipliers, number of microns per optical degree, and the fill-fraction. | 1.0 | Extracting imaging FOV from ScanImage is slow. Should cache it. - It takes about 200 ms to read the imaging FOV each time `SIBT.returnScanSettings` is called and sometimes this is repeatedly. To minimise this reading, we should cache all scan settings in the scanner class and only re-read if any settings that might alter the imaging FOV have changed. i.e. The scan angle, scan angle multipliers, number of microns per optical degree, and the fill-fraction. | priority | extracting imaging fov from scanimage is slow should cache it it takes about ms to read the imaging fov each time sibt returnscansettings is called and sometimes this is repeatedly to minimise this reading we should cache all scan settings in the scanner class and only re read if any settings that might alter the imaging fov have changed i e the scan angle scan angle multipliers number of microns per optical degree and the fill fraction | 1 |
437,287 | 12,566,705,353 | IssuesEvent | 2020-06-08 11:42:38 | CatalogueOfLife/clearinghouse-ui | https://api.github.com/repos/CatalogueOfLife/clearinghouse-ui | closed | Need a place to release a catalogue draft and watch its progress | API change assembly high priority | Releasing a (draft) catalogue requires a button to kick that off. It also needs a place to monitor the long running process. Maybe its wise to think about the generic long running monitoring idea Geoff had? | 1.0 | Need a place to release a catalogue draft and watch its progress - Releasing a (draft) catalogue requires a button to kick that off. It also needs a place to monitor the long running process. Maybe its wise to think about the generic long running monitoring idea Geoff had? | priority | need a place to release a catalogue draft and watch its progress releasing a draft catalogue requires a button to kick that off it also needs a place to monitor the long running process maybe its wise to think about the generic long running monitoring idea geoff had | 1 |
667,611 | 22,494,017,137 | IssuesEvent | 2022-06-23 05:33:45 | jhudsl/intro_to_r | https://api.github.com/repos/jhudsl/intro_to_r | closed | statistics updates | high priority | - update broom usage
- resources for classes on models in R??
- case studies have lots of stats
- tidyverse chapter series on modeling/ML (some from jhudatascience.org and data trail)
| 1.0 | statistics updates - - update broom usage
- resources for classes on models in R??
- case studies have lots of stats
- tidyverse chapter series on modeling/ML (some from jhudatascience.org and data trail)
| priority | statistics updates update broom usage resources for classes on models in r case studies have lots of stats tidyverse chapter series on modeling ml some from jhudatascience org and data trail | 1 |
286,114 | 8,784,043,771 | IssuesEvent | 2018-12-20 08:38:01 | projectacrn/acrn-hypervisor | https://api.github.com/repos/projectacrn/acrn-hypervisor | closed | GPU Mediator shall support virtual displays for guest domains | area: hypervisor priority: high status: closed type: feature | GPU Mediator shall support virtual displays for guest domains Some domains on the platform may have no assigned display pipes or planes and yet may be producing output that will be directed to some display (local through indirect display or remote) With that in mind the domain will require some means of determining the parameters of the physical display target (i.e. virtual VBT or cached EDID information) and GPU mediator will need to in some way provide MMIO state the driver requires to determine a display device is present (when it is not physically) and that the graphics stack should produce a display output buffer. | 1.0 | GPU Mediator shall support virtual displays for guest domains - GPU Mediator shall support virtual displays for guest domains Some domains on the platform may have no assigned display pipes or planes and yet may be producing output that will be directed to some display (local through indirect display or remote) With that in mind the domain will require some means of determining the parameters of the physical display target (i.e. virtual VBT or cached EDID information) and GPU mediator will need to in some way provide MMIO state the driver requires to determine a display device is present (when it is not physically) and that the graphics stack should produce a display output buffer. | priority | gpu mediator shall support virtual displays for guest domains gpu mediator shall support virtual displays for guest domains some domains on the platform may have no assigned display pipes or planes and yet may be producing output that will be directed to some display local through indirect display or remote with that in mind the domain will require some means of determining the parameters of the physical display target i e virtual vbt or cached edid information and gpu mediator will need to in some way provide mmio state the driver requires to determine a display device is present when it is not physically and that the graphics stack should produce a display output buffer | 1 |
767,899 | 26,946,641,997 | IssuesEvent | 2023-02-08 08:38:41 | telstra/open-kilda | https://api.github.com/repos/telstra/open-kilda | opened | flow_id in the request line is ignored while flow update | bug area/api priority/1-highest | Steps
case 1
1) create flow (e.g. with flow_id = flow1) - POST [/v2/flows]
2) make **wrong** put request to update it - - PUT [/v2/flows/flow_not_match]
where body of the request contains flow_id = flow1
note: system shouldn't contain any flow with "flow_not_match" id
actual result: flow is updated successfully
expected result: 404 with the corresponding error message like "flow not found"
case 2
1) create flow (e.g. with flow_id = flow1) - POST [/v2/flows]
2) make put request to update it - - PUT [/v2/flows/flow1]
where body of the request **doesn't contains** flow_id = flow1
e.g. body should contain flow_id = flow_not_match
note: system shouldn't contain any flow with "flow_not_match" id
actual result: 404 with the corresponding error message like "flow not found"
expected result: flow should be updated, flow_id from the body should be ignored
actually we are taking flow_id from body, but it should be taken from the request line
also flow_id update is not supported, so it can be set as nullable in the body
| 1.0 | flow_id in the request line is ignored while flow update - Steps
case 1
1) create flow (e.g. with flow_id = flow1) - POST [/v2/flows]
2) make **wrong** put request to update it - - PUT [/v2/flows/flow_not_match]
where body of the request contains flow_id = flow1
note: system shouldn't contain any flow with "flow_not_match" id
actual result: flow is updated successfully
expected result: 404 with the corresponding error message like "flow not found"
case 2
1) create flow (e.g. with flow_id = flow1) - POST [/v2/flows]
2) make put request to update it - - PUT [/v2/flows/flow1]
where body of the request **doesn't contains** flow_id = flow1
e.g. body should contain flow_id = flow_not_match
note: system shouldn't contain any flow with "flow_not_match" id
actual result: 404 with the corresponding error message like "flow not found"
expected result: flow should be updated, flow_id from the body should be ignored
actually we are taking flow_id from body, but it should be taken from the request line
also flow_id update is not supported, so it can be set as nullable in the body
| priority | flow id in the request line is ignored while flow update steps case create flow e g with flow id post make wrong put request to update it put where body of the request contains flow id note system shouldn t contain any flow with flow not match id actual result flow is updated successfully expected result with the corresponding error message like flow not found case create flow e g with flow id post make put request to update it put where body of the request doesn t contains flow id e g body should contain flow id flow not match note system shouldn t contain any flow with flow not match id actual result with the corresponding error message like flow not found expected result flow should be updated flow id from the body should be ignored actually we are taking flow id from body but it should be taken from the request line also flow id update is not supported so it can be set as nullable in the body | 1 |
632,082 | 20,171,234,560 | IssuesEvent | 2022-02-10 10:35:58 | neo4j/graphql | https://api.github.com/repos/neo4j/graphql | closed | *ConnectOrCreateFieldInputOnCreate and *CreateFieldInput are completely identical | confirmed bug report high priority | While harmonizing the logic for schema augmentation in java, I realized, that the types for `${prefix}ConnectOrCreateFieldInputOnCreate` and `${prefix}CreateFieldInput` are identical.
For example the following schema:
```graphql
type Movie {
title: String!
id: ID! @id
}
type Actor {
name: String!
movies: [Movie!]! @relationship(type: "ACTED_IN", direction: OUT)
}
```
will generate the following types among others:
```graphql
input ActorMoviesConnectOrCreateFieldInputOnCreate {
node: MovieCreateInput!
}
input ActorMoviesCreateFieldInput {
node: MovieCreateInput!
}
```
And if there are relation properties like:
```graphql
type Movie {
title: String!
isan: String! @unique
}
type Actor {
name: String!
movies: [Movie!]! @relationship(type: "ACTED_IN", direction: OUT, properties: "ActedIn")
}
interface ActedIn {
screentime: Int!
characterName: String
}
```
The following types are generated among others:
```graphql
input ActorMoviesConnectOrCreateFieldInputOnCreate {
edge: ActedInCreateInput!
node: MovieCreateInput!
}
input ActorMoviesCreateFieldInput {
edge: ActedInCreateInput!
node: MovieCreateInput!
}
```
| 1.0 | *ConnectOrCreateFieldInputOnCreate and *CreateFieldInput are completely identical - While harmonizing the logic for schema augmentation in java, I realized, that the types for `${prefix}ConnectOrCreateFieldInputOnCreate` and `${prefix}CreateFieldInput` are identical.
For example the following schema:
```graphql
type Movie {
title: String!
id: ID! @id
}
type Actor {
name: String!
movies: [Movie!]! @relationship(type: "ACTED_IN", direction: OUT)
}
```
will generate the following types among others:
```graphql
input ActorMoviesConnectOrCreateFieldInputOnCreate {
node: MovieCreateInput!
}
input ActorMoviesCreateFieldInput {
node: MovieCreateInput!
}
```
And if there are relation properties like:
```graphql
type Movie {
title: String!
isan: String! @unique
}
type Actor {
name: String!
movies: [Movie!]! @relationship(type: "ACTED_IN", direction: OUT, properties: "ActedIn")
}
interface ActedIn {
screentime: Int!
characterName: String
}
```
The following types are generated among others:
```graphql
input ActorMoviesConnectOrCreateFieldInputOnCreate {
edge: ActedInCreateInput!
node: MovieCreateInput!
}
input ActorMoviesCreateFieldInput {
edge: ActedInCreateInput!
node: MovieCreateInput!
}
```
| priority | connectorcreatefieldinputoncreate and createfieldinput are completely identical while harmonizing the logic for schema augmentation in java i realized that the types for prefix connectorcreatefieldinputoncreate and prefix createfieldinput are identical for example the following schema graphql type movie title string id id id type actor name string movies relationship type acted in direction out will generate the following types among others graphql input actormoviesconnectorcreatefieldinputoncreate node moviecreateinput input actormoviescreatefieldinput node moviecreateinput and if there are relation properties like graphql type movie title string isan string unique type actor name string movies relationship type acted in direction out properties actedin interface actedin screentime int charactername string the following types are generated among others graphql input actormoviesconnectorcreatefieldinputoncreate edge actedincreateinput node moviecreateinput input actormoviescreatefieldinput edge actedincreateinput node moviecreateinput | 1 |
442,477 | 12,746,087,829 | IssuesEvent | 2020-06-26 15:19:38 | a2000-erp-team/WEBERP | https://api.github.com/repos/a2000-erp-team/WEBERP | opened | Remove extra numbers behind memory in GJ transaction | ABIGAIL High Priority | FINANCE-LEDGER-GLOPERATION-Edit [GJ] transaction
 | 1.0 | Remove extra numbers behind memory in GJ transaction - FINANCE-LEDGER-GLOPERATION-Edit [GJ] transaction
 | priority | remove extra numbers behind memory in gj transaction finance ledger gloperation edit transaction | 1 |
369,274 | 10,894,956,385 | IssuesEvent | 2019-11-19 09:44:02 | canonical-web-and-design/ubuntu.com | https://api.github.com/repos/canonical-web-and-design/ubuntu.com | closed | /livepatch incorrectly points to https://ubuntu.com/advantage | Priority: High | https://ubuntu.com/livepatch currently, amongst other, states this:
> Free for personal use
>
> All you need is an Ubuntu One account. Free for 3 machines.
>
> [Get Livepatch](https://ubuntu.com/advantage)
After reading this text and clicking the button, a random user will expect to find information on how to enable live kernel patching on any currently supported Ubuntu release.
Instead, what they currently find at /advantage is a statement that
> Initially, free subscription is available for Ubuntu 14.04 LTS only.

The location where, today, someone wants who wants to enable live kernel patching, should be pointed to, seems to rather be https://auth.livepatch.canonical.com/ though.
---
*Reported from: https://ubuntu.com/livepatch* | 1.0 | /livepatch incorrectly points to https://ubuntu.com/advantage - https://ubuntu.com/livepatch currently, amongst other, states this:
> Free for personal use
>
> All you need is an Ubuntu One account. Free for 3 machines.
>
> [Get Livepatch](https://ubuntu.com/advantage)
After reading this text and clicking the button, a random user will expect to find information on how to enable live kernel patching on any currently supported Ubuntu release.
Instead, what they currently find at /advantage is a statement that
> Initially, free subscription is available for Ubuntu 14.04 LTS only.

The location where, today, someone wants who wants to enable live kernel patching, should be pointed to, seems to rather be https://auth.livepatch.canonical.com/ though.
---
*Reported from: https://ubuntu.com/livepatch* | priority | livepatch incorrectly points to currently amongst other states this free for personal use all you need is an ubuntu one account free for machines after reading this text and clicking the button a random user will expect to find information on how to enable live kernel patching on any currently supported ubuntu release instead what they currently find at advantage is a statement that initially free subscription is available for ubuntu lts only the location where today someone wants who wants to enable live kernel patching should be pointed to seems to rather be though reported from | 1 |
252,098 | 8,031,831,326 | IssuesEvent | 2018-07-28 07:20:36 | opencv/opencv | https://api.github.com/repos/opencv/opencv | opened | Nightly: test_imgproc(Imgproc_InitUndistortMap.accuracy) failure | bug category: imgproc priority: high | [The first failed build](http://pullrequest.opencv.org/buildbot/builders/3_4_coverage-lin64-debug/builds/102/steps/test_imgproc/logs/stdio):
```
[ RUN ] Imgproc_InitUndistortMap.accuracy
/build/3_4_coverage-lin64-debug/opencv/modules/ts/src/ts.cpp:549: Failure
Failed
failure reason: Arithmetic exception
test case #12
seed: 09491e2273330ddc
-----------------------------------
LOG:
General failure:
Arithmetic exception (-6)
-----------------------------------
CONSOLE: .
-----------------------------------
[ FAILED ] Imgproc_InitUndistortMap.accuracy (71 ms)
``` | 1.0 | Nightly: test_imgproc(Imgproc_InitUndistortMap.accuracy) failure - [The first failed build](http://pullrequest.opencv.org/buildbot/builders/3_4_coverage-lin64-debug/builds/102/steps/test_imgproc/logs/stdio):
```
[ RUN ] Imgproc_InitUndistortMap.accuracy
/build/3_4_coverage-lin64-debug/opencv/modules/ts/src/ts.cpp:549: Failure
Failed
failure reason: Arithmetic exception
test case #12
seed: 09491e2273330ddc
-----------------------------------
LOG:
General failure:
Arithmetic exception (-6)
-----------------------------------
CONSOLE: .
-----------------------------------
[ FAILED ] Imgproc_InitUndistortMap.accuracy (71 ms)
``` | priority | nightly test imgproc imgproc initundistortmap accuracy failure imgproc initundistortmap accuracy build coverage debug opencv modules ts src ts cpp failure failed failure reason arithmetic exception test case seed log general failure arithmetic exception console imgproc initundistortmap accuracy ms | 1 |
482,282 | 13,903,942,442 | IssuesEvent | 2020-10-20 07:57:26 | bryntum/support | https://api.github.com/repos/bryntum/support | closed | Calendar sync doesn't work | bug forum high-priority resolved | [Forum post](https://www.bryntum.com/forum/viewtopic.php?f=54&t=15105)
Use this config for Basic Calendar demo. No sync is called after creating a new event.
```
window.calendar = new Calendar({
crudManager : {
eventStore : {
fields : [
{ name : 'room' }
]
},
transport : {
load : {
url : 'data/data.json'
},
sync : {
url : 'data/sync.json'
}
},
autoLoad : true,
autoSync : true
},
``` | 1.0 | Calendar sync doesn't work - [Forum post](https://www.bryntum.com/forum/viewtopic.php?f=54&t=15105)
Use this config for Basic Calendar demo. No sync is called after creating a new event.
```
window.calendar = new Calendar({
crudManager : {
eventStore : {
fields : [
{ name : 'room' }
]
},
transport : {
load : {
url : 'data/data.json'
},
sync : {
url : 'data/sync.json'
}
},
autoLoad : true,
autoSync : true
},
``` | priority | calendar sync doesn t work use this config for basic calendar demo no sync is called after creating a new event window calendar new calendar crudmanager eventstore fields name room transport load url data data json sync url data sync json autoload true autosync true | 1 |
375,719 | 11,133,375,577 | IssuesEvent | 2019-12-20 09:15:28 | projectacrn/acrn-hypervisor | https://api.github.com/repos/projectacrn/acrn-hypervisor | closed | acrn failed to boot when generate hypervisor config source from config app | priority: P2-High type: bug | Environment
git clone https://github.com/projectacrn/acrn-hypervisor.git
HW/Board
KBLNUC
Build link
build from source
Image info
1fe1afd4 (HEAD -> master, origin/master, origin/HEAD) acrn-config: Add ramdisk tag parsing support
Steps
git clone https://github.com/projectacrn/acrn-hypervisor.git
follow guide to launch config app: https://projectacrn.github.io/latest/tutorials/acrn_configuration_tool.html#use-the-acrn-configuration-app
import default board info from misc/acrn-config/xmls/board-xmls/nuc7i7dnb.xml
select scenario: industry
click "generate board src" and "generate scenario src"
make all BOARD=nuc7i7dnb SCENARIO=industry
replace acrn.efi into kblnuc, and launch acrn
Expected result
acrn and sos can launch successfully
Actual result
acrn launch failed
Reproduce rate
3/3
Debugging info
debug logs could be put here or attached in an attachment. | 1.0 | acrn failed to boot when generate hypervisor config source from config app - Environment
git clone https://github.com/projectacrn/acrn-hypervisor.git
HW/Board
KBLNUC
Build link
build from source
Image info
1fe1afd4 (HEAD -> master, origin/master, origin/HEAD) acrn-config: Add ramdisk tag parsing support
Steps
git clone https://github.com/projectacrn/acrn-hypervisor.git
follow guide to launch config app: https://projectacrn.github.io/latest/tutorials/acrn_configuration_tool.html#use-the-acrn-configuration-app
import default board info from misc/acrn-config/xmls/board-xmls/nuc7i7dnb.xml
select scenario: industry
click "generate board src" and "generate scenario src"
make all BOARD=nuc7i7dnb SCENARIO=industry
replace acrn.efi into kblnuc, and launch acrn
Expected result
acrn and sos can launch successfully
Actual result
acrn launch failed
Reproduce rate
3/3
Debugging info
debug logs could be put here or attached in an attachment. | priority | acrn failed to boot when generate hypervisor config source from config app environment git clone hw board kblnuc build link build from source image info head master origin master origin head acrn config add ramdisk tag parsing support steps git clone follow guide to launch config app import default board info from misc acrn config xmls board xmls xml select scenario industry click generate board src and generate scenario src make all board scenario industry replace acrn efi into kblnuc and launch acrn expected result acrn and sos can launch successfully actual result acrn launch failed reproduce rate debugging info debug logs could be put here or attached in an attachment | 1 |
601,434 | 18,408,692,428 | IssuesEvent | 2021-10-13 01:01:13 | gambitph/Stackable | https://api.github.com/repos/gambitph/Stackable | opened | Content Vertical Align in Style tab is not working in backend | bug high priority [version] V3 | <!--
Before posting, make sure that:
1. you are running the latest version of Stackable, and
2. you have searched whether your issue has already been reported
-->
**To Reproduce**
Steps to reproduce the behavior:
1. Add a v3 Feature > Default layout
2. Set block alignment to full width
3. upload an image and set image height to at least 700px
4. Go to Feature > Style tab > Container Size & Spacing
5. Set Content Vertical Align to flex-end
6. see bug in backend (this works in frontend)
this is applicable to all blocks with Container in Style tab
https://user-images.githubusercontent.com/28699204/137048941-3ce51deb-4c75-4389-a253-461d70e9ad64.mov
| 1.0 | Content Vertical Align in Style tab is not working in backend - <!--
Before posting, make sure that:
1. you are running the latest version of Stackable, and
2. you have searched whether your issue has already been reported
-->
**To Reproduce**
Steps to reproduce the behavior:
1. Add a v3 Feature > Default layout
2. Set block alignment to full width
3. upload an image and set image height to at least 700px
4. Go to Feature > Style tab > Container Size & Spacing
5. Set Content Vertical Align to flex-end
6. see bug in backend (this works in frontend)
this is applicable to all blocks with Container in Style tab
https://user-images.githubusercontent.com/28699204/137048941-3ce51deb-4c75-4389-a253-461d70e9ad64.mov
| priority | content vertical align in style tab is not working in backend before posting make sure that you are running the latest version of stackable and you have searched whether your issue has already been reported to reproduce steps to reproduce the behavior add a feature default layout set block alignment to full width upload an image and set image height to at least go to feature style tab container size spacing set content vertical align to flex end see bug in backend this works in frontend this is applicable to all blocks with container in style tab | 1 |
561,168 | 16,612,379,447 | IssuesEvent | 2021-06-02 13:06:03 | telstra/open-kilda | https://api.github.com/repos/telstra/open-kilda | closed | Re-implement Feature toggles in separate topology or Northbound. | area/arch priority/2-high | The feature toggles need to be re-designed and re-implemented as they were disabled in #1128. | 1.0 | Re-implement Feature toggles in separate topology or Northbound. - The feature toggles need to be re-designed and re-implemented as they were disabled in #1128. | priority | re implement feature toggles in separate topology or northbound the feature toggles need to be re designed and re implemented as they were disabled in | 1 |
364,380 | 10,763,452,701 | IssuesEvent | 2019-11-01 04:03:52 | ppy/osu | https://api.github.com/repos/ppy/osu | closed | osu!mania doesn't register misses at high scroll speeds | gameplay high priority osu!mania osu!taiko replay | Dunno if already reported but most Osu!mania Notes are getting ignored when missing them with scrollspeed 50 , sliders are working.
Video:
https://youtu.be/vP2gY7rSDlk
in the other diff of this track are some single notes that are still registering.
osu!lazer version: 2019.919.00
Logs if needed:
[Logs.zip](https://github.com/ppy/osu/files/3650602/Logs.zip)
| 1.0 | osu!mania doesn't register misses at high scroll speeds - Dunno if already reported but most Osu!mania Notes are getting ignored when missing them with scrollspeed 50 , sliders are working.
Video:
https://youtu.be/vP2gY7rSDlk
in the other diff of this track are some single notes that are still registering.
osu!lazer version: 2019.919.00
Logs if needed:
[Logs.zip](https://github.com/ppy/osu/files/3650602/Logs.zip)
| priority | osu mania doesn t register misses at high scroll speeds dunno if already reported but most osu mania notes are getting ignored when missing them with scrollspeed sliders are working video in the other diff of this track are some single notes that are still registering osu lazer version logs if needed | 1 |
634,750 | 20,372,180,152 | IssuesEvent | 2022-02-21 12:20:05 | Cordobro/D0020E | https://api.github.com/repos/Cordobro/D0020E | closed | Create server class | enhancement Priority High Diffculty Medium | Create server class accordingly to classdiagram and make it able to take service requests | 1.0 | Create server class - Create server class accordingly to classdiagram and make it able to take service requests | priority | create server class create server class accordingly to classdiagram and make it able to take service requests | 1 |
534,394 | 15,615,693,592 | IssuesEvent | 2021-03-19 19:36:32 | ContinualAI/avalanche | https://api.github.com/repos/ContinualAI/avalanche | closed | Generalize AvalancheDataset __getitem__ for Handling General Y Tensors | Benchmarks Feature - High Priority | We should better generalize the `AvalancheDataset` __getitem__ so that it can handle any task (regression, segmentation, detection, ...) with arbitrary y tensors (not just integer). | 1.0 | Generalize AvalancheDataset __getitem__ for Handling General Y Tensors - We should better generalize the `AvalancheDataset` __getitem__ so that it can handle any task (regression, segmentation, detection, ...) with arbitrary y tensors (not just integer). | priority | generalize avalanchedataset getitem for handling general y tensors we should better generalize the avalanchedataset getitem so that it can handle any task regression segmentation detection with arbitrary y tensors not just integer | 1 |
431,549 | 12,481,226,904 | IssuesEvent | 2020-05-29 22:01:06 | javascript-obfuscator/javascript-obfuscator | https://api.github.com/repos/javascript-obfuscator/javascript-obfuscator | closed | memory issue | bug high priority | I'm running my app in a low resource environment. After a few calls to JavaScriptObfuscator.obfuscate my nodejs app runs out of memory.
Each obfuscation increases the heap by ~10mb and never gets garbage collected.
var app = require('express')();
var JavaScriptObfuscator = require('javascript-obfuscator');
.....
app.use(async function (req, res, next) {
if (path.basename(req.url) == 'somefile.js') {
var file = fs.readFileSync('somefile.js', 'ascii');
var obfuscationResult = JavaScriptObfuscator.obfuscate(file, {
compact: true,
controlFlowFlattening: true,
controlFlowFlatteningThreshold: 0.75,
deadCodeInjection: true,
deadCodeInjectionThreshold: 0.4,
debugProtection: false,
debugProtectionInterval: false,
disableConsoleOutput: true,
identifierNamesGenerator: 'mangled',
log: false,
renameGlobals: false,
rotateStringArray: true,
selfDefending: true,
shuffleStringArray: true,
splitStrings: true,
splitStringsChunkLength: 2,
stringArray: true,
stringArrayEncoding: 'base64',
stringArrayThreshold: 0.75,
transformObjectKeys: true,
unicodeEscapeSequence: false
});
res.end(obfuscationResult.getObfuscatedCode());
}
next();
} | 1.0 | memory issue - I'm running my app in a low resource environment. After a few calls to JavaScriptObfuscator.obfuscate my nodejs app runs out of memory.
Each obfuscation increases the heap by ~10mb and never gets garbage collected.
var app = require('express')();
var JavaScriptObfuscator = require('javascript-obfuscator');
.....
app.use(async function (req, res, next) {
if (path.basename(req.url) == 'somefile.js') {
var file = fs.readFileSync('somefile.js', 'ascii');
var obfuscationResult = JavaScriptObfuscator.obfuscate(file, {
compact: true,
controlFlowFlattening: true,
controlFlowFlatteningThreshold: 0.75,
deadCodeInjection: true,
deadCodeInjectionThreshold: 0.4,
debugProtection: false,
debugProtectionInterval: false,
disableConsoleOutput: true,
identifierNamesGenerator: 'mangled',
log: false,
renameGlobals: false,
rotateStringArray: true,
selfDefending: true,
shuffleStringArray: true,
splitStrings: true,
splitStringsChunkLength: 2,
stringArray: true,
stringArrayEncoding: 'base64',
stringArrayThreshold: 0.75,
transformObjectKeys: true,
unicodeEscapeSequence: false
});
res.end(obfuscationResult.getObfuscatedCode());
}
next();
} | priority | memory issue i m running my app in a low resource environment after a few calls to javascriptobfuscator obfuscate my nodejs app runs out of memory each obfuscation increases the heap by and never gets garbage collected var app require express var javascriptobfuscator require javascript obfuscator app use async function req res next if path basename req url somefile js var file fs readfilesync somefile js ascii var obfuscationresult javascriptobfuscator obfuscate file compact true controlflowflattening true controlflowflatteningthreshold deadcodeinjection true deadcodeinjectionthreshold debugprotection false debugprotectioninterval false disableconsoleoutput true identifiernamesgenerator mangled log false renameglobals false rotatestringarray true selfdefending true shufflestringarray true splitstrings true splitstringschunklength stringarray true stringarrayencoding stringarraythreshold transformobjectkeys true unicodeescapesequence false res end obfuscationresult getobfuscatedcode next | 1 |
251,465 | 8,015,823,347 | IssuesEvent | 2018-07-25 11:21:14 | PMEAL/OpenPNM | https://api.github.com/repos/PMEAL/OpenPNM | closed | setup.py file needs work | Priority - High bug | Some of the modules are not being imported in a vanilla install, must be missing some 'models.misc' type lines | 1.0 | setup.py file needs work - Some of the modules are not being imported in a vanilla install, must be missing some 'models.misc' type lines | priority | setup py file needs work some of the modules are not being imported in a vanilla install must be missing some models misc type lines | 1 |
765,151 | 26,835,284,071 | IssuesEvent | 2023-02-02 18:59:51 | andrea-bistacchi/PZero | https://api.github.com/repos/andrea-bistacchi/PZero | closed | [bug] Opening old projects crashes | bug enhancement high priority | With the new pushes opacity and point size were added as a legend dict. Now old project crash. Adding a way to append default values of missing columns should avoid these kind of problems. | 1.0 | [bug] Opening old projects crashes - With the new pushes opacity and point size were added as a legend dict. Now old project crash. Adding a way to append default values of missing columns should avoid these kind of problems. | priority | opening old projects crashes with the new pushes opacity and point size were added as a legend dict now old project crash adding a way to append default values of missing columns should avoid these kind of problems | 1 |
76,734 | 3,491,654,082 | IssuesEvent | 2016-01-04 16:38:33 | ualbertalib/HydraNorth | https://api.github.com/repos/ualbertalib/HydraNorth | closed | Ensure OAI service uses original deposit timestamps | enhancement priority:high size:small | It is crucial that the OAI service use the timestamp of an item's original ingest into old ERA, not the timestamp for ingest into HydraNorth (see #574). This should be automatic but it must be tested.
See http://www.openarchives.org/OAI/openarchivesprotocol.html#SelectiveHarvestingandDatestamps
When a request include "from" and "until" parameters, the response must include items whose records "have changed" or "have become available" within that period. The response does not distinguish to two reasons for inclusion. It is therefore sufficient that e.g. an item coming out of embargo have its modification date changed to force its inclusion, even though harvesters wion't have seen it before.
Part of #570 | 1.0 | Ensure OAI service uses original deposit timestamps - It is crucial that the OAI service use the timestamp of an item's original ingest into old ERA, not the timestamp for ingest into HydraNorth (see #574). This should be automatic but it must be tested.
See http://www.openarchives.org/OAI/openarchivesprotocol.html#SelectiveHarvestingandDatestamps
When a request include "from" and "until" parameters, the response must include items whose records "have changed" or "have become available" within that period. The response does not distinguish to two reasons for inclusion. It is therefore sufficient that e.g. an item coming out of embargo have its modification date changed to force its inclusion, even though harvesters wion't have seen it before.
Part of #570 | priority | ensure oai service uses original deposit timestamps it is crucial that the oai service use the timestamp of an item s original ingest into old era not the timestamp for ingest into hydranorth see this should be automatic but it must be tested see when a request include from and until parameters the response must include items whose records have changed or have become available within that period the response does not distinguish to two reasons for inclusion it is therefore sufficient that e g an item coming out of embargo have its modification date changed to force its inclusion even though harvesters wion t have seen it before part of | 1 |
315,508 | 9,621,490,478 | IssuesEvent | 2019-05-14 10:46:03 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | reopened | Timeline Animation | Priority: High Task Timeline enhancement | The animation should start from the current position of the tick/range and not from the beginning of the time domain
## Acceptance criteria
- [x] If the animation range is not set, the animation should start from current time
- [x] Once the cursor is at the last domain value of the guide layer, the play button should be disabled. | 1.0 | Timeline Animation - The animation should start from the current position of the tick/range and not from the beginning of the time domain
## Acceptance criteria
- [x] If the animation range is not set, the animation should start from current time
- [x] Once the cursor is at the last domain value of the guide layer, the play button should be disabled. | priority | timeline animation the animation should start from the current position of the tick range and not from the beginning of the time domain acceptance criteria if the animation range is not set the animation should start from current time once the cursor is at the last domain value of the guide layer the play button should be disabled | 1 |
182,433 | 6,670,080,472 | IssuesEvent | 2017-10-03 21:47:23 | Bedework/bedework | https://api.github.com/repos/Bedework/bedework | opened | Networked web client | priority: high type: enhancement | The current client implementation is a direct call on the svci methods.
New implementation as a networked client using enhanced CalDAV (because it exists). For the public client it's read only and would allow the insertion of caches and moving the web front end to different servers. | 1.0 | Networked web client - The current client implementation is a direct call on the svci methods.
New implementation as a networked client using enhanced CalDAV (because it exists). For the public client it's read only and would allow the insertion of caches and moving the web front end to different servers. | priority | networked web client the current client implementation is a direct call on the svci methods new implementation as a networked client using enhanced caldav because it exists for the public client it s read only and would allow the insertion of caches and moving the web front end to different servers | 1 |
780,772 | 27,407,642,732 | IssuesEvent | 2023-03-01 08:17:08 | ita-social-projects/TeachUA | https://api.github.com/repos/ita-social-projects/TeachUA | opened | [Challenge] Flag "Додати реєстрацію" is absent from the "Додайте челендж" modal window | bug Priority: High Severity: medium | **Environment:** Win 11 OS, Google Chrome, 110.0.5481.104
**Reproducible:** always
**Preconditions**
1. Go to TeachUA website : https://speak-ukrainian.org.ua/dev/
2. Log in as a 'Адміністратор' (email: admin@gmail.com, password: admin)
3. Click on the user icon
4. In the dropdown menu, hover over the "Контент" and then "Челенджі" links
5. Click on the "Челенджі" link
**Steps to reproduce**
1. Click on the "Додати челендж" button
2. Verify that flag "Додати реєстрацію" is available.
**Actual result**
A flag "Додати реєстрацію" is absent from the modal window.

**Expected result**
A flag "Додати реєстрацію" is available and clickable.

**User story and test case links**
"User story #1418
[Test case](https://jira.softserve.academy/browse/TUA-641)"
**Labels to be added**
"Bug", Type ("UI, "Functional")
| 1.0 | [Challenge] Flag "Додати реєстрацію" is absent from the "Додайте челендж" modal window - **Environment:** Win 11 OS, Google Chrome, 110.0.5481.104
**Reproducible:** always
**Preconditions**
1. Go to TeachUA website : https://speak-ukrainian.org.ua/dev/
2. Log in as a 'Адміністратор' (email: admin@gmail.com, password: admin)
3. Click on the user icon
4. In the dropdown menu, hover over the "Контент" and then "Челенджі" links
5. Click on the "Челенджі" link
**Steps to reproduce**
1. Click on the "Додати челендж" button
2. Verify that flag "Додати реєстрацію" is available.
**Actual result**
A flag "Додати реєстрацію" is absent from the modal window.

**Expected result**
A flag "Додати реєстрацію" is available and clickable.

**User story and test case links**
"User story #1418
[Test case](https://jira.softserve.academy/browse/TUA-641)"
**Labels to be added**
"Bug", Type ("UI, "Functional")
| priority | flag додати реєстрацію is absent from the додайте челендж modal window environment win os google chrome reproducible always preconditions go to teachua website log in as a адміністратор email admin gmail com password admin click on the user icon in the dropdown menu hover over the контент and then челенджі links click on the челенджі link steps to reproduce click on the додати челендж button verify that flag додати реєстрацію is available actual result a flag додати реєстрацію is absent from the modal window expected result a flag додати реєстрацію is available and clickable user story and test case links user story labels to be added bug type ui functional | 1 |
353,667 | 10,555,819,273 | IssuesEvent | 2019-10-03 23:08:46 | freeorion/freeorion | https://api.github.com/repos/freeorion/freeorion | closed | 'Ghost' colony renders planet unusable | category:bug component:game mechanic priority:high | ### Environment
* **FreeOrion Version**: 09 Sep
* **Operating System**: Win 7
* **Graphic card used**: RadeOn 5750
* **Fetched as** <!-- delete lines below that don't apply -->
* Binary release
* Weekly development build
### Description
When my opponent conceded, his pops vanished but all his buildings remained. One of them, at an outpost, was a completed Exobot colony, which had been built but not used up. I cannot scrap it, and I cannot build an exobot colony of my own.
### Expected Result
The completed colony building should activate and create an exobot pop.
### Steps to reproduce
Conquer your enemy until he quits in between the turn that a colony completes and the turn it activates.
| 1.0 | 'Ghost' colony renders planet unusable - ### Environment
* **FreeOrion Version**: 09 Sep
* **Operating System**: Win 7
* **Graphic card used**: RadeOn 5750
* **Fetched as** <!-- delete lines below that don't apply -->
* Binary release
* Weekly development build
### Description
When my opponent conceded, his pops vanished but all his buildings remained. One of them, at an outpost, was a completed Exobot colony, which had been built but not used up. I cannot scrap it, and I cannot build an exobot colony of my own.
### Expected Result
The completed colony building should activate and create an exobot pop.
### Steps to reproduce
Conquer your enemy until he quits in between the turn that a colony completes and the turn it activates.
| priority | ghost colony renders planet unusable environment freeorion version sep operating system win graphic card used radeon fetched as binary release weekly development build description when my opponent conceded his pops vanished but all his buildings remained one of them at an outpost was a completed exobot colony which had been built but not used up i cannot scrap it and i cannot build an exobot colony of my own expected result the completed colony building should activate and create an exobot pop steps to reproduce conquer your enemy until he quits in between the turn that a colony completes and the turn it activates | 1 |
34,356 | 2,777,154,920 | IssuesEvent | 2015-05-05 05:00:47 | punongbayan-araullo/tickets | https://api.github.com/repos/punongbayan-araullo/tickets | opened | Show risk factor of a company upon creation of new pursuit | priority - high status - accepted system - pursuits type - enhancement | As discussed please see information that should be shown as a widget in the new pursuits page (should appear after a company is selected from the company list):
- Most recent company risk assessment and date of that risk assessment
- Partners with recent jobs ongoing/completed with the company (within 2 years)
- Outstanding billings if any and how long outstanding (if multiple bills across multiple divisions, then just summarize and provide link to company data)
- Average collection period (for all billings / services)
- Any bills written-off (not reversed -- must be actual write-offs) | 1.0 | Show risk factor of a company upon creation of new pursuit - As discussed please see information that should be shown as a widget in the new pursuits page (should appear after a company is selected from the company list):
- Most recent company risk assessment and date of that risk assessment
- Partners with recent jobs ongoing/completed with the company (within 2 years)
- Outstanding billings if any and how long outstanding (if multiple bills across multiple divisions, then just summarize and provide link to company data)
- Average collection period (for all billings / services)
- Any bills written-off (not reversed -- must be actual write-offs) | priority | show risk factor of a company upon creation of new pursuit as discussed please see information that should be shown as a widget in the new pursuits page should appear after a company is selected from the company list most recent company risk assessment and date of that risk assessment partners with recent jobs ongoing completed with the company within years outstanding billings if any and how long outstanding if multiple bills across multiple divisions then just summarize and provide link to company data average collection period for all billings services any bills written off not reversed must be actual write offs | 1 |
405,016 | 11,866,047,264 | IssuesEvent | 2020-03-26 02:25:20 | earthlab/earthpy | https://api.github.com/repos/earthlab/earthpy | closed | DEPRECATE clip | high-priority | Clip is now in geopandas. so we can deprecate it from earthpy and remove all tests.
this should allow us to completely drop rtree as a dependency which causes significant pip install issues see #507 | 1.0 | DEPRECATE clip - Clip is now in geopandas. so we can deprecate it from earthpy and remove all tests.
this should allow us to completely drop rtree as a dependency which causes significant pip install issues see #507 | priority | deprecate clip clip is now in geopandas so we can deprecate it from earthpy and remove all tests this should allow us to completely drop rtree as a dependency which causes significant pip install issues see | 1 |
813,900 | 30,478,502,092 | IssuesEvent | 2023-07-17 18:22:55 | everyday-as/gmodstore-issues | https://api.github.com/repos/everyday-as/gmodstore-issues | closed | Refund/revoke reasons | 🌟 Feature High Priority | ## Describe a feature you would like to have added to GmodStore
We have recently noticed that a select few creators instantly refund customers and close their support ticket, without saying why. This has caused many customers to come to us and say that they did not want a refund, and they're now worried their issue will never be fixed and they won't get future updates.
## Proposed solution
**Solution 1**
In order to improve the customer experience I'm proposing a solution where refunds would now require a valid reason. There would be a dropdown to select a reason, with the following selectable reasons:
- "The product does not work on the customer's server" -> must include a link to a message, image or a support ticket
- "The customer requested a refund" -> must include a link to a message, image or a support ticket
- "The customer has not downloaded the product yet" -> we will automatically check if the customer has downloaded the product and process the refund immediately if they have not
- "Other" -> must provide a reason in free text of a minimum of x amount of characters
~~In all cases, except in the case where the user has not downloaded the addon the refund request will be flagged for review by a moderator/admin.~~ (**edit after feedback:** only flagged creators should be subject to a review)
**Solution 2**
N/A. I had another idea but I'm still trying to map it.
| 1.0 | Refund/revoke reasons - ## Describe a feature you would like to have added to GmodStore
We have recently noticed that a select few creators instantly refund customers and close their support ticket, without saying why. This has caused many customers to come to us and say that they did not want a refund, and they're now worried their issue will never be fixed and they won't get future updates.
## Proposed solution
**Solution 1**
In order to improve the customer experience I'm proposing a solution where refunds would now require a valid reason. There would be a dropdown to select a reason, with the following selectable reasons:
- "The product does not work on the customer's server" -> must include a link to a message, image or a support ticket
- "The customer requested a refund" -> must include a link to a message, image or a support ticket
- "The customer has not downloaded the product yet" -> we will automatically check if the customer has downloaded the product and process the refund immediately if they have not
- "Other" -> must provide a reason in free text of a minimum of x amount of characters
~~In all cases, except in the case where the user has not downloaded the addon the refund request will be flagged for review by a moderator/admin.~~ (**edit after feedback:** only flagged creators should be subject to a review)
**Solution 2**
N/A. I had another idea but I'm still trying to map it.
| priority | refund revoke reasons describe a feature you would like to have added to gmodstore we have recently noticed that a select few creators instantly refund customers and close their support ticket without saying why this has caused many customers to come to us and say that they did not want a refund and they re now worried their issue will never be fixed and they won t get future updates proposed solution solution in order to improve the customer experience i m proposing a solution where refunds would now require a valid reason there would be a dropdown to select a reason with the following selectable reasons the product does not work on the customer s server must include a link to a message image or a support ticket the customer requested a refund must include a link to a message image or a support ticket the customer has not downloaded the product yet we will automatically check if the customer has downloaded the product and process the refund immediately if they have not other must provide a reason in free text of a minimum of x amount of characters in all cases except in the case where the user has not downloaded the addon the refund request will be flagged for review by a moderator admin edit after feedback only flagged creators should be subject to a review solution n a i had another idea but i m still trying to map it | 1 |
306,507 | 9,395,823,635 | IssuesEvent | 2019-04-08 04:40:31 | metasfresh/metasfresh-webui-frontend | https://api.github.com/repos/metasfresh/metasfresh-webui-frontend | closed | input field value sometimes swallowed | branch:master branch:release priority:high topic:Cypress type:bug | ### Is this a bug or feature request?
bug
### What is the current behavior?
#### Which are the steps to reproduce?
### What is the expected or desired behavior?
| 1.0 | input field value sometimes swallowed - ### Is this a bug or feature request?
bug
### What is the current behavior?
#### Which are the steps to reproduce?
### What is the expected or desired behavior?
| priority | input field value sometimes swallowed is this a bug or feature request bug what is the current behavior which are the steps to reproduce what is the expected or desired behavior | 1 |
307,869 | 9,423,375,643 | IssuesEvent | 2019-04-11 11:43:46 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.8.1.2] Lavish Talent wrong calculation | High Priority Medium Priority Quality Assurance | [0.8.1.2]
no mods
related to #11330
Lavish Talent doesn`t work
**Steps**
1. Take tailoring
2. Place tailoting table in the room tier 1
3, Check resources requirement
4. Level up to 7
5. Take talant Lavish
**Expected result**
- tier requirement 1
- resources recuirement decrease (if level 1 cloth has 7 fibers - 50% decrease, after leveling to 7 it has 3 - 80% decrease, and there is no additional 10% from talent) )
**Actual result**
No bonuses from talent (and to small bounce from level)
before leveling



After leveling



| 2.0 | [0.8.1.2] Lavish Talent wrong calculation - [0.8.1.2]
no mods
related to #11330
Lavish Talent doesn`t work
**Steps**
1. Take tailoring
2. Place tailoting table in the room tier 1
3, Check resources requirement
4. Level up to 7
5. Take talant Lavish
**Expected result**
- tier requirement 1
- resources recuirement decrease (if level 1 cloth has 7 fibers - 50% decrease, after leveling to 7 it has 3 - 80% decrease, and there is no additional 10% from talent) )
**Actual result**
No bonuses from talent (and to small bounce from level)
before leveling



After leveling



| priority | lavish talent wrong calculation no mods related to lavish talent doesn t work steps take tailoring place tailoting table in the room tier check resources requirement level up to take talant lavish expected result tier requirement resources recuirement decrease if level cloth has fibers decrease after leveling to it has decrease and there is no additional from talent actual result no bonuses from talent and to small bounce from level before leveling after leveling | 1 |
736,977 | 25,495,164,042 | IssuesEvent | 2022-11-27 15:34:17 | bigbio/quantms | https://api.github.com/repos/bigbio/quantms | closed | error pmultiqc with lfq dataset | bug high-priority | ### Description of the bug
URL of the working folder: http://ftp.pride.ebi.ac.uk/pub/databases/pride/resources/proteomes/proteogenomics/benchmakrs/21fc23ff8c4e26d68bd614cfaa9d39/
```
nf-core/quantms execution completed unsuccessfully!
The exit status of the task that caused the workflow execution to fail was: 1.
The full error message was:
Error executing process > 'NFCORE_QUANTMS:QUANTMS:SUMMARYPIPELINE (1)'
Caused by:
Process `NFCORE_QUANTMS:QUANTMS:SUMMARYPIPELINE (1)` terminated with an error exit status (1)
Command executed:
multiqc \
-f \
--config ./results/multiqc_config.yml \
\
\
\
--quantification_method feature_intensity \
./results \
-o .
cat <<-END_VERSIONS > versions.yml
"NFCORE_QUANTMS:QUANTMS:SUMMARYPIPELINE":
pmultiqc: $(multiqc --pmultiqc_version | sed -e "s/pmultiqc, version //g")
END_VERSIONS
Command exit status:
1
Command output:
No decoy peptides found -> only showing target peptides
Command error:
KeyError: 0
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/site-packages/multiqc/multiqc.py", line 654, in run
output = mod()
File "/usr/local/lib/python3.10/site-packages/pmultiqc/modules/quantms/quantms.py", line 173, in __init__
self.CalHeatMapScore()
File "/usr/local/lib/python3.10/site-packages/pmultiqc/modules/quantms/quantms.py", line 930, in CalHeatMapScore
self.MissedCleavages_heatmap_score[name] = sc[0] / sc[:].sum()
File "/usr/local/lib/python3.10/site-packages/pandas/core/series.py", line 981, in __getitem__
return self._get_value(key)
File "/usr/local/lib/python3.10/site-packages/pandas/core/series.py", line 1089, in _get_value
loc = self.index.get_loc(label)
File "/usr/local/lib/python3.10/site-packages/pandas/core/indexes/base.py", line 3805, in get_loc
raise KeyError(key) from err
KeyError: 0
================================================================================
| custom_content | nf-core-quantms-methods-description: Found 1 sample (html)
INFO:multiqc.modules.custom_content.custom_content:nf-core-quantms-methods-description: Found 1 sample (html)
| custom_content | software_versions: Found 1 sample (html)
INFO:multiqc.modules.custom_content.custom_content:software_versions: Found 1 sample (html)
| custom_content | nf-core-quantms-summary: Found 1 sample (html)
INFO:multiqc.modules.custom_content.custom_content:nf-core-quantms-summary: Found 1 sample (html)
DEBUG:multiqc:Reordering sections: anchor 'pmultiqc' not found.
DEBUG:multiqc:Reordering sections: anchor 'pmultiqc' not found for module 'nf-core/quantms Methods Description'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-methods-description' not found for module 'nf-core/quantms Methods Description'.
DEBUG:multiqc:Reordering sections: anchor 'software_versions' not found for module 'nf-core/quantms Methods Description'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-summary' not found for module 'nf-core/quantms Methods Description'.
DEBUG:multiqc:Reordering sections: anchor 'pmultiqc' not found for module 'nf-core/quantms Software Versions'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-methods-description' not found for module 'nf-core/quantms Software Versions'.
DEBUG:multiqc:Reordering sections: anchor 'software_versions' not found for module 'nf-core/quantms Software Versions'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-summary' not found for module 'nf-core/quantms Software Versions'.
DEBUG:multiqc:Reordering sections: anchor 'pmultiqc' not found for module 'nf-core/quantms Workflow Summary'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-methods-description' not found for module 'nf-core/quantms Workflow Summary'.
DEBUG:multiqc:Reordering sections: anchor 'software_versions' not found for module 'nf-core/quantms Workflow Summary'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-summary' not found for module 'nf-core/quantms Workflow Summary'.
| multiqc | Compressing plot data
INFO:multiqc:Compressing plot data
| multiqc | Report : multiqc_report.html
INFO:multiqc:Report : multiqc_report.html
| multiqc | Data : multiqc_data
INFO:multiqc:Data : multiqc_data
DEBUG:multiqc:Moving data file from '/hps/scratch/lsf_tmpdir/hl-codon-28-03/tmp7umnqaqa/multiqc_data' to './multiqc_data'
| multiqc | Plots : multiqc_plots
INFO:multiqc:Plots : multiqc_plots
DEBUG:multiqc:Moving plots directory from '/hps/scratch/lsf_tmpdir/hl-codon-28-03/tmp7umnqaqa/multiqc_plots' to './multiqc_plots'
| multiqc | MultiQC complete
INFO:multiqc:MultiQC complete
Work dir:
/hps/nobackup/juan/pride/reanalysis/differential-expression/lfq/PXD002395/work/99/21fc23ff8c4e26d68bd614cfaa9d39
Tip: when you have fixed the problem you can continue the execution adding the option `-resume` to the run command line
```
### Command used and terminal output
_No response_
### Relevant files
_No response_
### System information
_No response_ | 1.0 | error pmultiqc with lfq dataset - ### Description of the bug
URL of the working folder: http://ftp.pride.ebi.ac.uk/pub/databases/pride/resources/proteomes/proteogenomics/benchmakrs/21fc23ff8c4e26d68bd614cfaa9d39/
```
nf-core/quantms execution completed unsuccessfully!
The exit status of the task that caused the workflow execution to fail was: 1.
The full error message was:
Error executing process > 'NFCORE_QUANTMS:QUANTMS:SUMMARYPIPELINE (1)'
Caused by:
Process `NFCORE_QUANTMS:QUANTMS:SUMMARYPIPELINE (1)` terminated with an error exit status (1)
Command executed:
multiqc \
-f \
--config ./results/multiqc_config.yml \
\
\
\
--quantification_method feature_intensity \
./results \
-o .
cat <<-END_VERSIONS > versions.yml
"NFCORE_QUANTMS:QUANTMS:SUMMARYPIPELINE":
pmultiqc: $(multiqc --pmultiqc_version | sed -e "s/pmultiqc, version //g")
END_VERSIONS
Command exit status:
1
Command output:
No decoy peptides found -> only showing target peptides
Command error:
KeyError: 0
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/site-packages/multiqc/multiqc.py", line 654, in run
output = mod()
File "/usr/local/lib/python3.10/site-packages/pmultiqc/modules/quantms/quantms.py", line 173, in __init__
self.CalHeatMapScore()
File "/usr/local/lib/python3.10/site-packages/pmultiqc/modules/quantms/quantms.py", line 930, in CalHeatMapScore
self.MissedCleavages_heatmap_score[name] = sc[0] / sc[:].sum()
File "/usr/local/lib/python3.10/site-packages/pandas/core/series.py", line 981, in __getitem__
return self._get_value(key)
File "/usr/local/lib/python3.10/site-packages/pandas/core/series.py", line 1089, in _get_value
loc = self.index.get_loc(label)
File "/usr/local/lib/python3.10/site-packages/pandas/core/indexes/base.py", line 3805, in get_loc
raise KeyError(key) from err
KeyError: 0
================================================================================
| custom_content | nf-core-quantms-methods-description: Found 1 sample (html)
INFO:multiqc.modules.custom_content.custom_content:nf-core-quantms-methods-description: Found 1 sample (html)
| custom_content | software_versions: Found 1 sample (html)
INFO:multiqc.modules.custom_content.custom_content:software_versions: Found 1 sample (html)
| custom_content | nf-core-quantms-summary: Found 1 sample (html)
INFO:multiqc.modules.custom_content.custom_content:nf-core-quantms-summary: Found 1 sample (html)
DEBUG:multiqc:Reordering sections: anchor 'pmultiqc' not found.
DEBUG:multiqc:Reordering sections: anchor 'pmultiqc' not found for module 'nf-core/quantms Methods Description'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-methods-description' not found for module 'nf-core/quantms Methods Description'.
DEBUG:multiqc:Reordering sections: anchor 'software_versions' not found for module 'nf-core/quantms Methods Description'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-summary' not found for module 'nf-core/quantms Methods Description'.
DEBUG:multiqc:Reordering sections: anchor 'pmultiqc' not found for module 'nf-core/quantms Software Versions'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-methods-description' not found for module 'nf-core/quantms Software Versions'.
DEBUG:multiqc:Reordering sections: anchor 'software_versions' not found for module 'nf-core/quantms Software Versions'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-summary' not found for module 'nf-core/quantms Software Versions'.
DEBUG:multiqc:Reordering sections: anchor 'pmultiqc' not found for module 'nf-core/quantms Workflow Summary'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-methods-description' not found for module 'nf-core/quantms Workflow Summary'.
DEBUG:multiqc:Reordering sections: anchor 'software_versions' not found for module 'nf-core/quantms Workflow Summary'.
DEBUG:multiqc:Reordering sections: anchor 'nf-core-quantms-summary' not found for module 'nf-core/quantms Workflow Summary'.
| multiqc | Compressing plot data
INFO:multiqc:Compressing plot data
| multiqc | Report : multiqc_report.html
INFO:multiqc:Report : multiqc_report.html
| multiqc | Data : multiqc_data
INFO:multiqc:Data : multiqc_data
DEBUG:multiqc:Moving data file from '/hps/scratch/lsf_tmpdir/hl-codon-28-03/tmp7umnqaqa/multiqc_data' to './multiqc_data'
| multiqc | Plots : multiqc_plots
INFO:multiqc:Plots : multiqc_plots
DEBUG:multiqc:Moving plots directory from '/hps/scratch/lsf_tmpdir/hl-codon-28-03/tmp7umnqaqa/multiqc_plots' to './multiqc_plots'
| multiqc | MultiQC complete
INFO:multiqc:MultiQC complete
Work dir:
/hps/nobackup/juan/pride/reanalysis/differential-expression/lfq/PXD002395/work/99/21fc23ff8c4e26d68bd614cfaa9d39
Tip: when you have fixed the problem you can continue the execution adding the option `-resume` to the run command line
```
### Command used and terminal output
_No response_
### Relevant files
_No response_
### System information
_No response_ | priority | error pmultiqc with lfq dataset description of the bug url of the working folder nf core quantms execution completed unsuccessfully the exit status of the task that caused the workflow execution to fail was the full error message was error executing process nfcore quantms quantms summarypipeline caused by process nfcore quantms quantms summarypipeline terminated with an error exit status command executed multiqc f config results multiqc config yml quantification method feature intensity results o cat versions yml nfcore quantms quantms summarypipeline pmultiqc multiqc pmultiqc version sed e s pmultiqc version g end versions command exit status command output no decoy peptides found only showing target peptides command error keyerror the above exception was the direct cause of the following exception traceback most recent call last file usr local lib site packages multiqc multiqc py line in run output mod file usr local lib site packages pmultiqc modules quantms quantms py line in init self calheatmapscore file usr local lib site packages pmultiqc modules quantms quantms py line in calheatmapscore self missedcleavages heatmap score sc sc sum file usr local lib site packages pandas core series py line in getitem return self get value key file usr local lib site packages pandas core series py line in get value loc self index get loc label file usr local lib site packages pandas core indexes base py line in get loc raise keyerror key from err keyerror custom content nf core quantms methods description found sample html info multiqc modules custom content custom content nf core quantms methods description found sample html custom content software versions found sample html info multiqc modules custom content custom content software versions found sample html custom content nf core quantms summary found sample html info multiqc modules custom content custom content nf core quantms summary found sample html debug multiqc reordering sections anchor pmultiqc not found debug multiqc reordering sections anchor pmultiqc not found for module nf core quantms methods description debug multiqc reordering sections anchor nf core quantms methods description not found for module nf core quantms methods description debug multiqc reordering sections anchor software versions not found for module nf core quantms methods description debug multiqc reordering sections anchor nf core quantms summary not found for module nf core quantms methods description debug multiqc reordering sections anchor pmultiqc not found for module nf core quantms software versions debug multiqc reordering sections anchor nf core quantms methods description not found for module nf core quantms software versions debug multiqc reordering sections anchor software versions not found for module nf core quantms software versions debug multiqc reordering sections anchor nf core quantms summary not found for module nf core quantms software versions debug multiqc reordering sections anchor pmultiqc not found for module nf core quantms workflow summary debug multiqc reordering sections anchor nf core quantms methods description not found for module nf core quantms workflow summary debug multiqc reordering sections anchor software versions not found for module nf core quantms workflow summary debug multiqc reordering sections anchor nf core quantms summary not found for module nf core quantms workflow summary multiqc compressing plot data info multiqc compressing plot data multiqc report multiqc report html info multiqc report multiqc report html multiqc data multiqc data info multiqc data multiqc data debug multiqc moving data file from hps scratch lsf tmpdir hl codon multiqc data to multiqc data multiqc plots multiqc plots info multiqc plots multiqc plots debug multiqc moving plots directory from hps scratch lsf tmpdir hl codon multiqc plots to multiqc plots multiqc multiqc complete info multiqc multiqc complete work dir hps nobackup juan pride reanalysis differential expression lfq work tip when you have fixed the problem you can continue the execution adding the option resume to the run command line command used and terminal output no response relevant files no response system information no response | 1 |
326,655 | 9,958,975,804 | IssuesEvent | 2019-07-06 01:14:06 | default51400/familynet | https://api.github.com/repos/default51400/familynet | opened | Create database using EF | Database high priority question task | Need to write all models with their data annotation (for example, reqiured, etc.). Create a database with the entire context, use the template repository (you need to clarify what Slava meant) | 1.0 | Create database using EF - Need to write all models with their data annotation (for example, reqiured, etc.). Create a database with the entire context, use the template repository (you need to clarify what Slava meant) | priority | create database using ef need to write all models with their data annotation for example reqiured etc create a database with the entire context use the template repository you need to clarify what slava meant | 1 |
592,652 | 17,926,655,476 | IssuesEvent | 2021-09-10 01:04:16 | IsaacCheng9/student-network | https://api.github.com/repos/IsaacCheng9/student-network | opened | Fix broken tests caused by relative imports | bug high priority | Tests are currently broken - this is largely caused by relative imports not working. | 1.0 | Fix broken tests caused by relative imports - Tests are currently broken - this is largely caused by relative imports not working. | priority | fix broken tests caused by relative imports tests are currently broken this is largely caused by relative imports not working | 1 |
229,831 | 7,595,700,834 | IssuesEvent | 2018-04-27 06:54:58 | braun-robotics/rust-lpc82x-hal | https://api.github.com/repos/braun-robotics/rust-lpc82x-hal | closed | Release on crates.io | priority: high type: enhancement | This project can't currently be released on crates.io, as two of its dependencies aren't:
- https://github.com/japaric/nb
- https://github.com/japaric/embedded-hal
Once those two are available, we should release a crates.io version ourselves.
Things to remember when releasing:
- Add additional badges to README
- Update links to documentation in README and Cargo.toml | 1.0 | Release on crates.io - This project can't currently be released on crates.io, as two of its dependencies aren't:
- https://github.com/japaric/nb
- https://github.com/japaric/embedded-hal
Once those two are available, we should release a crates.io version ourselves.
Things to remember when releasing:
- Add additional badges to README
- Update links to documentation in README and Cargo.toml | priority | release on crates io this project can t currently be released on crates io as two of its dependencies aren t once those two are available we should release a crates io version ourselves things to remember when releasing add additional badges to readme update links to documentation in readme and cargo toml | 1 |
606,852 | 18,769,323,788 | IssuesEvent | 2021-11-06 14:41:29 | CDCgov/prime-reportstream | https://api.github.com/repos/CDCgov/prime-reportstream | opened | Reduce errors and improve sender and receiver visibility for fields currently used and populated by the LivdMapper() | bug Epic high-priority data-issue | There have been multiple issues created since May, 2021 related to fields used and populated by the LivdMapper(). This Epic will consolidate those issues and create new user stories and acceptance criteria for the issues to be addressed.
**Problem Summary:**
If a sender does not send in the _exact_ Model_ID or Test_Kit_Name_ID, multiple fields will fail to populate in the outbound file to the STLT. Importantly, even if the Test_Performed_LOINC_CODE is supplied by the sender, this value will be overwritten with a blank in the outbound file to the STLT if the Model_ID or Test_Kit_Name_ID is incorrect (i.e. the sender missed an asterisk in the string name). The Test_Performed_LOINC_Code is a critical value for STLTs, as many use that value for routing records from their Rhapsody integration engine to the correct disease surveillance system or module within that STLT. This issue appears to have affected over 266,000 records received between September 19th and November 5, 2021. STLTs have reported ongoing issues with not receiving LOINC codes, and is a driver of ongoing tickets received by the Onboarding and Operations Team.
Based on the attached analysis, there appears to be some simple fixes that will address the majority of the records which are failing to populate a LOINC code. Specifically, 99.75% of the records that had an incorrect Model_ID or Test_Kit_Name_ID between 9/19 and 11/5 we only off by an asterisk (the sender added an asterisk unnecessarily, or didn't include an asterisk at the end of the test name when it should have).
[https://app.zenhub.com/files/304423150/33c3d86b-8e8c-41bb-8384-5439a97db74f/download] | 1.0 | Reduce errors and improve sender and receiver visibility for fields currently used and populated by the LivdMapper() - There have been multiple issues created since May, 2021 related to fields used and populated by the LivdMapper(). This Epic will consolidate those issues and create new user stories and acceptance criteria for the issues to be addressed.
**Problem Summary:**
If a sender does not send in the _exact_ Model_ID or Test_Kit_Name_ID, multiple fields will fail to populate in the outbound file to the STLT. Importantly, even if the Test_Performed_LOINC_CODE is supplied by the sender, this value will be overwritten with a blank in the outbound file to the STLT if the Model_ID or Test_Kit_Name_ID is incorrect (i.e. the sender missed an asterisk in the string name). The Test_Performed_LOINC_Code is a critical value for STLTs, as many use that value for routing records from their Rhapsody integration engine to the correct disease surveillance system or module within that STLT. This issue appears to have affected over 266,000 records received between September 19th and November 5, 2021. STLTs have reported ongoing issues with not receiving LOINC codes, and is a driver of ongoing tickets received by the Onboarding and Operations Team.
Based on the attached analysis, there appears to be some simple fixes that will address the majority of the records which are failing to populate a LOINC code. Specifically, 99.75% of the records that had an incorrect Model_ID or Test_Kit_Name_ID between 9/19 and 11/5 we only off by an asterisk (the sender added an asterisk unnecessarily, or didn't include an asterisk at the end of the test name when it should have).
[https://app.zenhub.com/files/304423150/33c3d86b-8e8c-41bb-8384-5439a97db74f/download] | priority | reduce errors and improve sender and receiver visibility for fields currently used and populated by the livdmapper there have been multiple issues created since may related to fields used and populated by the livdmapper this epic will consolidate those issues and create new user stories and acceptance criteria for the issues to be addressed problem summary if a sender does not send in the exact model id or test kit name id multiple fields will fail to populate in the outbound file to the stlt importantly even if the test performed loinc code is supplied by the sender this value will be overwritten with a blank in the outbound file to the stlt if the model id or test kit name id is incorrect i e the sender missed an asterisk in the string name the test performed loinc code is a critical value for stlts as many use that value for routing records from their rhapsody integration engine to the correct disease surveillance system or module within that stlt this issue appears to have affected over records received between september and november stlts have reported ongoing issues with not receiving loinc codes and is a driver of ongoing tickets received by the onboarding and operations team based on the attached analysis there appears to be some simple fixes that will address the majority of the records which are failing to populate a loinc code specifically of the records that had an incorrect model id or test kit name id between and we only off by an asterisk the sender added an asterisk unnecessarily or didn t include an asterisk at the end of the test name when it should have | 1 |
7,728 | 2,603,751,501 | IssuesEvent | 2015-02-24 17:44:15 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | closed | OnMorph incomplete and GetBuildType inconsistency | auto-migrated Component-Logic Priority-High Type-Enhancement Usability | ```
2 problems with buildings "morphing" to other buildings, such as a creep
colony transforming into a sunken.
First, there's an inconsistancy with onMorph. onMorph will usually get
called when the building starts for zergs (e.g., when I start a spawning
pool). However, it will get called when the building FINISHES if it was
morphed from an already existing building (so it will get called when the
sunken is finished for instance). On a side note, I'm not sure how it
works with other races, but shouldn't an event be launched both when a
building starts AND when a building finishes? Well... even if it was a
design decision to have an event launched only for one of those, it would
be nice if it could be consistent.
The second problem is that I have been unable to find how to get what
building it's morphing to. If you call the "getType()" method on a creep
colony that is morphing into a sunken colony, it will
return "Zerg_Creep_Colony". If you then call "getBuildType()", it will
return "None". Maybe there's a method I've overlooked, but I would expect
at least one of those, if not both would return "Zerg_Sunken_Colony".
```
-----
Original issue reported on code.google.com by `PhilBell...@gmail.com` on 25 Mar 2010 at 2:08 | 1.0 | OnMorph incomplete and GetBuildType inconsistency - ```
2 problems with buildings "morphing" to other buildings, such as a creep
colony transforming into a sunken.
First, there's an inconsistancy with onMorph. onMorph will usually get
called when the building starts for zergs (e.g., when I start a spawning
pool). However, it will get called when the building FINISHES if it was
morphed from an already existing building (so it will get called when the
sunken is finished for instance). On a side note, I'm not sure how it
works with other races, but shouldn't an event be launched both when a
building starts AND when a building finishes? Well... even if it was a
design decision to have an event launched only for one of those, it would
be nice if it could be consistent.
The second problem is that I have been unable to find how to get what
building it's morphing to. If you call the "getType()" method on a creep
colony that is morphing into a sunken colony, it will
return "Zerg_Creep_Colony". If you then call "getBuildType()", it will
return "None". Maybe there's a method I've overlooked, but I would expect
at least one of those, if not both would return "Zerg_Sunken_Colony".
```
-----
Original issue reported on code.google.com by `PhilBell...@gmail.com` on 25 Mar 2010 at 2:08 | priority | onmorph incomplete and getbuildtype inconsistency problems with buildings morphing to other buildings such as a creep colony transforming into a sunken first there s an inconsistancy with onmorph onmorph will usually get called when the building starts for zergs e g when i start a spawning pool however it will get called when the building finishes if it was morphed from an already existing building so it will get called when the sunken is finished for instance on a side note i m not sure how it works with other races but shouldn t an event be launched both when a building starts and when a building finishes well even if it was a design decision to have an event launched only for one of those it would be nice if it could be consistent the second problem is that i have been unable to find how to get what building it s morphing to if you call the gettype method on a creep colony that is morphing into a sunken colony it will return zerg creep colony if you then call getbuildtype it will return none maybe there s a method i ve overlooked but i would expect at least one of those if not both would return zerg sunken colony original issue reported on code google com by philbell gmail com on mar at | 1 |
350,457 | 10,490,645,012 | IssuesEvent | 2019-09-25 09:25:22 | firecracker-microvm/firecracker | https://api.github.com/repos/firecracker-microvm/firecracker | closed | Decouple logger logic from the flush destination | Priority: High Quality: Improvement | Decouple the destination for flushing logs and metrics (currently a FIFO) from the logger crate.
| 1.0 | Decouple logger logic from the flush destination - Decouple the destination for flushing logs and metrics (currently a FIFO) from the logger crate.
| priority | decouple logger logic from the flush destination decouple the destination for flushing logs and metrics currently a fifo from the logger crate | 1 |
411,439 | 12,018,061,229 | IssuesEvent | 2020-04-10 19:53:29 | cdnjs/cdnjs | https://api.github.com/repos/cdnjs/cdnjs | closed | [Request] Add videojs-vimeo | :arrows_counterclockwise: Wait for response :rotating_light: High Priority 🏷 Library Request | **Library name:** videojs-vimeo
**Git repository url:** https://github.com/videojs/videojs-vimeo
**npm package url(optional):** https://www.npmjs.com/package/videojs-vimeo
**License(s):** MIT
**Official homepage:** http://videojs.com/
Thanks :)
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/39032573-request-add-videojs-vimeo?utm_campaign=plugin&utm_content=tracker%2F32893&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F32893&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | [Request] Add videojs-vimeo - **Library name:** videojs-vimeo
**Git repository url:** https://github.com/videojs/videojs-vimeo
**npm package url(optional):** https://www.npmjs.com/package/videojs-vimeo
**License(s):** MIT
**Official homepage:** http://videojs.com/
Thanks :)
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/39032573-request-add-videojs-vimeo?utm_campaign=plugin&utm_content=tracker%2F32893&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F32893&utm_medium=issues&utm_source=github).
</bountysource-plugin> | priority | add videojs vimeo library name videojs vimeo git repository url npm package url optional license s mit official homepage thanks want to back this issue we accept bounties via | 1 |
620,503 | 19,563,652,744 | IssuesEvent | 2022-01-03 20:02:55 | bounswe/2021SpringGroup10 | https://api.github.com/repos/bounswe/2021SpringGroup10 | closed | AWS: Transform community_page endpoints to API Gateway | Type: Enhancement Priority: High Coding: Backend Platform: AWS | Newly produced endpoints should be transformed to the API Gateway to disable CORS error. The mentioned endpoints are:
* api/community_page/admin
* api/community_page/ban
* api/community_page/unban
* api/community_page/change_privacy
* api/community_page/request
* api/community_page/subscribe
* api/community_page/unsubscribe | 1.0 | AWS: Transform community_page endpoints to API Gateway - Newly produced endpoints should be transformed to the API Gateway to disable CORS error. The mentioned endpoints are:
* api/community_page/admin
* api/community_page/ban
* api/community_page/unban
* api/community_page/change_privacy
* api/community_page/request
* api/community_page/subscribe
* api/community_page/unsubscribe | priority | aws transform community page endpoints to api gateway newly produced endpoints should be transformed to the api gateway to disable cors error the mentioned endpoints are api community page admin api community page ban api community page unban api community page change privacy api community page request api community page subscribe api community page unsubscribe | 1 |
79,016 | 3,520,015,014 | IssuesEvent | 2016-01-12 19:05:17 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | Email: Create account email says welcome to Dataverse 4.0 Beta, should remove beta and maybe 4.0? | Component: Migration Component: UX & Upgrade Priority: High Status: QA Type: Bug | @eaquigley
Just created an account and got a nice welcome email. Saw it said welcome to Dataverse 4.0 Beta. Since we are close to releasing, we should remove beta and depending on how it gets 4.0, we may consider removing that. If automatically populated, then ok. | 1.0 | Email: Create account email says welcome to Dataverse 4.0 Beta, should remove beta and maybe 4.0? - @eaquigley
Just created an account and got a nice welcome email. Saw it said welcome to Dataverse 4.0 Beta. Since we are close to releasing, we should remove beta and depending on how it gets 4.0, we may consider removing that. If automatically populated, then ok. | priority | email create account email says welcome to dataverse beta should remove beta and maybe eaquigley just created an account and got a nice welcome email saw it said welcome to dataverse beta since we are close to releasing we should remove beta and depending on how it gets we may consider removing that if automatically populated then ok | 1 |
483,714 | 13,929,367,070 | IssuesEvent | 2020-10-21 23:29:38 | infiniteautomation/ma-core-public | https://api.github.com/repos/infiniteautomation/ma-core-public | closed | EventDao.returnEventsToNormal() | High Priority Item | This method should use a prepared statement and chunk into good size queries. Look up these limits in H2 and MySQL. | 1.0 | EventDao.returnEventsToNormal() - This method should use a prepared statement and chunk into good size queries. Look up these limits in H2 and MySQL. | priority | eventdao returneventstonormal this method should use a prepared statement and chunk into good size queries look up these limits in and mysql | 1 |
276,563 | 8,600,431,512 | IssuesEvent | 2018-11-16 07:34:03 | whereToApplication/whereTo | https://api.github.com/repos/whereToApplication/whereTo | closed | /feedback route not posting data to backend | Backend bug high priority | Feedback route not posting data correctly to backend. object is being created but the feedback is not being stored there.
| 1.0 | /feedback route not posting data to backend - Feedback route not posting data correctly to backend. object is being created but the feedback is not being stored there.
| priority | feedback route not posting data to backend feedback route not posting data correctly to backend object is being created but the feedback is not being stored there | 1 |
56,303 | 3,078,785,809 | IssuesEvent | 2015-08-21 12:49:14 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Предупреждение пользователей о несовместимом ПО. | Compatibility Component-Logic Component-UI enhancement imported Priority-High | _From [a.rain...@gmail.com](https://code.google.com/u/117892482479228821242/) on October 16, 2011 00:11:49_
Пользователи приложений использующих p2p нередко встречаются с проблемами совместимости этих приложений с различными программами контроля сети (фаерволами, антивирусами, и другими комплексными решениями). Предлагаю при старте как минимум выдавать предупреждение о несовместимости с каким либо приложением.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=574_ | 1.0 | Предупреждение пользователей о несовместимом ПО. - _From [a.rain...@gmail.com](https://code.google.com/u/117892482479228821242/) on October 16, 2011 00:11:49_
Пользователи приложений использующих p2p нередко встречаются с проблемами совместимости этих приложений с различными программами контроля сети (фаерволами, антивирусами, и другими комплексными решениями). Предлагаю при старте как минимум выдавать предупреждение о несовместимости с каким либо приложением.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=574_ | priority | предупреждение пользователей о несовместимом по from on october пользователи приложений использующих нередко встречаются с проблемами совместимости этих приложений с различными программами контроля сети фаерволами антивирусами и другими комплексными решениями предлагаю при старте как минимум выдавать предупреждение о несовместимости с каким либо приложением original issue | 1 |
65,905 | 3,248,128,961 | IssuesEvent | 2015-10-17 02:05:08 | pvpgn/pvpgn-server | https://api.github.com/repos/pvpgn/pvpgn-server | closed | won't run Ubuntu 14.04 LTS | priority high | I've got it compiled and installed but when I try to run bnetd, d2cs or d2dbs they don't work.
The error message I get when running in gdb is "Inferior 1 (process pid) exited normally"
64-bit issue? I'm stumped. | 1.0 | won't run Ubuntu 14.04 LTS - I've got it compiled and installed but when I try to run bnetd, d2cs or d2dbs they don't work.
The error message I get when running in gdb is "Inferior 1 (process pid) exited normally"
64-bit issue? I'm stumped. | priority | won t run ubuntu lts i ve got it compiled and installed but when i try to run bnetd or they don t work the error message i get when running in gdb is inferior process pid exited normally bit issue i m stumped | 1 |
700,046 | 24,043,313,713 | IssuesEvent | 2022-09-16 05:32:39 | prisma/prisma | https://api.github.com/repos/prisma/prisma | closed | `updateMany()` causes lost-updates | bug/2-confirmed kind/bug topic: prisma-client tech/engines team/client topic: performance/queries topic: optimistic concurrency control priority/high size/m topic: query race condition | ### Bug description
`updateMany()` returns incorrect updated counts when multiple processes or asynchronous functions run simultaneously.
The SQL log of `updateMany()` shows it emits `SELECT` and then `UPDATE`. Since prisma uses the default isolation level of DBMS, those queries are NON-REPEATEBLE READ in many DBMS and can cause lost-updates.
I believe `SELECT ... FOR UPDATE` or single query `UPDATE .... WHERE ...` is required.
### How to reproduce
<details>
<summary>bookSeat.js</summary>
```js
const {PrismaClient} = require('@prisma/client')
const client = new PrismaClient({
// log: ["query"]
})
async function bookSeat(userId) {
const movieName = 'Hidden Figures'
// Find the first available seat
// availableSeat.version might be 0
const availableSeat = await client.seat.findFirst({
where: {
// Movie: {
// name: movieName,
// },
claimedBy: null,
},
orderBy: [{id: "asc"}]
})
if (!availableSeat) {
throw new Error(`Oh no! ${movieName} is all booked.`)
}
// Only mark the seat as claimed if the availableSeat.version
// matches the version we're updating. Additionally, increment the
// version when we perform this update so all other clients trying
// to book this same seat will have an outdated version.
const seats = await client.seat.updateMany({
data: {
claimedBy: userId,
version: {
increment: 1,
},
},
where: {
id: availableSeat.id,
version: availableSeat.version, // This version field is the key; only claim seat if in-memory version matches database version, indicating that the field has not been updated
},
})
if (seats.count === 0) {
throw new Error(`That seat is already booked! Please try again.`)
}
return seats.count
}
async function demonstrateLostUpdate() {
if (process.argv[2] === "createData") {
await client.seat.deleteMany()
for (let i = 0; i < 1000; i++) {
await client.seat.create({
data: {
id: i,
version: 0,
movieId: 1,
claimedBy: null,
}
})
}
process.exit()
}
const userId = process.argv[2]
let updatedCount = 0
for (let i = 0; i < 1000; i++) {
try {
updatedCount += await bookSeat(userId)
} catch {
// ignore lock failure
}
}
// Detect lost-updates
const actualCount = await client.seat.count({
where: {
claimedBy: userId
},
})
console.log({
userId,
updatedCountByUpdateMany: updatedCount,
actualUpdatedCount: actualCount
})
process.exit()
}
demonstrateLostUpdate()
```
</details>
With the above script, run
```sh
$ node bookSeat.js createData
$ node bookSeat.js Sorcha & node bookSeat.js Ellen
```
The schema and logic in the script are taken from the optimistic concurrency control example in the doc: https://www.prisma.io/docs/guides/performance-and-optimization/prisma-client-transactions-guide#optimistic-concurrency-control
Outputs:
```
{
userId: 'Sorcha',
updatedCountByUpdateMany: 968,
actualUpdatedCount: 461
}
{
userId: 'Ellen',
updatedCountByUpdateMany: 974,
actualUpdatedCount: 539
}
```
### Expected behavior
`updatedCountByUpdateMany` should be equal to `actualUpdatedCount` in the outputs.
### Prisma information
schema.prisma:
```
// This is your Prisma schema file,
// learn more about it in the docs: https://pris.ly/d/prisma-schema
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
}
model Seat {
id Int @id @default(autoincrement())
userId Int?
// claimedBy User? @relation(fields: [userId], references: [id])
claimedBy String?
movieId Int
// movie Movie @relation(fields: [movieId], references: [id])
version Int
}
```
SQL query Logs of repro:
> prisma:query SELECT "public"."Seat"."id", "public"."Seat"."userId", "public"."Seat"."claimedBy", "public"."Seat"."movieId", "public"."Seat"."version" FROM "public"."Seat" WHERE "public"."Seat"."claimedBy" IS NULL ORDER BY "public"."Seat"."id" ASC LIMIT $1 OFFSET $2
> prisma:query BEGIN
> prisma:query SELECT "public"."Seat"."id" FROM "public"."Seat" WHERE ("public"."Seat"."id" = $1 AND "public"."Seat"."version" = $2)
> prisma:query UPDATE "public"."Seat" SET "claimedBy" = $1, "version" = ("version" + $2) WHERE "public"."Seat"."id" IN ($3)
> prisma:query COMMIT
### Environment & setup
- OS: Mac OS
- Database: PostgreSQL
- Node.js version: v14.17.0
### Prisma Version
```
Environment variables loaded from .env
prisma : 2.28.0
@prisma/client : 2.28.0
Current platform : darwin
Query Engine : query-engine 89facabd0366f63911d089156a7a70125bfbcd27 (at node_modules/prisma/node_modules/@prisma/engines/query-engine-darwin)
Migration Engine : migration-engine-cli 89facabd0366f63911d089156a7a70125bfbcd27 (at node_modules/prisma/node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 89facabd0366f63911d089156a7a70125bfbcd27 (at node_modules/prisma/node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 89facabd0366f63911d089156a7a70125bfbcd27 (at node_modules/prisma/node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 89facabd0366f63911d089156a7a70125bfbcd27
Studio : 0.417.0
```
| 1.0 | `updateMany()` causes lost-updates - ### Bug description
`updateMany()` returns incorrect updated counts when multiple processes or asynchronous functions run simultaneously.
The SQL log of `updateMany()` shows it emits `SELECT` and then `UPDATE`. Since prisma uses the default isolation level of DBMS, those queries are NON-REPEATEBLE READ in many DBMS and can cause lost-updates.
I believe `SELECT ... FOR UPDATE` or single query `UPDATE .... WHERE ...` is required.
### How to reproduce
<details>
<summary>bookSeat.js</summary>
```js
const {PrismaClient} = require('@prisma/client')
const client = new PrismaClient({
// log: ["query"]
})
async function bookSeat(userId) {
const movieName = 'Hidden Figures'
// Find the first available seat
// availableSeat.version might be 0
const availableSeat = await client.seat.findFirst({
where: {
// Movie: {
// name: movieName,
// },
claimedBy: null,
},
orderBy: [{id: "asc"}]
})
if (!availableSeat) {
throw new Error(`Oh no! ${movieName} is all booked.`)
}
// Only mark the seat as claimed if the availableSeat.version
// matches the version we're updating. Additionally, increment the
// version when we perform this update so all other clients trying
// to book this same seat will have an outdated version.
const seats = await client.seat.updateMany({
data: {
claimedBy: userId,
version: {
increment: 1,
},
},
where: {
id: availableSeat.id,
version: availableSeat.version, // This version field is the key; only claim seat if in-memory version matches database version, indicating that the field has not been updated
},
})
if (seats.count === 0) {
throw new Error(`That seat is already booked! Please try again.`)
}
return seats.count
}
async function demonstrateLostUpdate() {
if (process.argv[2] === "createData") {
await client.seat.deleteMany()
for (let i = 0; i < 1000; i++) {
await client.seat.create({
data: {
id: i,
version: 0,
movieId: 1,
claimedBy: null,
}
})
}
process.exit()
}
const userId = process.argv[2]
let updatedCount = 0
for (let i = 0; i < 1000; i++) {
try {
updatedCount += await bookSeat(userId)
} catch {
// ignore lock failure
}
}
// Detect lost-updates
const actualCount = await client.seat.count({
where: {
claimedBy: userId
},
})
console.log({
userId,
updatedCountByUpdateMany: updatedCount,
actualUpdatedCount: actualCount
})
process.exit()
}
demonstrateLostUpdate()
```
</details>
With the above script, run
```sh
$ node bookSeat.js createData
$ node bookSeat.js Sorcha & node bookSeat.js Ellen
```
The schema and logic in the script are taken from the optimistic concurrency control example in the doc: https://www.prisma.io/docs/guides/performance-and-optimization/prisma-client-transactions-guide#optimistic-concurrency-control
Outputs:
```
{
userId: 'Sorcha',
updatedCountByUpdateMany: 968,
actualUpdatedCount: 461
}
{
userId: 'Ellen',
updatedCountByUpdateMany: 974,
actualUpdatedCount: 539
}
```
### Expected behavior
`updatedCountByUpdateMany` should be equal to `actualUpdatedCount` in the outputs.
### Prisma information
schema.prisma:
```
// This is your Prisma schema file,
// learn more about it in the docs: https://pris.ly/d/prisma-schema
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
}
model Seat {
id Int @id @default(autoincrement())
userId Int?
// claimedBy User? @relation(fields: [userId], references: [id])
claimedBy String?
movieId Int
// movie Movie @relation(fields: [movieId], references: [id])
version Int
}
```
SQL query Logs of repro:
> prisma:query SELECT "public"."Seat"."id", "public"."Seat"."userId", "public"."Seat"."claimedBy", "public"."Seat"."movieId", "public"."Seat"."version" FROM "public"."Seat" WHERE "public"."Seat"."claimedBy" IS NULL ORDER BY "public"."Seat"."id" ASC LIMIT $1 OFFSET $2
> prisma:query BEGIN
> prisma:query SELECT "public"."Seat"."id" FROM "public"."Seat" WHERE ("public"."Seat"."id" = $1 AND "public"."Seat"."version" = $2)
> prisma:query UPDATE "public"."Seat" SET "claimedBy" = $1, "version" = ("version" + $2) WHERE "public"."Seat"."id" IN ($3)
> prisma:query COMMIT
### Environment & setup
- OS: Mac OS
- Database: PostgreSQL
- Node.js version: v14.17.0
### Prisma Version
```
Environment variables loaded from .env
prisma : 2.28.0
@prisma/client : 2.28.0
Current platform : darwin
Query Engine : query-engine 89facabd0366f63911d089156a7a70125bfbcd27 (at node_modules/prisma/node_modules/@prisma/engines/query-engine-darwin)
Migration Engine : migration-engine-cli 89facabd0366f63911d089156a7a70125bfbcd27 (at node_modules/prisma/node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 89facabd0366f63911d089156a7a70125bfbcd27 (at node_modules/prisma/node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 89facabd0366f63911d089156a7a70125bfbcd27 (at node_modules/prisma/node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 89facabd0366f63911d089156a7a70125bfbcd27
Studio : 0.417.0
```
| priority | updatemany causes lost updates bug description updatemany returns incorrect updated counts when multiple processes or asynchronous functions run simultaneously the sql log of updatemany shows it emits select and then update since prisma uses the default isolation level of dbms those queries are non repeateble read in many dbms and can cause lost updates i believe select for update or single query update where is required how to reproduce bookseat js js const prismaclient require prisma client const client new prismaclient log async function bookseat userid const moviename hidden figures find the first available seat availableseat version might be const availableseat await client seat findfirst where movie name moviename claimedby null orderby if availableseat throw new error oh no moviename is all booked only mark the seat as claimed if the availableseat version matches the version we re updating additionally increment the version when we perform this update so all other clients trying to book this same seat will have an outdated version const seats await client seat updatemany data claimedby userid version increment where id availableseat id version availableseat version this version field is the key only claim seat if in memory version matches database version indicating that the field has not been updated if seats count throw new error that seat is already booked please try again return seats count async function demonstratelostupdate if process argv createdata await client seat deletemany for let i i i await client seat create data id i version movieid claimedby null process exit const userid process argv let updatedcount for let i i i try updatedcount await bookseat userid catch ignore lock failure detect lost updates const actualcount await client seat count where claimedby userid console log userid updatedcountbyupdatemany updatedcount actualupdatedcount actualcount process exit demonstratelostupdate with the above script run sh node bookseat js createdata node bookseat js sorcha node bookseat js ellen the schema and logic in the script are taken from the optimistic concurrency control example in the doc outputs userid sorcha updatedcountbyupdatemany actualupdatedcount userid ellen updatedcountbyupdatemany actualupdatedcount expected behavior updatedcountbyupdatemany should be equal to actualupdatedcount in the outputs prisma information schema prisma this is your prisma schema file learn more about it in the docs datasource db provider postgresql url env database url generator client provider prisma client js model seat id int id default autoincrement userid int claimedby user relation fields references claimedby string movieid int movie movie relation fields references version int sql query logs of repro prisma query select public seat id public seat userid public seat claimedby public seat movieid public seat version from public seat where public seat claimedby is null order by public seat id asc limit offset prisma query begin prisma query select public seat id from public seat where public seat id and public seat version prisma query update public seat set claimedby version version where public seat id in prisma query commit environment setup os mac os database postgresql node js version prisma version environment variables loaded from env prisma prisma client current platform darwin query engine query engine at node modules prisma node modules prisma engines query engine darwin migration engine migration engine cli at node modules prisma node modules prisma engines migration engine darwin introspection engine introspection core at node modules prisma node modules prisma engines introspection engine darwin format binary prisma fmt at node modules prisma node modules prisma engines prisma fmt darwin default engines hash studio | 1 |
294,864 | 9,049,580,300 | IssuesEvent | 2019-02-12 05:23:47 | LordNed/Winditor | https://api.github.com/repos/LordNed/Winditor | opened | Implement an ActorReferenceControl | Priority: High | Several entity types, among them paths and environment lighting, have references to other objects within the world. These references should be handled by an `ActorReferenceControl`, which will allow the user to select, remove, or replace the actor which it references.
When replacing the stored reference, the control will display a list of the entities matching the referenced object's type. While the reference will be stored as type `WDOMNode`, this type filtering will prevent the user from, for example, selecting a waypoint as a reference for the environment lighting's palette data.
When removing or replacing a reference, the previous reference must be marked as `destroyed`. | 1.0 | Implement an ActorReferenceControl - Several entity types, among them paths and environment lighting, have references to other objects within the world. These references should be handled by an `ActorReferenceControl`, which will allow the user to select, remove, or replace the actor which it references.
When replacing the stored reference, the control will display a list of the entities matching the referenced object's type. While the reference will be stored as type `WDOMNode`, this type filtering will prevent the user from, for example, selecting a waypoint as a reference for the environment lighting's palette data.
When removing or replacing a reference, the previous reference must be marked as `destroyed`. | priority | implement an actorreferencecontrol several entity types among them paths and environment lighting have references to other objects within the world these references should be handled by an actorreferencecontrol which will allow the user to select remove or replace the actor which it references when replacing the stored reference the control will display a list of the entities matching the referenced object s type while the reference will be stored as type wdomnode this type filtering will prevent the user from for example selecting a waypoint as a reference for the environment lighting s palette data when removing or replacing a reference the previous reference must be marked as destroyed | 1 |
662,235 | 22,105,411,192 | IssuesEvent | 2022-06-01 16:37:36 | encorelab/ck-board | https://api.github.com/repos/encorelab/ck-board | closed | Groupings - Model & API Endpoints | enhancement high priority | Inside `/backend`, we need to create a `Group` model that will store all data associated with a group.
```
class Group {
groupID: string // unique id for group
name: string // group name
members: string[] // array of user ids of all members in group
... other data as you may seem fit
}
```
We also need to create API endpoints to `create`, `update`, `read`, and `delete` groups.
You should synch with @DCH25 regarding whatever endpoints he may need to create his UI. | 1.0 | Groupings - Model & API Endpoints - Inside `/backend`, we need to create a `Group` model that will store all data associated with a group.
```
class Group {
groupID: string // unique id for group
name: string // group name
members: string[] // array of user ids of all members in group
... other data as you may seem fit
}
```
We also need to create API endpoints to `create`, `update`, `read`, and `delete` groups.
You should synch with @DCH25 regarding whatever endpoints he may need to create his UI. | priority | groupings model api endpoints inside backend we need to create a group model that will store all data associated with a group class group groupid string unique id for group name string group name members string array of user ids of all members in group other data as you may seem fit we also need to create api endpoints to create update read and delete groups you should synch with regarding whatever endpoints he may need to create his ui | 1 |
792,052 | 27,944,346,972 | IssuesEvent | 2023-03-24 00:56:26 | RTXteam/RTX | https://api.github.com/repos/RTXteam/RTX | closed | Primary knowledge source: deadline 3/24/2023 | high priority | From Tyler on March 7th: of utmost priority (ASAP) is having a `primary_knowledge_source` on _every_ edge which indicates the original data source that provided the assertion leading to the edge. We should check that:
1. All RTX-KG2 edges have this
2. All ARAX virtual edges have this
3. All edges `expand` gets preserves this information
See page 2 of [this](https://docs.google.com/document/d/177sOmjTueIK4XKJ0GjxsARg909CaU71tReIehAp5DDo/edit#heading=h.9mu3cpnwwefy) re: format (basically, infores)
@chunyuma , can you check the xDTD and DTD edges? This is related to #1970. I will have Kevin start on the other edges when he gets onboarded. | 1.0 | Primary knowledge source: deadline 3/24/2023 - From Tyler on March 7th: of utmost priority (ASAP) is having a `primary_knowledge_source` on _every_ edge which indicates the original data source that provided the assertion leading to the edge. We should check that:
1. All RTX-KG2 edges have this
2. All ARAX virtual edges have this
3. All edges `expand` gets preserves this information
See page 2 of [this](https://docs.google.com/document/d/177sOmjTueIK4XKJ0GjxsARg909CaU71tReIehAp5DDo/edit#heading=h.9mu3cpnwwefy) re: format (basically, infores)
@chunyuma , can you check the xDTD and DTD edges? This is related to #1970. I will have Kevin start on the other edges when he gets onboarded. | priority | primary knowledge source deadline from tyler on march of utmost priority asap is having a primary knowledge source on every edge which indicates the original data source that provided the assertion leading to the edge we should check that all rtx edges have this all arax virtual edges have this all edges expand gets preserves this information see page of re format basically infores chunyuma can you check the xdtd and dtd edges this is related to i will have kevin start on the other edges when he gets onboarded | 1 |
68,813 | 3,292,782,007 | IssuesEvent | 2015-10-30 16:04:10 | cs2103aug2015-w13-2j/main | https://api.github.com/repos/cs2103aug2015-w13-2j/main | closed | Implement search function (which adopts the 'filter is:search' method) | component.logic priority.high status.ongoing type.task | search command (As discussed on 29/10, after the CS2101 Software Demo Conference), is as follows:
search lunch with
OR
search "this is some phrase in one of my tasks"
The command without double quotations refers to the multi-worded search feature.
The command with double quotations refers to searching for a specific phrase in a task name. | 1.0 | Implement search function (which adopts the 'filter is:search' method) - search command (As discussed on 29/10, after the CS2101 Software Demo Conference), is as follows:
search lunch with
OR
search "this is some phrase in one of my tasks"
The command without double quotations refers to the multi-worded search feature.
The command with double quotations refers to searching for a specific phrase in a task name. | priority | implement search function which adopts the filter is search method search command as discussed on after the software demo conference is as follows search lunch with or search this is some phrase in one of my tasks the command without double quotations refers to the multi worded search feature the command with double quotations refers to searching for a specific phrase in a task name | 1 |
483,278 | 13,921,813,365 | IssuesEvent | 2020-10-21 12:28:34 | fossasia/open-event-frontend | https://api.github.com/repos/fossasia/open-event-frontend | opened | Organizer: Adding speaker without email results in sessions added to "my sessions" from organizer him/herself | Priority: High bug | If an organizer adds a speaker "without email" the overview page shows the email of the organizer and sessions show up under "my sessions".

| 1.0 | Organizer: Adding speaker without email results in sessions added to "my sessions" from organizer him/herself - If an organizer adds a speaker "without email" the overview page shows the email of the organizer and sessions show up under "my sessions".

| priority | organizer adding speaker without email results in sessions added to my sessions from organizer him herself if an organizer adds a speaker without email the overview page shows the email of the organizer and sessions show up under my sessions | 1 |
505,926 | 14,654,728,346 | IssuesEvent | 2020-12-28 09:21:58 | schemathesis/schemathesis | https://api.github.com/repos/schemathesis/schemathesis | opened | [FEATURE] A way to refresh auth | Difficulty: Medium Priority: High Type: Feature | **Is your feature request related to a problem? Please describe.**
Currently, if the auth token expires during the test run, there is no built-in way to refresh it. It will be nice to have some way to register a function that will refresh auth. And maybe some refresh interval
**Describe the solution you'd like**
A decorator that will register such a function that will return headers required for authorization
| 1.0 | [FEATURE] A way to refresh auth - **Is your feature request related to a problem? Please describe.**
Currently, if the auth token expires during the test run, there is no built-in way to refresh it. It will be nice to have some way to register a function that will refresh auth. And maybe some refresh interval
**Describe the solution you'd like**
A decorator that will register such a function that will return headers required for authorization
| priority | a way to refresh auth is your feature request related to a problem please describe currently if the auth token expires during the test run there is no built in way to refresh it it will be nice to have some way to register a function that will refresh auth and maybe some refresh interval describe the solution you d like a decorator that will register such a function that will return headers required for authorization | 1 |
658,333 | 21,885,019,832 | IssuesEvent | 2022-05-19 17:43:12 | comses/comses.net | https://api.github.com/repos/comses/comses.net | closed | discourse sync: match username validation | bug high priority | It appears that Discourse usernames can't have the `@` symbol but Django usernames are OK with it, which causes errors like the following:
```
failed to sync user <User: example@colorado.edu> with discourse:
{u'message': u'Username must only include numbers, letters, dashes, dots, and underscores', u'is_developer': False, u'values': {u'username': u'example@colorado.edu', u'name': u'Example Name', u'email': u'example@colorado.edu'}, u'errors': {u'username': ["u'must only include numbers, letters, dashes, dots, and underscores'"]}, u'success': False}
```
We can fix this by forcing the same username validation logic on our django usernames (probably the easiest since we have a bit more control over that) or we could check if it's possible to make the username validation on Discourse more lax.. | 1.0 | discourse sync: match username validation - It appears that Discourse usernames can't have the `@` symbol but Django usernames are OK with it, which causes errors like the following:
```
failed to sync user <User: example@colorado.edu> with discourse:
{u'message': u'Username must only include numbers, letters, dashes, dots, and underscores', u'is_developer': False, u'values': {u'username': u'example@colorado.edu', u'name': u'Example Name', u'email': u'example@colorado.edu'}, u'errors': {u'username': ["u'must only include numbers, letters, dashes, dots, and underscores'"]}, u'success': False}
```
We can fix this by forcing the same username validation logic on our django usernames (probably the easiest since we have a bit more control over that) or we could check if it's possible to make the username validation on Discourse more lax.. | priority | discourse sync match username validation it appears that discourse usernames can t have the symbol but django usernames are ok with it which causes errors like the following failed to sync user with discourse u message u username must only include numbers letters dashes dots and underscores u is developer false u values u username u example colorado edu u name u example name u email u example colorado edu u errors u username u success false we can fix this by forcing the same username validation logic on our django usernames probably the easiest since we have a bit more control over that or we could check if it s possible to make the username validation on discourse more lax | 1 |
343,374 | 10,329,003,855 | IssuesEvent | 2019-09-02 11:00:12 | geosolutions-it/geonode | https://api.github.com/repos/geosolutions-it/geonode | opened | Return count of resources in metric_data/request.count | Priority: High analytics enhancement | We need a query param to return the count of resources in a requested interval.
This is the current state:
Request
```
/monitoring/api/metric_data/request.count/
group_by=resource
&valid_from=2019-01-01+00:00:00
&valid_to=2019-12-31+23:59:59
&interval=31536000
&resource_type=layer
&event_type=upload
```
Response:
```
{
"data": {
"input_valid_from": "2019-01-01T00:00:00-06:00",
"input_valid_to": "2019-12-31T23:59:59-06:00",
"data": [
{
"valid_from": "2019-01-01T00:00:00-06:00",
"data": [
{
"resource": {
"type": "document",
"name": "test.txt",
"id": 36
},
"samples_count": 1,
"val": "1.0000",
"min": "1.0000",
"max": "1.0000",
"sum": "1.0000",
"label": "Count",
"metric_count": 1
},
{
"resource": {
"type": "document",
"name": "test001.txt",
"id": 37
},
"samples_count": 1,
"val": "1.0000",
"min": "1.0000",
"max": "1.0000",
"sum": "1.0000",
"label": "Count",
"metric_count": 1
}
],
"valid_to": "2020-01-01T00:00:00-06:00"
}
],
"metric": "request.count",
"interval": 31536000,
"type": "count",
"axis_label": "Count",
"label": null
}
}
```
We need a query param to return the data length so val equal to 2:
[PROPOSAL] `data`=count
Request
```
/monitoring/api/metric_data/request.count/
group_by=resource
&data=count
&valid_from=2019-01-01+00:00:00
&valid_to=2019-12-31+23:59:59
&interval=31536000
&resource_type=layer
&event_type=upload
```
Response:
```
{
"data": {
"input_valid_from": "2019-01-01T00:00:00-06:00",
"input_valid_to": "2019-12-31T23:59:59-06:00",
"data": [
{
"valid_from": "2019-01-01T00:00:00-06:00",
"data": [
{
"samples_count": 1,
"val": 2,
"min": "1.0000",
"max": "1.0000",
"sum": "1.0000",
"label": "Count",
"metric_count": 1
}
],
"valid_to": "2020-01-01T00:00:00-06:00"
}
],
"metric": "request.count",
"interval": 31536000,
"type": "count",
"axis_label": "Count",
"label": null
}
}
```
| 1.0 | Return count of resources in metric_data/request.count - We need a query param to return the count of resources in a requested interval.
This is the current state:
Request
```
/monitoring/api/metric_data/request.count/
group_by=resource
&valid_from=2019-01-01+00:00:00
&valid_to=2019-12-31+23:59:59
&interval=31536000
&resource_type=layer
&event_type=upload
```
Response:
```
{
"data": {
"input_valid_from": "2019-01-01T00:00:00-06:00",
"input_valid_to": "2019-12-31T23:59:59-06:00",
"data": [
{
"valid_from": "2019-01-01T00:00:00-06:00",
"data": [
{
"resource": {
"type": "document",
"name": "test.txt",
"id": 36
},
"samples_count": 1,
"val": "1.0000",
"min": "1.0000",
"max": "1.0000",
"sum": "1.0000",
"label": "Count",
"metric_count": 1
},
{
"resource": {
"type": "document",
"name": "test001.txt",
"id": 37
},
"samples_count": 1,
"val": "1.0000",
"min": "1.0000",
"max": "1.0000",
"sum": "1.0000",
"label": "Count",
"metric_count": 1
}
],
"valid_to": "2020-01-01T00:00:00-06:00"
}
],
"metric": "request.count",
"interval": 31536000,
"type": "count",
"axis_label": "Count",
"label": null
}
}
```
We need a query param to return the data length so val equal to 2:
[PROPOSAL] `data`=count
Request
```
/monitoring/api/metric_data/request.count/
group_by=resource
&data=count
&valid_from=2019-01-01+00:00:00
&valid_to=2019-12-31+23:59:59
&interval=31536000
&resource_type=layer
&event_type=upload
```
Response:
```
{
"data": {
"input_valid_from": "2019-01-01T00:00:00-06:00",
"input_valid_to": "2019-12-31T23:59:59-06:00",
"data": [
{
"valid_from": "2019-01-01T00:00:00-06:00",
"data": [
{
"samples_count": 1,
"val": 2,
"min": "1.0000",
"max": "1.0000",
"sum": "1.0000",
"label": "Count",
"metric_count": 1
}
],
"valid_to": "2020-01-01T00:00:00-06:00"
}
],
"metric": "request.count",
"interval": 31536000,
"type": "count",
"axis_label": "Count",
"label": null
}
}
```
| priority | return count of resources in metric data request count we need a query param to return the count of resources in a requested interval this is the current state request monitoring api metric data request count group by resource valid from valid to interval resource type layer event type upload response data input valid from input valid to data valid from data resource type document name test txt id samples count val min max sum label count metric count resource type document name txt id samples count val min max sum label count metric count valid to metric request count interval type count axis label count label null we need a query param to return the data length so val equal to data count request monitoring api metric data request count group by resource data count valid from valid to interval resource type layer event type upload response data input valid from input valid to data valid from data samples count val min max sum label count metric count valid to metric request count interval type count axis label count label null | 1 |
361,795 | 10,720,256,310 | IssuesEvent | 2019-10-26 16:22:47 | AY1920S1-CS2113-T14-1/main | https://api.github.com/repos/AY1920S1-CS2113-T14-1/main | closed | Create [new] command in [Impression] context | priority.High type.Task | **Add a new Treatment or Evidence item to this Impression**
Format: `new _type_ <relevant switches> [-g[o]]`
[View relevant switches here](https://github.com/AY1920S1-CS2113-T14-1/main/blob/master/docs/UserGuide.adoc#131-new---add-a-new-treatment-or-evidence-item-to-this-impression) | 1.0 | Create [new] command in [Impression] context - **Add a new Treatment or Evidence item to this Impression**
Format: `new _type_ <relevant switches> [-g[o]]`
[View relevant switches here](https://github.com/AY1920S1-CS2113-T14-1/main/blob/master/docs/UserGuide.adoc#131-new---add-a-new-treatment-or-evidence-item-to-this-impression) | priority | create command in context add a new treatment or evidence item to this impression format new type | 1 |
775,442 | 27,234,589,231 | IssuesEvent | 2023-02-21 15:28:26 | alexrhowe/APOLLO | https://api.github.com/repos/alexrhowe/APOLLO | closed | lnprior() can underflow with a large number of parameters. | Bug High priority | For a uniform prior, each parameter multiplies the total prior by 1/(upper-lower). If there are many parameters with wide priors, this can result in an underflow error. The final line in lnprior():
return np.log(np.prod(priors))+penalty
Should read:
return np.sum(np.log(priors))+penalty | 1.0 | lnprior() can underflow with a large number of parameters. - For a uniform prior, each parameter multiplies the total prior by 1/(upper-lower). If there are many parameters with wide priors, this can result in an underflow error. The final line in lnprior():
return np.log(np.prod(priors))+penalty
Should read:
return np.sum(np.log(priors))+penalty | priority | lnprior can underflow with a large number of parameters for a uniform prior each parameter multiplies the total prior by upper lower if there are many parameters with wide priors this can result in an underflow error the final line in lnprior return np log np prod priors penalty should read return np sum np log priors penalty | 1 |
824,704 | 31,167,535,830 | IssuesEvent | 2023-08-16 21:02:19 | bcgov/foi-flow | https://api.github.com/repos/bcgov/foi-flow | opened | If we delete a portfolio file, Its attachment files still appears on the records tab | bug high priority | **Describe the bug in current situation**
If we delete a portfolio file, then its attachment files are showing up on the records tab
**Link bug to the User Story**
**Impact of this bug**
Describe the impact, i.e. what the impact is, and number of users impacted.
High
**Chance of Occurring (high/medium/low/very low)**
**Pre Conditions: which Env, any pre-requesites or assumptions to execute steps?**
Test-marshal
**Steps to Reproduce**
Steps to reproduce the behavior:
1. Upload some portfolio file to records tab as a ministry user
2. Then try to delete the parent portfolio file
3. Then its attachment files remains on the records tab as it is appeared to be separate ones
**Actual/ observed behaviour/ results**
**Expected behaviour**
A clear and concise description of what you expected to happen. Use the gherking language.
**Screenshots/ Visual Reference/ Source**
If applicable, add screenshots to help explain your problem. You an use screengrab.

| 1.0 | If we delete a portfolio file, Its attachment files still appears on the records tab - **Describe the bug in current situation**
If we delete a portfolio file, then its attachment files are showing up on the records tab
**Link bug to the User Story**
**Impact of this bug**
Describe the impact, i.e. what the impact is, and number of users impacted.
High
**Chance of Occurring (high/medium/low/very low)**
**Pre Conditions: which Env, any pre-requesites or assumptions to execute steps?**
Test-marshal
**Steps to Reproduce**
Steps to reproduce the behavior:
1. Upload some portfolio file to records tab as a ministry user
2. Then try to delete the parent portfolio file
3. Then its attachment files remains on the records tab as it is appeared to be separate ones
**Actual/ observed behaviour/ results**
**Expected behaviour**
A clear and concise description of what you expected to happen. Use the gherking language.
**Screenshots/ Visual Reference/ Source**
If applicable, add screenshots to help explain your problem. You an use screengrab.

| priority | if we delete a portfolio file its attachment files still appears on the records tab describe the bug in current situation if we delete a portfolio file then its attachment files are showing up on the records tab link bug to the user story impact of this bug describe the impact i e what the impact is and number of users impacted high chance of occurring high medium low very low pre conditions which env any pre requesites or assumptions to execute steps test marshal steps to reproduce steps to reproduce the behavior upload some portfolio file to records tab as a ministry user then try to delete the parent portfolio file then its attachment files remains on the records tab as it is appeared to be separate ones actual observed behaviour results expected behaviour a clear and concise description of what you expected to happen use the gherking language screenshots visual reference source if applicable add screenshots to help explain your problem you an use screengrab | 1 |
220,463 | 7,360,109,514 | IssuesEvent | 2018-03-10 15:11:49 | imjuangarcia/personal-website | https://api.github.com/repos/imjuangarcia/personal-website | closed | Notification when cookie had expired | Development Frontend Priority: High | To warn the users that they're on the checkout page, but they don't have nothing to check out | 1.0 | Notification when cookie had expired - To warn the users that they're on the checkout page, but they don't have nothing to check out | priority | notification when cookie had expired to warn the users that they re on the checkout page but they don t have nothing to check out | 1 |
763,431 | 26,756,182,550 | IssuesEvent | 2023-01-31 00:31:24 | ShahzaibParacha/COMP-4350-Project | https://api.github.com/repos/ShahzaibParacha/COMP-4350-Project | opened | [dev-task] Create schema for Comment | backend high priority dev-task | Create the MongoDB model for Comment. It will contain at least a post id the comment is for, contents, and a reference to the user that made it.
#10: Comment on posts | 1.0 | [dev-task] Create schema for Comment - Create the MongoDB model for Comment. It will contain at least a post id the comment is for, contents, and a reference to the user that made it.
#10: Comment on posts | priority | create schema for comment create the mongodb model for comment it will contain at least a post id the comment is for contents and a reference to the user that made it comment on posts | 1 |
400,811 | 11,781,013,923 | IssuesEvent | 2020-03-16 21:21:36 | QuantEcon/lecture-source-py | https://api.github.com/repos/QuantEcon/lecture-source-py | closed | [contributors] update python landing page to include contributors | high-priority | @DrDrij we would like to update the front page (the current python.quantecon.org) or the `index.rst` file for `lecture-source-py` to include `contributors.
We would like to:
1. move the `commit` link to be an item below the `Last compiled: {DATE}`:
1. add another `contributors` item and link to contributors page on github https://github.com/QuantEcon/lecture-source-py/graphs/contributors (for now)
<img width="495" alt="image" src="https://user-images.githubusercontent.com/8263752/76051417-c7e01d80-5fbe-11ea-8d55-869cb408117a.png">
In the future we will want to make an `contributors.rst` page that is similar to https://datascience.quantecon.org/contributors.html.
This will address Item 3 in https://github.com/QuantEcon/lecture-source-py/issues/946
| 1.0 | [contributors] update python landing page to include contributors - @DrDrij we would like to update the front page (the current python.quantecon.org) or the `index.rst` file for `lecture-source-py` to include `contributors.
We would like to:
1. move the `commit` link to be an item below the `Last compiled: {DATE}`:
1. add another `contributors` item and link to contributors page on github https://github.com/QuantEcon/lecture-source-py/graphs/contributors (for now)
<img width="495" alt="image" src="https://user-images.githubusercontent.com/8263752/76051417-c7e01d80-5fbe-11ea-8d55-869cb408117a.png">
In the future we will want to make an `contributors.rst` page that is similar to https://datascience.quantecon.org/contributors.html.
This will address Item 3 in https://github.com/QuantEcon/lecture-source-py/issues/946
| priority | update python landing page to include contributors drdrij we would like to update the front page the current python quantecon org or the index rst file for lecture source py to include contributors we would like to move the commit link to be an item below the last compiled date add another contributors item and link to contributors page on github for now img width alt image src in the future we will want to make an contributors rst page that is similar to this will address item in | 1 |
529,531 | 15,390,734,672 | IssuesEvent | 2021-03-03 13:47:12 | Jlll1/Expelibrum | https://api.github.com/repos/Jlll1/Expelibrum | closed | Before moving an item, check the Move() parameters are not the same | bug good first issue high priority | Currently, if the book doesn't contain a valid MSDN, the Process() method will try to move it while keeping the original name of the file.
This will throw an exception if Origin and Target directories selected by the user are the same.
Before moving we should check if the paths are different or not. | 1.0 | Before moving an item, check the Move() parameters are not the same - Currently, if the book doesn't contain a valid MSDN, the Process() method will try to move it while keeping the original name of the file.
This will throw an exception if Origin and Target directories selected by the user are the same.
Before moving we should check if the paths are different or not. | priority | before moving an item check the move parameters are not the same currently if the book doesn t contain a valid msdn the process method will try to move it while keeping the original name of the file this will throw an exception if origin and target directories selected by the user are the same before moving we should check if the paths are different or not | 1 |
482,196 | 13,902,624,791 | IssuesEvent | 2020-10-20 05:45:26 | AY2021S1-CS2103T-W11-2/tp | https://api.github.com/repos/AY2021S1-CS2103T-W11-2/tp | closed | Retrieve all data related to a student | priority.High severity.High type.Story | As a user I want to view all data pertaining to a student in order to have a quick overview of information such as contacts and consultations related to a student. | 1.0 | Retrieve all data related to a student - As a user I want to view all data pertaining to a student in order to have a quick overview of information such as contacts and consultations related to a student. | priority | retrieve all data related to a student as a user i want to view all data pertaining to a student in order to have a quick overview of information such as contacts and consultations related to a student | 1 |
426,031 | 12,366,132,679 | IssuesEvent | 2020-05-18 09:56:12 | ComPWA/tensorwaves | https://api.github.com/repos/ComPWA/tensorwaves | closed | Refactor amplitude building code | Priority: High Type: Maintenance | The current code is combining building and computation/execution which is inconvenient for testing.
So as an example rather than
```python
class _CoefficientAmplitude():
def __init__(self, recipe: dict, builder):
if isinstance(recipe, dict):
pars = recipe['Parameters']
self.mag = builder._get_parameter(pars['Magnitude'])
self.phase = builder._get_parameter(pars['Phase'])
self.amp = builder._create_amplitude(recipe['Amplitude'])
def __call__(self, x):
coefficient = atfi.complex(
self.mag * atfi.cos(self.phase),
self.mag * atfi.sin(self.phase)
)
return coefficient * self.amp(x)
```
define an computation part
```python
class _CoefficientAmplitude():
def __init__(self, amplitude, mag, phase):
self.mag = mag
self.phase = phase
self.amp = amplitude
def __call__(self, x):
coefficient = atfi.complex(
self.mag * atfi.cos(self.phase),
self.mag * atfi.sin(self.phase)
)
return coefficient * self.amp(x)
```
and a building part
```python
def create_cofficient_amplitude(self, recipe: dict, builder):
if isinstance(recipe, dict):
pars = recipe['Parameters']
mag = builder._get_parameter(pars['Magnitude'])
phase = builder._get_parameter(pars['Phase'])
amp = builder._create_amplitude(recipe['Amplitude'])
return _CoefficientAmplitude(amp, mag, phase)
``` | 1.0 | Refactor amplitude building code - The current code is combining building and computation/execution which is inconvenient for testing.
So as an example rather than
```python
class _CoefficientAmplitude():
def __init__(self, recipe: dict, builder):
if isinstance(recipe, dict):
pars = recipe['Parameters']
self.mag = builder._get_parameter(pars['Magnitude'])
self.phase = builder._get_parameter(pars['Phase'])
self.amp = builder._create_amplitude(recipe['Amplitude'])
def __call__(self, x):
coefficient = atfi.complex(
self.mag * atfi.cos(self.phase),
self.mag * atfi.sin(self.phase)
)
return coefficient * self.amp(x)
```
define an computation part
```python
class _CoefficientAmplitude():
def __init__(self, amplitude, mag, phase):
self.mag = mag
self.phase = phase
self.amp = amplitude
def __call__(self, x):
coefficient = atfi.complex(
self.mag * atfi.cos(self.phase),
self.mag * atfi.sin(self.phase)
)
return coefficient * self.amp(x)
```
and a building part
```python
def create_cofficient_amplitude(self, recipe: dict, builder):
if isinstance(recipe, dict):
pars = recipe['Parameters']
mag = builder._get_parameter(pars['Magnitude'])
phase = builder._get_parameter(pars['Phase'])
amp = builder._create_amplitude(recipe['Amplitude'])
return _CoefficientAmplitude(amp, mag, phase)
``` | priority | refactor amplitude building code the current code is combining building and computation execution which is inconvenient for testing so as an example rather than python class coefficientamplitude def init self recipe dict builder if isinstance recipe dict pars recipe self mag builder get parameter pars self phase builder get parameter pars self amp builder create amplitude recipe def call self x coefficient atfi complex self mag atfi cos self phase self mag atfi sin self phase return coefficient self amp x define an computation part python class coefficientamplitude def init self amplitude mag phase self mag mag self phase phase self amp amplitude def call self x coefficient atfi complex self mag atfi cos self phase self mag atfi sin self phase return coefficient self amp x and a building part python def create cofficient amplitude self recipe dict builder if isinstance recipe dict pars recipe mag builder get parameter pars phase builder get parameter pars amp builder create amplitude recipe return coefficientamplitude amp mag phase | 1 |
191,706 | 6,841,011,996 | IssuesEvent | 2017-11-11 07:37:25 | dhowe/AdNauseam | https://api.github.com/repos/dhowe/AdNauseam | closed | 'Delete this ad' causes date-slider to reset | Bug PRIORITY: High | ### Steps for anyone to reproduce the issue
1. Open the vault with more than 300 ads
1. Slide the date slider to the left to show more than 300 ads
1. Right-click and select 'delete this ad'
1. Notice that the date-slider has reset
- OS/version: OS X
- Browser/version: Chrome
- AdNauseam version: v3.4.100
| 1.0 | 'Delete this ad' causes date-slider to reset - ### Steps for anyone to reproduce the issue
1. Open the vault with more than 300 ads
1. Slide the date slider to the left to show more than 300 ads
1. Right-click and select 'delete this ad'
1. Notice that the date-slider has reset
- OS/version: OS X
- Browser/version: Chrome
- AdNauseam version: v3.4.100
| priority | delete this ad causes date slider to reset steps for anyone to reproduce the issue open the vault with more than ads slide the date slider to the left to show more than ads right click and select delete this ad notice that the date slider has reset os version os x browser version chrome adnauseam version | 1 |
514,044 | 14,931,995,477 | IssuesEvent | 2021-01-25 06:56:57 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | edistrict.up.nic.in - desktop site instead of mobile site | browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal | <!-- @browser: Firefox 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:85.0) Gecko/20100101 Firefox/85.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/66175 -->
**URL**: http://edistrict.up.nic.in/GOs/Integratedservicequota.pdf
**Browser / Version**: Firefox 85.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes Other
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
deepak sinngh
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210118153634</li><li>channel: release</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/f225fb7a-45a5-416a-8d51-389804bf26cf)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | edistrict.up.nic.in - desktop site instead of mobile site - <!-- @browser: Firefox 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:85.0) Gecko/20100101 Firefox/85.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/66175 -->
**URL**: http://edistrict.up.nic.in/GOs/Integratedservicequota.pdf
**Browser / Version**: Firefox 85.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes Other
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
deepak sinngh
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210118153634</li><li>channel: release</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/f225fb7a-45a5-416a-8d51-389804bf26cf)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | edistrict up nic in desktop site instead of mobile site url browser version firefox operating system windows tested another browser yes other problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce deepak sinngh browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
202,281 | 7,046,235,423 | IssuesEvent | 2018-01-02 06:08:29 | RSPluto/Web-UI | https://api.github.com/repos/RSPluto/Web-UI | closed | 实时监测 - 曲线图 - 在界面选择button切换“天”“周”“月”时,系统没有再次发送新的请求 | bug Fixed High Priority | 测试步骤:
1. 进入曲线图 (默认查询“天”);
2. 选择“周”;
3. 选择“月”;
4. 选择“天”
期望结果:
2.&3&4 系统会重新触发请求,并且显示周对应的曲线图;
实际结果:
后面三个操作,界面都显示空白,并没有发送进行新的请求
| 1.0 | 实时监测 - 曲线图 - 在界面选择button切换“天”“周”“月”时,系统没有再次发送新的请求 - 测试步骤:
1. 进入曲线图 (默认查询“天”);
2. 选择“周”;
3. 选择“月”;
4. 选择“天”
期望结果:
2.&3&4 系统会重新触发请求,并且显示周对应的曲线图;
实际结果:
后面三个操作,界面都显示空白,并没有发送进行新的请求
| priority | 实时监测 曲线图 在界面选择button切换“天”“周”“月”时,系统没有再次发送新的请求 测试步骤: 进入曲线图 (默认查询“天”); 选择“周”; 选择“月”; 选择“天” 期望结果: 系统会重新触发请求,并且显示周对应的曲线图; 实际结果: 后面三个操作,界面都显示空白,并没有发送进行新的请求 | 1 |
411,338 | 12,017,118,018 | IssuesEvent | 2020-04-10 17:38:54 | AugurProject/augur | https://api.github.com/repos/AugurProject/augur | closed | Trade simulation for 24hr/7day/30 PnL chart change | Bug Needed for V2 launch Priority: High | Yes/No (do same test over different period time frames)
User 4) Adds Offer to sell 1000 shares at 0.4
User 1) Buys 1000 shares at 0.4
User 2) Adds bid to buy 350 shares at 0.6
User 3) Sells 100 shares at 0.6
PnL:
(user 1: unrealized = 200, unrealized Percent +50%)
User 4 (realized = 0) (unrealized = -$200) (-50%)
User 1) Sells 250 shares of 0.6
PnL:
(user 1: realized = $50, realized Percent +50%
(unrealized = $150, unrealized Percent +50%)
User 1) adds Offer to sell 250 @ 0.9
User 3) Buys 250 @ 0.9
PnL:
User 1 (realized = $50+$125=175)(87.5%) (unrealized = $250) (125%)
User 2 (realized = 0) (unrealized = $105) (50%)
User 3 (realized = 0) (unrealized = -$30) (-50%)
User 4 (realized = 0) (unrealized = -$500) (-125%) | 1.0 | Trade simulation for 24hr/7day/30 PnL chart change - Yes/No (do same test over different period time frames)
User 4) Adds Offer to sell 1000 shares at 0.4
User 1) Buys 1000 shares at 0.4
User 2) Adds bid to buy 350 shares at 0.6
User 3) Sells 100 shares at 0.6
PnL:
(user 1: unrealized = 200, unrealized Percent +50%)
User 4 (realized = 0) (unrealized = -$200) (-50%)
User 1) Sells 250 shares of 0.6
PnL:
(user 1: realized = $50, realized Percent +50%
(unrealized = $150, unrealized Percent +50%)
User 1) adds Offer to sell 250 @ 0.9
User 3) Buys 250 @ 0.9
PnL:
User 1 (realized = $50+$125=175)(87.5%) (unrealized = $250) (125%)
User 2 (realized = 0) (unrealized = $105) (50%)
User 3 (realized = 0) (unrealized = -$30) (-50%)
User 4 (realized = 0) (unrealized = -$500) (-125%) | priority | trade simulation for pnl chart change yes no do same test over different period time frames user adds offer to sell shares at user buys shares at user adds bid to buy shares at user sells shares at pnl user unrealized unrealized percent user realized unrealized user sells shares of pnl user realized realized percent unrealized unrealized percent user adds offer to sell user buys pnl user realized unrealized user realized unrealized user realized unrealized user realized unrealized | 1 |
279,451 | 8,665,502,843 | IssuesEvent | 2018-11-28 23:37:51 | JID8201/CHaRM-Web-App | https://api.github.com/repos/JID8201/CHaRM-Web-App | closed | Employee can authenticate to system | Priority: High | @bran100 commented on [Wed Sep 05 2018](https://github.com/JID8201/CHaRM-Android-App/issues/6)
As a Live Thrive employee, I would like the database to be secure, so only authenticated users can access the data.
Scenario 1: An authenticated user attempts to access data
Given: an employee is logged in to the web application;
When: the employee requests to see data on employees;
Then: the request gets processed and executes successfully
given the employee logged in with the right credentials.
| 1.0 | Employee can authenticate to system - @bran100 commented on [Wed Sep 05 2018](https://github.com/JID8201/CHaRM-Android-App/issues/6)
As a Live Thrive employee, I would like the database to be secure, so only authenticated users can access the data.
Scenario 1: An authenticated user attempts to access data
Given: an employee is logged in to the web application;
When: the employee requests to see data on employees;
Then: the request gets processed and executes successfully
given the employee logged in with the right credentials.
| priority | employee can authenticate to system commented on as a live thrive employee i would like the database to be secure so only authenticated users can access the data scenario an authenticated user attempts to access data given an employee is logged in to the web application when the employee requests to see data on employees then the request gets processed and executes successfully given the employee logged in with the right credentials | 1 |
404,135 | 11,852,589,782 | IssuesEvent | 2020-03-24 20:14:35 | perfectsense/gyro | https://api.github.com/repos/perfectsense/gyro | closed | Gyro does not suspend creation of a resource when there is a dependency | bug priority:high | **Describe the bug**
Gyro does not suspend the creation of azure::network sub-resources when they're dependent on another resource. The following is an failed gyro up:
```
ezapata@admins-MacBook-Pro ~/Documents/projects/ops-test/ops/gyro-azure master ● gyro up
↓ Loading plugin: gyro:gyro-azure-provider:0.99.1-SNAPSHOT
↓ Loading plugin: gyro:gyro-ssh-plugin:0.99.1-SNAPSHOT
⟳ Refreshed resources: 0
Looking for changes...
+ Create azure::dns-zone public-zone
+ Create azure::resource-group resource-group (ops-test-v1)
+ Create azure::network network
+ Create subnet private
+ Create subnet public
+ Create subnet lower
+ Create azure::route-table public
+ Create route internet
+ Create azure::route-table private
+ Create route nat
+ Create azure::route-table lower
+ Create route internet
Are you sure you want to change resources? (y/N) y
+ Creating azure::dns-zone public-zone OK
+ Creating azure::resource-group resource-group (ops-test-v1) OK
+ Creating azure::network network OK
+ Creating subnet private for azure::network network
Error: Can't create subnet private resource!
```
**To Reproduce**
Run a gyro up with the following configs.
network.gyro:
```
azure::network network
name: "Network-test"
resource-group: $(azure::resource-group resource-group)
address-spaces: [
"10.0.0.0/16"
]
subnet
address-prefix: 10.0.0.0/26
name: "private"
route-table: $(azure::route-table private)
end
end
```
routes.gyro:
```
azure::route-table private
bgp-route-propagation-disabled: true
resource-group: $(azure::resource-group resource-group)
route
destination-address-prefix: "0.0.0.0/0"
name: "nat"
next-hop-type: "VirtualAppliance"
next-hop-ip-address: '10.0.0.10'
end
end
```
**Expected behavior**
The creation of the network resource should be done after the route-table has been created.
**Stacktrace**
```
Caused by: Unexpected error: com.microsoft.azure.CloudException: Cannot parse the request.: Cannot parse the request.
at com.microsoft.azure.AzureClient.createExceptionFromResponse(AzureClient.java:740)
at com.microsoft.azure.AzureClient.access$100(AzureClient.java:33)
at com.microsoft.azure.AzureClient$3.call(AzureClient.java:160)
at com.microsoft.azure.AzureClient$3.call(AzureClient.java:157)
at rx.internal.operators.OnSubscribeMap$MapSubscriber.onNext(OnSubscribeMap.java:69)
at retrofit2.adapter.rxjava.CallArbiter.deliverResponse(CallArbiter.java:120)
at retrofit2.adapter.rxjava.CallArbiter.emitResponse(CallArbiter.java:102)
at retrofit2.adapter.rxjava.CallExecuteOnSubscribe.call(CallExecuteOnSubscribe.java:46)
at retrofit2.adapter.rxjava.CallExecuteOnSubscribe.call(CallExecuteOnSubscribe.java:24)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeSingle.call(OnSubscribeSingle.java:81)
at rx.internal.operators.OnSubscribeSingle.call(OnSubscribeSingle.java:27)
at rx.internal.operators.SingleToObservable.call(SingleToObservable.java:39)
at rx.internal.operators.SingleToObservable.call(SingleToObservable.java:27)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:48)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:30)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.DeferredScalarSubscriber.subscribeTo(DeferredScalarSubscriber.java:153)
at rx.internal.operators.OnSubscribeTakeLastOne.call(OnSubscribeTakeLastOne.java:32)
at rx.internal.operators.OnSubscribeTakeLastOne.call(OnSubscribeTakeLastOne.java:22)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:48)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:30)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:48)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:30)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OperatorSubscribeOn$SubscribeOnSubscriber.call(OperatorSubscribeOn.java:100)
at rx.internal.schedulers.CachedThreadScheduler$EventLoopWorker$1.call(CachedThreadScheduler.java:230)
at rx.internal.schedulers.ScheduledAction.run(ScheduledAction.java:55)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
Caused by: rx.exceptions.OnErrorThrowable$OnNextValue: OnError while emitting onNext value: retrofit2.Response.class
at rx.exceptions.OnErrorThrowable.addValueAsLastCause(OnErrorThrowable.java:118)
at rx.internal.operators.OnSubscribeMap$MapSubscriber.onNext(OnSubscribeMap.java:73)
... 47 more
``` | 1.0 | Gyro does not suspend creation of a resource when there is a dependency - **Describe the bug**
Gyro does not suspend the creation of azure::network sub-resources when they're dependent on another resource. The following is an failed gyro up:
```
ezapata@admins-MacBook-Pro ~/Documents/projects/ops-test/ops/gyro-azure master ● gyro up
↓ Loading plugin: gyro:gyro-azure-provider:0.99.1-SNAPSHOT
↓ Loading plugin: gyro:gyro-ssh-plugin:0.99.1-SNAPSHOT
⟳ Refreshed resources: 0
Looking for changes...
+ Create azure::dns-zone public-zone
+ Create azure::resource-group resource-group (ops-test-v1)
+ Create azure::network network
+ Create subnet private
+ Create subnet public
+ Create subnet lower
+ Create azure::route-table public
+ Create route internet
+ Create azure::route-table private
+ Create route nat
+ Create azure::route-table lower
+ Create route internet
Are you sure you want to change resources? (y/N) y
+ Creating azure::dns-zone public-zone OK
+ Creating azure::resource-group resource-group (ops-test-v1) OK
+ Creating azure::network network OK
+ Creating subnet private for azure::network network
Error: Can't create subnet private resource!
```
**To Reproduce**
Run a gyro up with the following configs.
network.gyro:
```
azure::network network
name: "Network-test"
resource-group: $(azure::resource-group resource-group)
address-spaces: [
"10.0.0.0/16"
]
subnet
address-prefix: 10.0.0.0/26
name: "private"
route-table: $(azure::route-table private)
end
end
```
routes.gyro:
```
azure::route-table private
bgp-route-propagation-disabled: true
resource-group: $(azure::resource-group resource-group)
route
destination-address-prefix: "0.0.0.0/0"
name: "nat"
next-hop-type: "VirtualAppliance"
next-hop-ip-address: '10.0.0.10'
end
end
```
**Expected behavior**
The creation of the network resource should be done after the route-table has been created.
**Stacktrace**
```
Caused by: Unexpected error: com.microsoft.azure.CloudException: Cannot parse the request.: Cannot parse the request.
at com.microsoft.azure.AzureClient.createExceptionFromResponse(AzureClient.java:740)
at com.microsoft.azure.AzureClient.access$100(AzureClient.java:33)
at com.microsoft.azure.AzureClient$3.call(AzureClient.java:160)
at com.microsoft.azure.AzureClient$3.call(AzureClient.java:157)
at rx.internal.operators.OnSubscribeMap$MapSubscriber.onNext(OnSubscribeMap.java:69)
at retrofit2.adapter.rxjava.CallArbiter.deliverResponse(CallArbiter.java:120)
at retrofit2.adapter.rxjava.CallArbiter.emitResponse(CallArbiter.java:102)
at retrofit2.adapter.rxjava.CallExecuteOnSubscribe.call(CallExecuteOnSubscribe.java:46)
at retrofit2.adapter.rxjava.CallExecuteOnSubscribe.call(CallExecuteOnSubscribe.java:24)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeSingle.call(OnSubscribeSingle.java:81)
at rx.internal.operators.OnSubscribeSingle.call(OnSubscribeSingle.java:27)
at rx.internal.operators.SingleToObservable.call(SingleToObservable.java:39)
at rx.internal.operators.SingleToObservable.call(SingleToObservable.java:27)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:48)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:30)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.DeferredScalarSubscriber.subscribeTo(DeferredScalarSubscriber.java:153)
at rx.internal.operators.OnSubscribeTakeLastOne.call(OnSubscribeTakeLastOne.java:32)
at rx.internal.operators.OnSubscribeTakeLastOne.call(OnSubscribeTakeLastOne.java:22)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:48)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:30)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:48)
at rx.internal.operators.OnSubscribeMap.call(OnSubscribeMap.java:33)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:48)
at rx.internal.operators.OnSubscribeLift.call(OnSubscribeLift.java:30)
at rx.Observable.unsafeSubscribe(Observable.java:10327)
at rx.internal.operators.OperatorSubscribeOn$SubscribeOnSubscriber.call(OperatorSubscribeOn.java:100)
at rx.internal.schedulers.CachedThreadScheduler$EventLoopWorker$1.call(CachedThreadScheduler.java:230)
at rx.internal.schedulers.ScheduledAction.run(ScheduledAction.java:55)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
Caused by: rx.exceptions.OnErrorThrowable$OnNextValue: OnError while emitting onNext value: retrofit2.Response.class
at rx.exceptions.OnErrorThrowable.addValueAsLastCause(OnErrorThrowable.java:118)
at rx.internal.operators.OnSubscribeMap$MapSubscriber.onNext(OnSubscribeMap.java:73)
... 47 more
``` | priority | gyro does not suspend creation of a resource when there is a dependency describe the bug gyro does not suspend the creation of azure network sub resources when they re dependent on another resource the following is an failed gyro up ezapata admins macbook pro documents projects ops test ops gyro azure master ● gyro up ↓ loading plugin gyro gyro azure provider snapshot ↓ loading plugin gyro gyro ssh plugin snapshot ⟳ refreshed resources looking for changes create azure dns zone public zone create azure resource group resource group ops test create azure network network create subnet private create subnet public create subnet lower create azure route table public create route internet create azure route table private create route nat create azure route table lower create route internet are you sure you want to change resources y n y creating azure dns zone public zone ok creating azure resource group resource group ops test ok creating azure network network ok creating subnet private for azure network network error can t create subnet private resource to reproduce run a gyro up with the following configs network gyro azure network network name network test resource group azure resource group resource group address spaces subnet address prefix name private route table azure route table private end end routes gyro azure route table private bgp route propagation disabled true resource group azure resource group resource group route destination address prefix name nat next hop type virtualappliance next hop ip address end end expected behavior the creation of the network resource should be done after the route table has been created stacktrace caused by unexpected error com microsoft azure cloudexception cannot parse the request cannot parse the request at com microsoft azure azureclient createexceptionfromresponse azureclient java at com microsoft azure azureclient access azureclient java at com microsoft azure azureclient call azureclient java at com microsoft azure azureclient call azureclient java at rx internal operators onsubscribemap mapsubscriber onnext onsubscribemap java at adapter rxjava callarbiter deliverresponse callarbiter java at adapter rxjava callarbiter emitresponse callarbiter java at adapter rxjava callexecuteonsubscribe call callexecuteonsubscribe java at adapter rxjava callexecuteonsubscribe call callexecuteonsubscribe java at rx observable unsafesubscribe observable java at rx internal operators onsubscribemap call onsubscribemap java at rx internal operators onsubscribemap call onsubscribemap java at rx observable unsafesubscribe observable java at rx internal operators onsubscribesingle call onsubscribesingle java at rx internal operators onsubscribesingle call onsubscribesingle java at rx internal operators singletoobservable call singletoobservable java at rx internal operators singletoobservable call singletoobservable java at rx observable unsafesubscribe observable java at rx internal operators onsubscribemap call onsubscribemap java at rx internal operators onsubscribemap call onsubscribemap java at rx internal operators onsubscribelift call onsubscribelift java at rx internal operators onsubscribelift call onsubscribelift java at rx observable unsafesubscribe observable java at rx internal operators deferredscalarsubscriber subscribeto deferredscalarsubscriber java at rx internal operators onsubscribetakelastone call onsubscribetakelastone java at rx internal operators onsubscribetakelastone call onsubscribetakelastone java at rx internal operators onsubscribelift call onsubscribelift java at rx internal operators onsubscribelift call onsubscribelift java at rx observable unsafesubscribe observable java at rx internal operators onsubscribemap call onsubscribemap java at rx internal operators onsubscribemap call onsubscribemap java at rx observable unsafesubscribe observable java at rx internal operators onsubscribemap call onsubscribemap java at rx internal operators onsubscribemap call onsubscribemap java at rx observable unsafesubscribe observable java at rx internal operators onsubscribemap call onsubscribemap java at rx internal operators onsubscribemap call onsubscribemap java at rx observable unsafesubscribe observable java at rx internal operators onsubscribemap call onsubscribemap java at rx internal operators onsubscribemap call onsubscribemap java at rx internal operators onsubscribelift call onsubscribelift java at rx internal operators onsubscribelift call onsubscribelift java at rx observable unsafesubscribe observable java at rx internal operators operatorsubscribeon subscribeonsubscriber call operatorsubscribeon java at rx internal schedulers cachedthreadscheduler eventloopworker call cachedthreadscheduler java at rx internal schedulers scheduledaction run scheduledaction java at java base java util concurrent executors runnableadapter call executors java at java base java util concurrent futuretask run futuretask java at java base java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by rx exceptions onerrorthrowable onnextvalue onerror while emitting onnext value response class at rx exceptions onerrorthrowable addvalueaslastcause onerrorthrowable java at rx internal operators onsubscribemap mapsubscriber onnext onsubscribemap java more | 1 |
375,506 | 11,104,675,405 | IssuesEvent | 2019-12-17 08:11:45 | wso2/ballerina-message-broker | https://api.github.com/repos/wso2/ballerina-message-broker | closed | Add user doc on Admin services | Complexity/Novice Priority/High Severity/Minor Type/Docs | ### Description
There should be a document explaining exposed admin services and related configurations. | 1.0 | Add user doc on Admin services - ### Description
There should be a document explaining exposed admin services and related configurations. | priority | add user doc on admin services description there should be a document explaining exposed admin services and related configurations | 1 |
121,919 | 4,822,998,866 | IssuesEvent | 2016-11-06 04:48:17 | CS2103AUG2016-W14-C2/main | https://api.github.com/repos/CS2103AUG2016-W14-C2/main | closed | UG, DG to include floatingTask is able to change to Task, Vice Versa. | priority.high | The command for FT to T would be, e.g : edit floating 1 by: tmr 7pm
The command for T to FT would be, e.g : edit task 1 to: floating | 1.0 | UG, DG to include floatingTask is able to change to Task, Vice Versa. - The command for FT to T would be, e.g : edit floating 1 by: tmr 7pm
The command for T to FT would be, e.g : edit task 1 to: floating | priority | ug dg to include floatingtask is able to change to task vice versa the command for ft to t would be e g edit floating by tmr the command for t to ft would be e g edit task to floating | 1 |
52,671 | 3,026,104,324 | IssuesEvent | 2015-08-03 13:20:28 | theodi/member-directory | https://api.github.com/repos/theodi/member-directory | opened | Invoice payments coming up paid in Xero | bug priority: high xero | For people who are paying for membership via invoice we are still getting the invoices displaying in Xero as marked as paid e.g.
ODI Supporter (GS660BY) Non Commercial
PAID BY CREDIT CARD; reference
The 'PAID BY CREDIT CARD; reference' needs to be removed from all invoice payments otherwise Finance team won't know to send the invoices out. | 1.0 | Invoice payments coming up paid in Xero - For people who are paying for membership via invoice we are still getting the invoices displaying in Xero as marked as paid e.g.
ODI Supporter (GS660BY) Non Commercial
PAID BY CREDIT CARD; reference
The 'PAID BY CREDIT CARD; reference' needs to be removed from all invoice payments otherwise Finance team won't know to send the invoices out. | priority | invoice payments coming up paid in xero for people who are paying for membership via invoice we are still getting the invoices displaying in xero as marked as paid e g odi supporter non commercial paid by credit card reference the paid by credit card reference needs to be removed from all invoice payments otherwise finance team won t know to send the invoices out | 1 |
306,880 | 9,412,482,752 | IssuesEvent | 2019-04-10 04:16:14 | wso2/product-ei | https://api.github.com/repos/wso2/product-ei | closed | "Payload could not be written as JSON" error when using call mediator in blocking mode | Priority/Highest Type/Bug | **Description:**
Error when call mediator used in blocking mode
**Affected Product Version:**
EI 6.1.1 (wso2ei-6.1.1.1512875284190)
**OS, DB, other environment details and versions:**
OS - ubuntu 15.10
DB - H2
**Steps to reproduce:**
1. create below API
```
<api xmlns="http://ws.apache.org/ns/synapse" name="Backend4" context="/userinfomation">
<resource methods="POST">
<inSequence>
<call blocking="true">
<endpoint>
<http uri-template="http://www.mocky.io/v2/5a6851eb2e00002509d5b303/userinfo"/>
</endpoint>
</call>
<respond/>
</inSequence>
</resource>
</api>
```
Observations
```
[2018-01-25 17:26:00,529] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "POST /userinfomation HTTP/1.1[\r][\n]"
[2018-01-25 17:26:00,529] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "Host: 10.100.8.9:8280[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "User-Agent: curl/7.43.0[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "Accept: */*[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "Content-Type: application/json[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "Content-Length: 11[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "{"a" : "b"}"
[2018-01-25 17:26:00,981] [EI-Core] ERROR - JsonUtil #writeAsJson. Payload could not be written as JSON. MessageID: urn:uuid:3f796012-2551-4a9c-a034-0d1365b3d60c
[2018-01-25 17:26:01,128] [EI-Core] ERROR - JsonUtil #writeAsJson. Payload could not be written as JSON. MessageID: urn:uuid:3f796012-2551-4a9c-a034-0d1365b3d60c
[2018-01-25 17:26:01,282] [EI-Core] ERROR - JsonUtil #writeAsJson. Payload could not be written as JSON. MessageID: urn:uuid:3f796012-2551-4a9c-a034-0d1365b3d60c
[2018-01-25 17:26:01,433] [EI-Core] ERROR - JsonUtil #writeAsJson. Payload could not be written as JSON. MessageID: urn:uuid:3f796012-2551-4a9c-a034-0d1365b3d60c
[2018-01-25 17:26:01,435] [EI-Core] INFO - HTTPSender Unable to sendViaPost to url[http://www.mocky.io/v2/5a6851eb2e00002509d5b303/userinfo]
org.apache.axis2.AxisFault: Payload could not be written as JSON.
at org.apache.synapse.commons.json.JsonUtil.writeAsJson(JsonUtil.java:289)
at org.apache.synapse.commons.json.JsonStreamFormatter.writeTo(JsonStreamFormatter.java:70)
at org.apache.axis2.transport.http.AxisRequestEntity.writeRequest(AxisRequestEntity.java:84)
at org.apache.commons.httpclient.methods.EntityEnclosingMethod.writeRequestBody(EntityEnclosingMethod.java:499)
at org.apache.commons.httpclient.HttpMethodBase.writeRequest(HttpMethodBase.java:2114)
at org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1096)
at org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398)
at org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171)
at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397)
at org.apache.axis2.transport.http.AbstractHTTPSender.executeMethod(AbstractHTTPSender.java:682)
at org.apache.axis2.transport.http.HTTPSender.sendViaPost(HTTPSender.java:195)
at org.apache.axis2.transport.http.HTTPSender.send(HTTPSender.java:77)
at org.apache.axis2.transport.http.CommonsHTTPTransportSender.writeMessageWithCommons(CommonsHTTPTransportSender.java:451)
at org.apache.axis2.transport.http.CommonsHTTPTransportSender.invoke(CommonsHTTPTransportSender.java:278)
at org.apache.axis2.engine.AxisEngine.send(AxisEngine.java:442)
at org.apache.axis2.description.OutInAxisOperationClient.send(OutInAxisOperation.java:430)
at org.apache.axis2.description.OutInAxisOperationClient.executeImpl(OutInAxisOperation.java:225)
at org.apache.axis2.client.OperationClient.execute(OperationClient.java:149)
at org.apache.synapse.message.senders.blocking.BlockingMsgSender.sendReceive(BlockingMsgSender.java:302)
at org.apache.synapse.message.senders.blocking.BlockingMsgSender.send(BlockingMsgSender.java:211)
at org.apache.synapse.mediators.builtin.CallMediator.handleBlockingCall(CallMediator.java:150)
at org.apache.synapse.mediators.builtin.CallMediator.mediate(CallMediator.java:113)
at org.apache.synapse.mediators.AbstractListMediator.mediate(AbstractListMediator.java:97)
at org.apache.synapse.mediators.AbstractListMediator.mediate(AbstractListMediator.java:59)
at org.apache.synapse.mediators.base.SequenceMediator.mediate(SequenceMediator.java:158)
at org.apache.synapse.rest.Resource.process(Resource.java:343)
at org.apache.synapse.rest.API.process(API.java:399)
at org.apache.synapse.rest.RESTRequestHandler.apiProcess(RESTRequestHandler.java:123)
at org.apache.synapse.rest.RESTRequestHandler.dispatchToAPI(RESTRequestHandler.java:101)
at org.apache.synapse.rest.RESTRequestHandler.process(RESTRequestHandler.java:69)
at org.apache.synapse.core.axis2.Axis2SynapseEnvironment.injectMessage(Axis2SynapseEnvironment.java:304)
at org.apache.synapse.core.axis2.SynapseMessageReceiver.receive(SynapseMessageReceiver.java:78)
at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180)
at org.apache.synapse.transport.passthru.ServerWorker.processNonEntityEnclosingRESTHandler(ServerWorker.java:326)
at org.apache.synapse.transport.passthru.ServerWorker.processEntityEnclosingRequest(ServerWorker.java:372)
at org.apache.synapse.transport.passthru.ServerWorker.run(ServerWorker.java:151)
at org.apache.axis2.transport.base.threads.NativeWorkerPool$1.run(NativeWorkerPool.java:172)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
[2018-01-25 17:26:01,460] [EI-Core] INFO - LogMediator To: http://www.mocky.io/v2/5a6851eb2e00002509d5b303/userinfo, MessageID: urn:uuid:1de7fc5b-8982-4507-bc9b-1f6aefb6f1e7, Direction: request, MESSAGE = Executing default 'fault' sequence, ERROR_CODE = 401001, ERROR_MESSAGE = Payload could not be written as JSON., Payload: {"a" : "b"}
[2018-01-25 17:26:01,462] [EI-Core] WARN - PassThroughTransportUtils Unable to set the HTTP status code from the property HTTP_SC with value:
[2018-01-25 17:26:01,467] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "HTTP/1.1 202 Accepted[\r][\n]"
[2018-01-25 17:26:01,467] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "Date: Thu, 25 Jan 2018 11:56:01 GMT[\r][\n]"
[2018-01-25 17:26:01,468] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "Transfer-Encoding: chunked[\r][\n]"
[2018-01-25 17:26:01,468] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "[\r][\n]"
[2018-01-25 17:26:01,468] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "0[\r][\n]"
[2018-01-25 17:26:01,468] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "[\r][\n]"
```
Note - If we add a log full this issue will not get reproduced | 1.0 | "Payload could not be written as JSON" error when using call mediator in blocking mode - **Description:**
Error when call mediator used in blocking mode
**Affected Product Version:**
EI 6.1.1 (wso2ei-6.1.1.1512875284190)
**OS, DB, other environment details and versions:**
OS - ubuntu 15.10
DB - H2
**Steps to reproduce:**
1. create below API
```
<api xmlns="http://ws.apache.org/ns/synapse" name="Backend4" context="/userinfomation">
<resource methods="POST">
<inSequence>
<call blocking="true">
<endpoint>
<http uri-template="http://www.mocky.io/v2/5a6851eb2e00002509d5b303/userinfo"/>
</endpoint>
</call>
<respond/>
</inSequence>
</resource>
</api>
```
Observations
```
[2018-01-25 17:26:00,529] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "POST /userinfomation HTTP/1.1[\r][\n]"
[2018-01-25 17:26:00,529] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "Host: 10.100.8.9:8280[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "User-Agent: curl/7.43.0[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "Accept: */*[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "Content-Type: application/json[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "Content-Length: 11[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "[\r][\n]"
[2018-01-25 17:26:00,530] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 >> "{"a" : "b"}"
[2018-01-25 17:26:00,981] [EI-Core] ERROR - JsonUtil #writeAsJson. Payload could not be written as JSON. MessageID: urn:uuid:3f796012-2551-4a9c-a034-0d1365b3d60c
[2018-01-25 17:26:01,128] [EI-Core] ERROR - JsonUtil #writeAsJson. Payload could not be written as JSON. MessageID: urn:uuid:3f796012-2551-4a9c-a034-0d1365b3d60c
[2018-01-25 17:26:01,282] [EI-Core] ERROR - JsonUtil #writeAsJson. Payload could not be written as JSON. MessageID: urn:uuid:3f796012-2551-4a9c-a034-0d1365b3d60c
[2018-01-25 17:26:01,433] [EI-Core] ERROR - JsonUtil #writeAsJson. Payload could not be written as JSON. MessageID: urn:uuid:3f796012-2551-4a9c-a034-0d1365b3d60c
[2018-01-25 17:26:01,435] [EI-Core] INFO - HTTPSender Unable to sendViaPost to url[http://www.mocky.io/v2/5a6851eb2e00002509d5b303/userinfo]
org.apache.axis2.AxisFault: Payload could not be written as JSON.
at org.apache.synapse.commons.json.JsonUtil.writeAsJson(JsonUtil.java:289)
at org.apache.synapse.commons.json.JsonStreamFormatter.writeTo(JsonStreamFormatter.java:70)
at org.apache.axis2.transport.http.AxisRequestEntity.writeRequest(AxisRequestEntity.java:84)
at org.apache.commons.httpclient.methods.EntityEnclosingMethod.writeRequestBody(EntityEnclosingMethod.java:499)
at org.apache.commons.httpclient.HttpMethodBase.writeRequest(HttpMethodBase.java:2114)
at org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1096)
at org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398)
at org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171)
at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397)
at org.apache.axis2.transport.http.AbstractHTTPSender.executeMethod(AbstractHTTPSender.java:682)
at org.apache.axis2.transport.http.HTTPSender.sendViaPost(HTTPSender.java:195)
at org.apache.axis2.transport.http.HTTPSender.send(HTTPSender.java:77)
at org.apache.axis2.transport.http.CommonsHTTPTransportSender.writeMessageWithCommons(CommonsHTTPTransportSender.java:451)
at org.apache.axis2.transport.http.CommonsHTTPTransportSender.invoke(CommonsHTTPTransportSender.java:278)
at org.apache.axis2.engine.AxisEngine.send(AxisEngine.java:442)
at org.apache.axis2.description.OutInAxisOperationClient.send(OutInAxisOperation.java:430)
at org.apache.axis2.description.OutInAxisOperationClient.executeImpl(OutInAxisOperation.java:225)
at org.apache.axis2.client.OperationClient.execute(OperationClient.java:149)
at org.apache.synapse.message.senders.blocking.BlockingMsgSender.sendReceive(BlockingMsgSender.java:302)
at org.apache.synapse.message.senders.blocking.BlockingMsgSender.send(BlockingMsgSender.java:211)
at org.apache.synapse.mediators.builtin.CallMediator.handleBlockingCall(CallMediator.java:150)
at org.apache.synapse.mediators.builtin.CallMediator.mediate(CallMediator.java:113)
at org.apache.synapse.mediators.AbstractListMediator.mediate(AbstractListMediator.java:97)
at org.apache.synapse.mediators.AbstractListMediator.mediate(AbstractListMediator.java:59)
at org.apache.synapse.mediators.base.SequenceMediator.mediate(SequenceMediator.java:158)
at org.apache.synapse.rest.Resource.process(Resource.java:343)
at org.apache.synapse.rest.API.process(API.java:399)
at org.apache.synapse.rest.RESTRequestHandler.apiProcess(RESTRequestHandler.java:123)
at org.apache.synapse.rest.RESTRequestHandler.dispatchToAPI(RESTRequestHandler.java:101)
at org.apache.synapse.rest.RESTRequestHandler.process(RESTRequestHandler.java:69)
at org.apache.synapse.core.axis2.Axis2SynapseEnvironment.injectMessage(Axis2SynapseEnvironment.java:304)
at org.apache.synapse.core.axis2.SynapseMessageReceiver.receive(SynapseMessageReceiver.java:78)
at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180)
at org.apache.synapse.transport.passthru.ServerWorker.processNonEntityEnclosingRESTHandler(ServerWorker.java:326)
at org.apache.synapse.transport.passthru.ServerWorker.processEntityEnclosingRequest(ServerWorker.java:372)
at org.apache.synapse.transport.passthru.ServerWorker.run(ServerWorker.java:151)
at org.apache.axis2.transport.base.threads.NativeWorkerPool$1.run(NativeWorkerPool.java:172)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
[2018-01-25 17:26:01,460] [EI-Core] INFO - LogMediator To: http://www.mocky.io/v2/5a6851eb2e00002509d5b303/userinfo, MessageID: urn:uuid:1de7fc5b-8982-4507-bc9b-1f6aefb6f1e7, Direction: request, MESSAGE = Executing default 'fault' sequence, ERROR_CODE = 401001, ERROR_MESSAGE = Payload could not be written as JSON., Payload: {"a" : "b"}
[2018-01-25 17:26:01,462] [EI-Core] WARN - PassThroughTransportUtils Unable to set the HTTP status code from the property HTTP_SC with value:
[2018-01-25 17:26:01,467] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "HTTP/1.1 202 Accepted[\r][\n]"
[2018-01-25 17:26:01,467] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "Date: Thu, 25 Jan 2018 11:56:01 GMT[\r][\n]"
[2018-01-25 17:26:01,468] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "Transfer-Encoding: chunked[\r][\n]"
[2018-01-25 17:26:01,468] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "[\r][\n]"
[2018-01-25 17:26:01,468] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "0[\r][\n]"
[2018-01-25 17:26:01,468] [EI-Core] DEBUG - wire HTTP-Listener I/O dispatcher-1 << "[\r][\n]"
```
Note - If we add a log full this issue will not get reproduced | priority | payload could not be written as json error when using call mediator in blocking mode description error when call mediator used in blocking mode affected product version ei os db other environment details and versions os ubuntu db steps to reproduce create below api http uri template observations debug wire http listener i o dispatcher post userinfomation http debug wire http listener i o dispatcher host debug wire http listener i o dispatcher user agent curl debug wire http listener i o dispatcher accept debug wire http listener i o dispatcher content type application json debug wire http listener i o dispatcher content length debug wire http listener i o dispatcher debug wire http listener i o dispatcher a b error jsonutil writeasjson payload could not be written as json messageid urn uuid error jsonutil writeasjson payload could not be written as json messageid urn uuid error jsonutil writeasjson payload could not be written as json messageid urn uuid error jsonutil writeasjson payload could not be written as json messageid urn uuid info httpsender unable to sendviapost to url org apache axisfault payload could not be written as json at org apache synapse commons json jsonutil writeasjson jsonutil java at org apache synapse commons json jsonstreamformatter writeto jsonstreamformatter java at org apache transport http axisrequestentity writerequest axisrequestentity java at org apache commons httpclient methods entityenclosingmethod writerequestbody entityenclosingmethod java at org apache commons httpclient httpmethodbase writerequest httpmethodbase java at org apache commons httpclient httpmethodbase execute httpmethodbase java at org apache commons httpclient httpmethoddirector executewithretry httpmethoddirector java at org apache commons httpclient httpmethoddirector executemethod httpmethoddirector java at org apache commons httpclient httpclient executemethod httpclient java at org apache transport http abstracthttpsender executemethod abstracthttpsender java at org apache transport http httpsender sendviapost httpsender java at org apache transport http httpsender send httpsender java at org apache transport http commonshttptransportsender writemessagewithcommons commonshttptransportsender java at org apache transport http commonshttptransportsender invoke commonshttptransportsender java at org apache engine axisengine send axisengine java at org apache description outinaxisoperationclient send outinaxisoperation java at org apache description outinaxisoperationclient executeimpl outinaxisoperation java at org apache client operationclient execute operationclient java at org apache synapse message senders blocking blockingmsgsender sendreceive blockingmsgsender java at org apache synapse message senders blocking blockingmsgsender send blockingmsgsender java at org apache synapse mediators builtin callmediator handleblockingcall callmediator java at org apache synapse mediators builtin callmediator mediate callmediator java at org apache synapse mediators abstractlistmediator mediate abstractlistmediator java at org apache synapse mediators abstractlistmediator mediate abstractlistmediator java at org apache synapse mediators base sequencemediator mediate sequencemediator java at org apache synapse rest resource process resource java at org apache synapse rest api process api java at org apache synapse rest restrequesthandler apiprocess restrequesthandler java at org apache synapse rest restrequesthandler dispatchtoapi restrequesthandler java at org apache synapse rest restrequesthandler process restrequesthandler java at org apache synapse core injectmessage java at org apache synapse core synapsemessagereceiver receive synapsemessagereceiver java at org apache engine axisengine receive axisengine java at org apache synapse transport passthru serverworker processnonentityenclosingresthandler serverworker java at org apache synapse transport passthru serverworker processentityenclosingrequest serverworker java at org apache synapse transport passthru serverworker run serverworker java at org apache transport base threads nativeworkerpool run nativeworkerpool java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java info logmediator to messageid urn uuid direction request message executing default fault sequence error code error message payload could not be written as json payload a b warn passthroughtransportutils unable to set the http status code from the property http sc with value debug wire http listener i o dispatcher http accepted debug wire http listener i o dispatcher date thu jan gmt debug wire http listener i o dispatcher transfer encoding chunked debug wire http listener i o dispatcher debug wire http listener i o dispatcher debug wire http listener i o dispatcher note if we add a log full this issue will not get reproduced | 1 |
260,270 | 8,205,709,028 | IssuesEvent | 2018-09-03 11:03:35 | hpcugent/vsc_user_docs | https://api.github.com/repos/hpcugent/vsc_user_docs | closed | troubleshooting: memory limits | Jasper (HPC-UGent student intern) priority:high | * see http://hpc.ugent.be/userwiki/index.php/User:Mem_limits
* memory limits that are in place on login nodes and in submitted jobs can manifest in different ways
* `Out of memory` errors
* segmentation faults & bus errors
* `Could not allocate memory` errors
* `Insufficient memory` errors
* explain how to check available memory (see `ulimit` output?)
* refer to relevant sections in manual, in particular how to request memory in jobs scripts
* also check whether default is explained well, i.e. how much memory is requested by default b (probably specific to HPC-UGent?)
* Java-specific stuff can be added in a subsection, but should be moved out to software-specific part of documentation when it's in place
related issue: #126 | 1.0 | troubleshooting: memory limits - * see http://hpc.ugent.be/userwiki/index.php/User:Mem_limits
* memory limits that are in place on login nodes and in submitted jobs can manifest in different ways
* `Out of memory` errors
* segmentation faults & bus errors
* `Could not allocate memory` errors
* `Insufficient memory` errors
* explain how to check available memory (see `ulimit` output?)
* refer to relevant sections in manual, in particular how to request memory in jobs scripts
* also check whether default is explained well, i.e. how much memory is requested by default b (probably specific to HPC-UGent?)
* Java-specific stuff can be added in a subsection, but should be moved out to software-specific part of documentation when it's in place
related issue: #126 | priority | troubleshooting memory limits see memory limits that are in place on login nodes and in submitted jobs can manifest in different ways out of memory errors segmentation faults bus errors could not allocate memory errors insufficient memory errors explain how to check available memory see ulimit output refer to relevant sections in manual in particular how to request memory in jobs scripts also check whether default is explained well i e how much memory is requested by default b probably specific to hpc ugent java specific stuff can be added in a subsection but should be moved out to software specific part of documentation when it s in place related issue | 1 |
321,113 | 9,793,707,438 | IssuesEvent | 2019-06-10 20:38:57 | infor-design/enterprise-ng | https://api.github.com/repos/infor-design/enterprise-ng | closed | XSS in soho-dropdown | [3] priority: high type: bug :bug: type: patch | **Describe the bug**
Clicking an entry with a <script> tag in a dropdown causes the script to execute. This only seem to happen when using `ids-enterprise-ng@5.2.1` and `ids-enterprise@4.18.1`.
**To Reproduce**
Given the followin component:
```html
<select soho-dropdown [(ngModel)]="selected">
<option *ngFor="let option of options" [value]="option">{{option}}</option>
</select>
```
```typescript
@Component({
selector: 'app-dropdown-xss',
templateUrl: './dropdown-xss.component.html',
styleUrls: ['./dropdown-xss.component.css']
})
export class DropdownXssComponent {
selected = 'Hello';
options = [
'Hello',
`<script>window.alert('dropdown xss')</script>XSS`,
'World',
];
}
```
Steps to reproduce the behavior:
1. Open the dropdown
2. Click the entry with XSS
3. Script executes
**Expected behavior**
The script should not execute.
**Version**
- ids-enterprise-ng: 5.2.1
- ids-enterprise: 4.18.1
**Platform**
Tested in Chrome on Windows.
**Additional context**
This only happens when `ids-enterprise@4.18.1` is installed, i.e not the bundled 4.17.1.
There might be other XSS vulnerabilities in 4.18.1, since I managed to get the script to execute again when it was bound with `ngModel` to an `<input>`.
| 1.0 | XSS in soho-dropdown - **Describe the bug**
Clicking an entry with a <script> tag in a dropdown causes the script to execute. This only seem to happen when using `ids-enterprise-ng@5.2.1` and `ids-enterprise@4.18.1`.
**To Reproduce**
Given the followin component:
```html
<select soho-dropdown [(ngModel)]="selected">
<option *ngFor="let option of options" [value]="option">{{option}}</option>
</select>
```
```typescript
@Component({
selector: 'app-dropdown-xss',
templateUrl: './dropdown-xss.component.html',
styleUrls: ['./dropdown-xss.component.css']
})
export class DropdownXssComponent {
selected = 'Hello';
options = [
'Hello',
`<script>window.alert('dropdown xss')</script>XSS`,
'World',
];
}
```
Steps to reproduce the behavior:
1. Open the dropdown
2. Click the entry with XSS
3. Script executes
**Expected behavior**
The script should not execute.
**Version**
- ids-enterprise-ng: 5.2.1
- ids-enterprise: 4.18.1
**Platform**
Tested in Chrome on Windows.
**Additional context**
This only happens when `ids-enterprise@4.18.1` is installed, i.e not the bundled 4.17.1.
There might be other XSS vulnerabilities in 4.18.1, since I managed to get the script to execute again when it was bound with `ngModel` to an `<input>`.
| priority | xss in soho dropdown describe the bug clicking an entry with a tag in a dropdown causes the script to execute this only seem to happen when using ids enterprise ng and ids enterprise to reproduce given the followin component html option typescript component selector app dropdown xss templateurl dropdown xss component html styleurls export class dropdownxsscomponent selected hello options hello window alert dropdown xss xss world steps to reproduce the behavior open the dropdown click the entry with xss script executes expected behavior the script should not execute version ids enterprise ng ids enterprise platform tested in chrome on windows additional context this only happens when ids enterprise is installed i e not the bundled there might be other xss vulnerabilities in since i managed to get the script to execute again when it was bound with ngmodel to an | 1 |
158,909 | 6,037,018,086 | IssuesEvent | 2017-06-09 17:36:16 | champagneabuelo/the-material-exchange | https://api.github.com/repos/champagneabuelo/the-material-exchange | opened | Implement sign-up page for nonauthenticated users | high priority | - for users that are not logged in, there can be a sign up/login element that directs them to sign up page
- social media authentication can be used here as well
| 1.0 | Implement sign-up page for nonauthenticated users - - for users that are not logged in, there can be a sign up/login element that directs them to sign up page
- social media authentication can be used here as well
| priority | implement sign up page for nonauthenticated users for users that are not logged in there can be a sign up login element that directs them to sign up page social media authentication can be used here as well | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.