Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
253,298 | 8,053,830,504 | IssuesEvent | 2018-08-02 01:26:36 | kjohnsen/MMAPPR2 | https://api.github.com/repos/kjohnsen/MMAPPR2 | closed | Does callSampleSpecificVariants eliminate variants that have low frequency in WT? | complexity-medium priority-medium | This would be a problem because one missorted mutant would get the causative mutation thrown out. | 1.0 | Does callSampleSpecificVariants eliminate variants that have low frequency in WT? - This would be a problem because one missorted mutant would get the causative mutation thrown out. | priority | does callsamplespecificvariants eliminate variants that have low frequency in wt this would be a problem because one missorted mutant would get the causative mutation thrown out | 1 |
405,920 | 11,884,405,607 | IssuesEvent | 2020-03-27 17:35:23 | visit-dav/visit | https://api.github.com/repos/visit-dav/visit | closed | Qt 5 enabled, plot List missing up/down arrows and x | bug impact medium likelihood medium priority reviewed | Our plot list, when expanded, has up/down arrows next to operators for moving them relative to other operators, and has an x button for deleting the operator.These are missing when VisIt is built with Qt 5.I've observed this on Windows, don't have access to linux at the moment to see if it occurs there, too.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2795
Status: Pending
Project: VisIt
Tracker: Bug
Priority: High
Subject: Qt 5 enabled, plot List missing up/down arrows and x
Assigned to: -
Category: -
Target version: -
Author: Kathleen Biagas
Start: 03/29/2017
Due date:
% Done: 0%
Estimated time:
Created: 03/29/2017 04:28 pm
Updated: 04/04/2017 07:04 pm
Likelihood: 3 - Occasional
Severity: 3 - Major Irritation
Found in version: trunk
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Our plot list, when expanded, has up/down arrows next to operators for moving them relative to other operators, and has an x button for deleting the operator.These are missing when VisIt is built with Qt 5.I've observed this on Windows, don't have access to linux at the moment to see if it occurs there, too.
Comments:
| 1.0 | Qt 5 enabled, plot List missing up/down arrows and x - Our plot list, when expanded, has up/down arrows next to operators for moving them relative to other operators, and has an x button for deleting the operator.These are missing when VisIt is built with Qt 5.I've observed this on Windows, don't have access to linux at the moment to see if it occurs there, too.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2795
Status: Pending
Project: VisIt
Tracker: Bug
Priority: High
Subject: Qt 5 enabled, plot List missing up/down arrows and x
Assigned to: -
Category: -
Target version: -
Author: Kathleen Biagas
Start: 03/29/2017
Due date:
% Done: 0%
Estimated time:
Created: 03/29/2017 04:28 pm
Updated: 04/04/2017 07:04 pm
Likelihood: 3 - Occasional
Severity: 3 - Major Irritation
Found in version: trunk
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Our plot list, when expanded, has up/down arrows next to operators for moving them relative to other operators, and has an x button for deleting the operator.These are missing when VisIt is built with Qt 5.I've observed this on Windows, don't have access to linux at the moment to see if it occurs there, too.
Comments:
| priority | qt enabled plot list missing up down arrows and x our plot list when expanded has up down arrows next to operators for moving them relative to other operators and has an x button for deleting the operator these are missing when visit is built with qt i ve observed this on windows don t have access to linux at the moment to see if it occurs there too redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status pending project visit tracker bug priority high subject qt enabled plot list missing up down arrows and x assigned to category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity major irritation found in version trunk impact expected use os all support group any description our plot list when expanded has up down arrows next to operators for moving them relative to other operators and has an x button for deleting the operator these are missing when visit is built with qt i ve observed this on windows don t have access to linux at the moment to see if it occurs there too comments | 1 |
822,959 | 30,922,230,884 | IssuesEvent | 2023-08-06 03:16:27 | hphothong/smite | https://api.github.com/repos/hphothong/smite | closed | Team Balance toggle button can be turned off | bug good first issue medium priority | The toggle button for Team Balance should not be able to be turned off. When you select the same button as the current selection, nothing should happen.
Repro Steps:
1. Load the page
2. Select the BALANCED toggle filter
3. Select the BALANCED toggle filter again
4. Notice that the toggle button now shows no buttons as selected
AC:
- [ ] Given the BALANCED button is selected when selecting the BALANCED button again, then the BALANCED button should still be selected
- [ ] Given the BALANCED button is selected when selecting the UNBALANCED button, then the toggle button should switch to UNBALANCED
- [ ] Given the UNBALANCED button is selected when selecting the BALANCED button, then the toggle button should switch to BALANCED
- [ ] Given the UNBALANCED button is selected when selecting the UNBALANCED button, then the UNBALANCED button should still be selected | 1.0 | Team Balance toggle button can be turned off - The toggle button for Team Balance should not be able to be turned off. When you select the same button as the current selection, nothing should happen.
Repro Steps:
1. Load the page
2. Select the BALANCED toggle filter
3. Select the BALANCED toggle filter again
4. Notice that the toggle button now shows no buttons as selected
AC:
- [ ] Given the BALANCED button is selected when selecting the BALANCED button again, then the BALANCED button should still be selected
- [ ] Given the BALANCED button is selected when selecting the UNBALANCED button, then the toggle button should switch to UNBALANCED
- [ ] Given the UNBALANCED button is selected when selecting the BALANCED button, then the toggle button should switch to BALANCED
- [ ] Given the UNBALANCED button is selected when selecting the UNBALANCED button, then the UNBALANCED button should still be selected | priority | team balance toggle button can be turned off the toggle button for team balance should not be able to be turned off when you select the same button as the current selection nothing should happen repro steps load the page select the balanced toggle filter select the balanced toggle filter again notice that the toggle button now shows no buttons as selected ac given the balanced button is selected when selecting the balanced button again then the balanced button should still be selected given the balanced button is selected when selecting the unbalanced button then the toggle button should switch to unbalanced given the unbalanced button is selected when selecting the balanced button then the toggle button should switch to balanced given the unbalanced button is selected when selecting the unbalanced button then the unbalanced button should still be selected | 1 |
711,333 | 24,458,950,347 | IssuesEvent | 2022-10-07 09:22:00 | unikraft/unikraft | https://api.github.com/repos/unikraft/unikraft | opened | GDB Stub | kind/enhancement kind/project area/arch arch/x86_64 arch/arm arch/arm64 priority/medium | ### Feature request summary
Enable GDB introspection of running Unikraft instance.
Provides the interface to list threads, memory allocations, open files in Unikraft etc.
Requires a second serial connection to Unikraft
**Summary of objectives**
- [ ] Add a GDB stub server in Unikraft, that a GDB client can connect to to debug the running unikernel instance.
- [ ] Add support for hardware breakpoints.
**Depends on**: TODO
### Describe alternatives
_No response_
### Related architectures
_No response_
### Related platforms
_No response_
### Additional context
_No response_ | 1.0 | GDB Stub - ### Feature request summary
Enable GDB introspection of running Unikraft instance.
Provides the interface to list threads, memory allocations, open files in Unikraft etc.
Requires a second serial connection to Unikraft
**Summary of objectives**
- [ ] Add a GDB stub server in Unikraft, that a GDB client can connect to to debug the running unikernel instance.
- [ ] Add support for hardware breakpoints.
**Depends on**: TODO
### Describe alternatives
_No response_
### Related architectures
_No response_
### Related platforms
_No response_
### Additional context
_No response_ | priority | gdb stub feature request summary enable gdb introspection of running unikraft instance provides the interface to list threads memory allocations open files in unikraft etc requires a second serial connection to unikraft summary of objectives add a gdb stub server in unikraft that a gdb client can connect to to debug the running unikernel instance add support for hardware breakpoints depends on todo describe alternatives no response related architectures no response related platforms no response additional context no response | 1 |
39,526 | 2,856,236,478 | IssuesEvent | 2015-06-02 14:08:22 | aseprite/aseprite | https://api.github.com/repos/aseprite/aseprite | closed | Cannot create new frame if on a disabled layer | bug imported medium priority | _From [st...@sprixelsoft.com](https://code.google.com/u/105605591483612568850/) on June 22, 2013 22:45:18_
What steps will reproduce the problem? 1.disable current layer
2.attempt to create new frame
3.fail
_Original issue: http://code.google.com/p/aseprite/issues/detail?id=243_ | 1.0 | Cannot create new frame if on a disabled layer - _From [st...@sprixelsoft.com](https://code.google.com/u/105605591483612568850/) on June 22, 2013 22:45:18_
What steps will reproduce the problem? 1.disable current layer
2.attempt to create new frame
3.fail
_Original issue: http://code.google.com/p/aseprite/issues/detail?id=243_ | priority | cannot create new frame if on a disabled layer from on june what steps will reproduce the problem disable current layer attempt to create new frame fail original issue | 1 |
280,547 | 8,683,541,435 | IssuesEvent | 2018-12-02 19:03:46 | GammaStation/Gamma-Station | https://api.github.com/repos/GammaStation/Gamma-Station | closed | Двери клятые двери | Priority: Medium bug | <!--
ВАЖНО: Если ваш ишью является не репортом о баге, а предложением для чего-либо, то ОБЯЗАТЕЛЬНО добавьте в название тег [Proposal]
1. ОТВЕТЫ ОСТАВЛЯТЬ ПОД СООТВЕТСТВУЮЩИЕ ЗАГОЛОВКИ
(они в самом низу, после всех правил)
2. В ОДНОМ РЕПОРТЕ ДОЛЖНО БЫТЬ ОПИСАНИЕ ТОЛЬКО ОДНОЙ ПРОБЛЕМЫ
3. КОРРЕКТНОЕ НАЗВАНИЕ РЕПОРТА НЕ МЕНЕЕ ВАЖНО ЧЕМ ОПИСАНИЕ
-. Ниже описание каждого пункта.
1. Весь данный текст что уже написан до вас -
НЕ УДАЛЯТЬ И НЕ РЕДАКТИРОВАТЬ.
Если нечего написать в тот или иной пункт -
просто оставить пустым.
2. Не надо описывать пачку багов в одном репорте,
(!даже если там все описать можно парой слов!)
шанс что их исправят за раз, крайне мал.
А вот использовать на гите удобную функцию -
автозакрытия репорта при мерже пулл реквеста -
исправляющего данный репорт, будет невозможно.
3. Корректное и в меру подробное название репорта -
тоже очень важно! Чтобы даже не заходя в сам репорт -
было понятно что за проблема.
Плохой пример: "Ковер." - что мы должны понять из такого названия?
Хороший пример: "Некорректное отображение спрайтов ковра." -
а вот так уже будет понятно о чем репорт.
Это надо как минимум для того, чтобы вам же самим -
было видно, что репорта_нейм еще нет или наоборот,
уже есть, и это можно было понять не углубляясь в -
чтение каждого репорта внутри. Когда название не имеет конкретики, из -
которого нельзя понять о чем репорт, это также затрудняет функцию поиска.
-->
#### Подробное описание проблемы
Сломались доступы у бармена и библиотекаря, конкретно стеклянные
#### Что должно было произойти
Они должны открываться
#### Что произошло на самом деле
Они не открылись
#### Как повторить
Попробовать открыть за библиотекаря стеклянную дверь у него в офисе
#### Дополнительная информация:
Возможно где-то ещё я не проверял | 1.0 | Двери клятые двери - <!--
ВАЖНО: Если ваш ишью является не репортом о баге, а предложением для чего-либо, то ОБЯЗАТЕЛЬНО добавьте в название тег [Proposal]
1. ОТВЕТЫ ОСТАВЛЯТЬ ПОД СООТВЕТСТВУЮЩИЕ ЗАГОЛОВКИ
(они в самом низу, после всех правил)
2. В ОДНОМ РЕПОРТЕ ДОЛЖНО БЫТЬ ОПИСАНИЕ ТОЛЬКО ОДНОЙ ПРОБЛЕМЫ
3. КОРРЕКТНОЕ НАЗВАНИЕ РЕПОРТА НЕ МЕНЕЕ ВАЖНО ЧЕМ ОПИСАНИЕ
-. Ниже описание каждого пункта.
1. Весь данный текст что уже написан до вас -
НЕ УДАЛЯТЬ И НЕ РЕДАКТИРОВАТЬ.
Если нечего написать в тот или иной пункт -
просто оставить пустым.
2. Не надо описывать пачку багов в одном репорте,
(!даже если там все описать можно парой слов!)
шанс что их исправят за раз, крайне мал.
А вот использовать на гите удобную функцию -
автозакрытия репорта при мерже пулл реквеста -
исправляющего данный репорт, будет невозможно.
3. Корректное и в меру подробное название репорта -
тоже очень важно! Чтобы даже не заходя в сам репорт -
было понятно что за проблема.
Плохой пример: "Ковер." - что мы должны понять из такого названия?
Хороший пример: "Некорректное отображение спрайтов ковра." -
а вот так уже будет понятно о чем репорт.
Это надо как минимум для того, чтобы вам же самим -
было видно, что репорта_нейм еще нет или наоборот,
уже есть, и это можно было понять не углубляясь в -
чтение каждого репорта внутри. Когда название не имеет конкретики, из -
которого нельзя понять о чем репорт, это также затрудняет функцию поиска.
-->
#### Подробное описание проблемы
Сломались доступы у бармена и библиотекаря, конкретно стеклянные
#### Что должно было произойти
Они должны открываться
#### Что произошло на самом деле
Они не открылись
#### Как повторить
Попробовать открыть за библиотекаря стеклянную дверь у него в офисе
#### Дополнительная информация:
Возможно где-то ещё я не проверял | priority | двери клятые двери важно если ваш ишью является не репортом о баге а предложением для чего либо то обязательно добавьте в название тег ответы оставлять под соответствующие заголовки они в самом низу после всех правил в одном репорте должно быть описание только одной проблемы корректное название репорта не менее важно чем описание ниже описание каждого пункта весь данный текст что уже написан до вас не удалять и не редактировать если нечего написать в тот или иной пункт просто оставить пустым не надо описывать пачку багов в одном репорте даже если там все описать можно парой слов шанс что их исправят за раз крайне мал а вот использовать на гите удобную функцию автозакрытия репорта при мерже пулл реквеста исправляющего данный репорт будет невозможно корректное и в меру подробное название репорта тоже очень важно чтобы даже не заходя в сам репорт было понятно что за проблема плохой пример ковер что мы должны понять из такого названия хороший пример некорректное отображение спрайтов ковра а вот так уже будет понятно о чем репорт это надо как минимум для того чтобы вам же самим было видно что репорта нейм еще нет или наоборот уже есть и это можно было понять не углубляясь в чтение каждого репорта внутри когда название не имеет конкретики из которого нельзя понять о чем репорт это также затрудняет функцию поиска подробное описание проблемы сломались доступы у бармена и библиотекаря конкретно стеклянные что должно было произойти они должны открываться что произошло на самом деле они не открылись как повторить попробовать открыть за библиотекаря стеклянную дверь у него в офисе дополнительная информация возможно где то ещё я не проверял | 1 |
701,786 | 24,108,176,721 | IssuesEvent | 2022-09-20 09:10:41 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Bluetooth: Controller: Syncing with devices with per. adv. int. < ~10ms eventually causes events from BT controller stop arriving | bug priority: medium area: Bluetooth platform: nRF area: Bluetooth Controller | **Describe the bug**
When using periodic adv. intervals <~10ms then sometimes BT Controller stops reporting any events, including sync termination, CTEs, scan events etc.
- What target platform are you using? nRF52833, Zephyr
- Is this a regression? Not checked.
**To Reproduce**
Have 10 tags sending per. adv. with 7.5ms interval with adv. train of 5 and CTE enabled.
I have not found replicatable way to reproduce, sometimes it's enough to just let all 10 tags sync and then wait a minute and it happens. Sometimes it's easier to reproduce by randomly resetting some tags. I have reproduced without CTE sampling enabled, however may or may not be easier to reproduce with it on. I have also reproduced with only 5-6 tags but I think it's easier with more.
Application:
[aoa_receiver_multiple_49915.zip](https://github.com/zephyrproject-rtos/zephyr/files/9488286/aoa_receiver_multiple_49915.zip)
**Expected behavior**
BT controller keeps reporting events.
**Impact**
BT Controller stops generating events, application will be in strange state.
**Logs and console output**
The application prints <wrn> main: Interval: 6 whenever a device sending per. adv. is seen, as can be seen it's stopped reporting at the end.
```
*** Booting Zephyr OS build zephyr-v3.1.0-4244-g1995d349db37 ***
Starting Connectionless Locator Demo
Bluetooth initialization...success
Scan callbacks register...success.
Periodic Advertising callbacks register...success.
Start scanning...success
[00:00:00.004,638] <inf> main: Creating Sync for tags in list
Waiting for periodic advertising...
Scan is running...
[00:00:00.328,277] <wrn> main: Interval: 6
[00:00:00.328,521] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.470,764] <wrn> main: Interval: 6
[00:00:00.470,977] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.498,992] <wrn> main: Interval: 6
[00:00:00.499,206] <dbg> main: scan_recv: Added tag from per adv list
Scan is running...
[00:00:00.709,167] <wrn> main: Interval: 6
(709) PER_ADV_SYNC[0]: [DEVICE]: 36:DC:42:90:4D:0E (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:00.710,174] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:00.710,174] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(710) success. CTE receive enabled.
[00:00:00.741,027] <wrn> main: Interval: 6
[00:00:00.741,241] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.762,207] <wrn> main: Interval: 6
[00:00:00.762,420] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.883,422] <wrn> main: Interval: 6
[00:00:00.883,666] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.972,381] <wrn> main: Interval: 6
(988) PER_ADV_SYNC[1]: [DEVICE]: 3B:66:BD:7F:64:67 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:00.988,983] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:00.989,013] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(989) success. CTE receive enabled.
Running...
TAGS:sync...,sync...,33,sync...,2,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 4 Synced: 2 Num err: 0/35
[00:00:01.096,862] <wrn> main: Interval: 6
[00:00:01.097,106] <dbg> main: scan_recv: Added tag from per adv list
Scan is running...
[00:00:01.310,974] <wrn> main: Interval: 6
(1312) PER_ADV_SYNC[2]: [DEVICE]: 08:F7:34:24:2B:28 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:01.312,622] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:01.312,622] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(1313) success. CTE receive enabled.
[00:00:01.730,895] <wrn> main: Interval: 6
Scan is running...
[00:00:01.944,915] <wrn> main: Interval: 6
[00:00:01.945,129] <dbg> main: scan_recv: Added tag from per adv list
Running...
TAGS:sync...,55,56,sync...,66,sync...,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 5 Synced: 3 Num err: 0/178
[00:00:02.179,016] <wrn> main: Interval: 6
[00:00:02.442,382] <wrn> main: Interval: 6
Scan is running...
[00:00:02.598,999] <wrn> main: Interval: 6
[00:00:02.862,396] <wrn> main: Interval: 6
Running...
TAGS:sync...,79,49,sync...,59,sync...,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 5 Synced: 3 Num err: 0/191
[00:00:03.018,981] <wrn> main: Interval: 6
[00:00:03.035,156] <wrn> main: Interval: 6
[00:00:03.035,369] <dbg> main: scan_recv: Added tag from per adv list
[00:00:03.197,052] <wrn> main: Interval: 6
(3197) PER_ADV_SYNC[3]: [DEVICE]: 28:B5:CF:0A:86:88 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:03.198,150] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:03.198,181] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(3198) success. CTE receive enabled.
[00:00:03.268,554] <wrn> main: Interval: 6
(3276) PER_ADV_SYNC[4]: [DEVICE]: 1D:29:0E:6A:E8:65 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:03.277,221] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:03.277,221] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(3277) success. CTE receive enabled.
Scan is running...
[00:00:03.613,586] <wrn> main: Interval: 6
(3621) PER_ADV_SYNC[5]: [DEVICE]: 17:50:92:34:97:F7 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:03.622,253] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:03.622,253] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(3622) success. CTE receive enabled.
[00:00:03.912,384] <wrn> main: Interval: 6
Running...
TAGS:19,82,44,sync...,53,13,144,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 3 Synced: 6 Num err: 0/358
[00:00:04.050,933] <wrn> main: Interval: 40
[00:00:04.069,000] <wrn> main: Interval: 6
[00:00:04.453,491] <wrn> main: Interval: 6
[00:00:04.671,020] <wrn> main: Interval: 6
[00:00:04.948,516] <wrn> main: Interval: 6
Running...
TAGS:24,64,21,sync...,35,30,204,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 3 Synced: 6 Num err: 0/384
Scan is running...
[00:00:05.368,530] <wrn> main: Interval: 6
Running...
TAGS:24,56,30,sync...,32,35,196,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 3 Synced: 6 Num err: 0/381
[00:00:06.150,970] <wrn> main: Interval: 40
[00:00:06.185,180] <wrn> main: Interval: 6
(6207) PER_ADV_SYNC[6]: [DEVICE]: 2B:D6:79:55:8E:32 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:06.207,916] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:06.207,916] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(6208) success. CTE receive enabled.
[00:00:06.418,548] <wrn> main: Interval: 6
[00:00:06.770,996] <wrn> main: Interval: 6
Scan is running...
Running...
TAGS:27,37,23,sync...,27,35,191,sync...,16,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 2 Synced: 7 Num err: 0/367
[00:00:07.191,009] <wrn> main: Interval: 6
[00:00:07.200,988] <wrn> main: Interval: 40
[00:00:07.654,937] <wrn> main: Interval: 6
[00:00:07.888,549] <wrn> main: Interval: 6
Running...
TAGS:27,43,36,sync...,37,44,170,sync...,20,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 2 Synced: 7 Num err: 0/380
[00:00:08.034,912] <wrn> main: Interval: 6
(8088) PER_ADV_SYNC[7]: [DEVICE]: 17:BB:4D:10:4D:39 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:08.089,447] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:08.089,447] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(8090) success. CTE receive enabled.
Scan is running...
[00:00:08.867,034] <wrn> main: Interval: 6
Running...
TAGS:40,55,21,sync...,25,32,17,14,18,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 1 Synced: 8 Num err: 0/223
[00:00:09.287,017] <wrn> main: Interval: 6
[00:00:09.560,943] <wrn> main: Interval: 6
(9570) PER_ADV_SYNC[8]: [DEVICE]: 3A:9E:4A:1C:6F:D1 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:09.570,434] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:09.570,465] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(9570) success. CTE receive enabled.
Scan is running...
[00:00:09.710,906] <wrn> main: Interval: 6
[00:00:09.783,935] <err> bt_scan: Prepare CTE conn IQ report failed -22
Running...
TAGS:33,53,30,15,31,38,18,21,21,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/263
[00:00:10.134,857] <wrn> main: Interval: 6
Running...
TAGS:37,50,27,44,31,34,21,23,21,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/289
[00:00:11.030,792] <wrn> main: Interval: 6
[00:00:11.038,604] <wrn> main: Interval: 6
Scan is running...
[00:00:12.008,056] <err> bt_scan: Prepare CTE conn IQ report failed -22
Running...
TAGS:39,49,22,43,34,40,11,18,17,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/281
[00:00:12.049,133] <wrn> main: Interval: 6
[00:00:12.436,981] <wrn> main: Interval: 6
[00:00:12.445,037] <wrn> main: Interval: 6
Running...
TAGS:39,55,24,39,24,28,16,16,19,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/262
[00:00:13.740,722] <err> bt_scan: Prepare CTE conn IQ report failed -22
Running...
TAGS:30,44,27,39,32,34,19,19,20,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/277
[00:00:14.116,912] <wrn> main: Interval: 6
Scan is running...
Running...
TAGS:23,49,26,24,32,31,14,18,17,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/244
[00:00:15.251,892] <wrn> main: Interval: 6
[00:00:15.586,822] <wrn> main: Interval: 6
[00:00:16.003,265] <wrn> main: Interval: 6
Running...
TAGS:28,46,23,29,32,33,18,19,20,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/260
[00:00:16.054,779] <wrn> main: Interval: 6
[00:00:16.424,407] <wrn> main: Interval: 6
[00:00:16.424,652] <dbg> main: scan_recv: Added tag from per adv list
[00:00:16.435,058] <wrn> main: Interval: 6
Scan is running...
Running...
TAGS:23,52,20,39,28,34,12,13,14,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 1 Synced: 9 Num err: 0/235
[00:00:17.270,812] <wrn> main: Interval: 6
[00:00:17.900,756] <wrn> main: Interval: 6
Running...
TAGS:25,64,23,41,31,36,14,20,21,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 1 Synced: 9 Num err: 0/280
[00:00:18.734,436] <wrn> main: Interval: 6
(18773) PER_ADV_SYNC[9]: [DEVICE]: 2B:DA:D3:2E:AE:89 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:18.774,322] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:18.774,353] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(18775) success. CTE receive enabled.
[00:00:18.961,120] <wrn> main: Interval: 40
Running...
TAGS:23,45,27,31,26,31,14,14,17,2,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/237
Running...
----->After here no events arrive from controller, even powering of tags etc. won't generate sync term<-----
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
```
**Environment (please complete the following information):**
- OS: Windows
- Toolchain: Zephyr
- Commit SHA or Version used: 894423e9aed34a394b97c995c33549513b0ac4a5
| 1.0 | Bluetooth: Controller: Syncing with devices with per. adv. int. < ~10ms eventually causes events from BT controller stop arriving - **Describe the bug**
When using periodic adv. intervals <~10ms then sometimes BT Controller stops reporting any events, including sync termination, CTEs, scan events etc.
- What target platform are you using? nRF52833, Zephyr
- Is this a regression? Not checked.
**To Reproduce**
Have 10 tags sending per. adv. with 7.5ms interval with adv. train of 5 and CTE enabled.
I have not found replicatable way to reproduce, sometimes it's enough to just let all 10 tags sync and then wait a minute and it happens. Sometimes it's easier to reproduce by randomly resetting some tags. I have reproduced without CTE sampling enabled, however may or may not be easier to reproduce with it on. I have also reproduced with only 5-6 tags but I think it's easier with more.
Application:
[aoa_receiver_multiple_49915.zip](https://github.com/zephyrproject-rtos/zephyr/files/9488286/aoa_receiver_multiple_49915.zip)
**Expected behavior**
BT controller keeps reporting events.
**Impact**
BT Controller stops generating events, application will be in strange state.
**Logs and console output**
The application prints <wrn> main: Interval: 6 whenever a device sending per. adv. is seen, as can be seen it's stopped reporting at the end.
```
*** Booting Zephyr OS build zephyr-v3.1.0-4244-g1995d349db37 ***
Starting Connectionless Locator Demo
Bluetooth initialization...success
Scan callbacks register...success.
Periodic Advertising callbacks register...success.
Start scanning...success
[00:00:00.004,638] <inf> main: Creating Sync for tags in list
Waiting for periodic advertising...
Scan is running...
[00:00:00.328,277] <wrn> main: Interval: 6
[00:00:00.328,521] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.470,764] <wrn> main: Interval: 6
[00:00:00.470,977] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.498,992] <wrn> main: Interval: 6
[00:00:00.499,206] <dbg> main: scan_recv: Added tag from per adv list
Scan is running...
[00:00:00.709,167] <wrn> main: Interval: 6
(709) PER_ADV_SYNC[0]: [DEVICE]: 36:DC:42:90:4D:0E (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:00.710,174] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:00.710,174] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(710) success. CTE receive enabled.
[00:00:00.741,027] <wrn> main: Interval: 6
[00:00:00.741,241] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.762,207] <wrn> main: Interval: 6
[00:00:00.762,420] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.883,422] <wrn> main: Interval: 6
[00:00:00.883,666] <dbg> main: scan_recv: Added tag from per adv list
[00:00:00.972,381] <wrn> main: Interval: 6
(988) PER_ADV_SYNC[1]: [DEVICE]: 3B:66:BD:7F:64:67 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:00.988,983] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:00.989,013] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(989) success. CTE receive enabled.
Running...
TAGS:sync...,sync...,33,sync...,2,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 4 Synced: 2 Num err: 0/35
[00:00:01.096,862] <wrn> main: Interval: 6
[00:00:01.097,106] <dbg> main: scan_recv: Added tag from per adv list
Scan is running...
[00:00:01.310,974] <wrn> main: Interval: 6
(1312) PER_ADV_SYNC[2]: [DEVICE]: 08:F7:34:24:2B:28 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:01.312,622] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:01.312,622] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(1313) success. CTE receive enabled.
[00:00:01.730,895] <wrn> main: Interval: 6
Scan is running...
[00:00:01.944,915] <wrn> main: Interval: 6
[00:00:01.945,129] <dbg> main: scan_recv: Added tag from per adv list
Running...
TAGS:sync...,55,56,sync...,66,sync...,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 5 Synced: 3 Num err: 0/178
[00:00:02.179,016] <wrn> main: Interval: 6
[00:00:02.442,382] <wrn> main: Interval: 6
Scan is running...
[00:00:02.598,999] <wrn> main: Interval: 6
[00:00:02.862,396] <wrn> main: Interval: 6
Running...
TAGS:sync...,79,49,sync...,59,sync...,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 5 Synced: 3 Num err: 0/191
[00:00:03.018,981] <wrn> main: Interval: 6
[00:00:03.035,156] <wrn> main: Interval: 6
[00:00:03.035,369] <dbg> main: scan_recv: Added tag from per adv list
[00:00:03.197,052] <wrn> main: Interval: 6
(3197) PER_ADV_SYNC[3]: [DEVICE]: 28:B5:CF:0A:86:88 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:03.198,150] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:03.198,181] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(3198) success. CTE receive enabled.
[00:00:03.268,554] <wrn> main: Interval: 6
(3276) PER_ADV_SYNC[4]: [DEVICE]: 1D:29:0E:6A:E8:65 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:03.277,221] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:03.277,221] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(3277) success. CTE receive enabled.
Scan is running...
[00:00:03.613,586] <wrn> main: Interval: 6
(3621) PER_ADV_SYNC[5]: [DEVICE]: 17:50:92:34:97:F7 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:03.622,253] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:03.622,253] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(3622) success. CTE receive enabled.
[00:00:03.912,384] <wrn> main: Interval: 6
Running...
TAGS:19,82,44,sync...,53,13,144,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 3 Synced: 6 Num err: 0/358
[00:00:04.050,933] <wrn> main: Interval: 40
[00:00:04.069,000] <wrn> main: Interval: 6
[00:00:04.453,491] <wrn> main: Interval: 6
[00:00:04.671,020] <wrn> main: Interval: 6
[00:00:04.948,516] <wrn> main: Interval: 6
Running...
TAGS:24,64,21,sync...,35,30,204,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 3 Synced: 6 Num err: 0/384
Scan is running...
[00:00:05.368,530] <wrn> main: Interval: 6
Running...
TAGS:24,56,30,sync...,32,35,196,sync...,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 3 Synced: 6 Num err: 0/381
[00:00:06.150,970] <wrn> main: Interval: 40
[00:00:06.185,180] <wrn> main: Interval: 6
(6207) PER_ADV_SYNC[6]: [DEVICE]: 2B:D6:79:55:8E:32 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:06.207,916] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:06.207,916] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(6208) success. CTE receive enabled.
[00:00:06.418,548] <wrn> main: Interval: 6
[00:00:06.770,996] <wrn> main: Interval: 6
Scan is running...
Running...
TAGS:27,37,23,sync...,27,35,191,sync...,16,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 2 Synced: 7 Num err: 0/367
[00:00:07.191,009] <wrn> main: Interval: 6
[00:00:07.200,988] <wrn> main: Interval: 40
[00:00:07.654,937] <wrn> main: Interval: 6
[00:00:07.888,549] <wrn> main: Interval: 6
Running...
TAGS:27,43,36,sync...,37,44,170,sync...,20,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 2 Synced: 7 Num err: 0/380
[00:00:08.034,912] <wrn> main: Interval: 6
(8088) PER_ADV_SYNC[7]: [DEVICE]: 17:BB:4D:10:4D:39 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:08.089,447] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:08.089,447] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(8090) success. CTE receive enabled.
Scan is running...
[00:00:08.867,034] <wrn> main: Interval: 6
Running...
TAGS:40,55,21,sync...,25,32,17,14,18,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 1 Synced: 8 Num err: 0/223
[00:00:09.287,017] <wrn> main: Interval: 6
[00:00:09.560,943] <wrn> main: Interval: 6
(9570) PER_ADV_SYNC[8]: [DEVICE]: 3A:9E:4A:1C:6F:D1 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:09.570,434] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:09.570,465] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(9570) success. CTE receive enabled.
Scan is running...
[00:00:09.710,906] <wrn> main: Interval: 6
[00:00:09.783,935] <err> bt_scan: Prepare CTE conn IQ report failed -22
Running...
TAGS:33,53,30,15,31,38,18,21,21,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/263
[00:00:10.134,857] <wrn> main: Interval: 6
Running...
TAGS:37,50,27,44,31,34,21,23,21,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/289
[00:00:11.030,792] <wrn> main: Interval: 6
[00:00:11.038,604] <wrn> main: Interval: 6
Scan is running...
[00:00:12.008,056] <err> bt_scan: Prepare CTE conn IQ report failed -22
Running...
TAGS:39,49,22,43,34,40,11,18,17,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/281
[00:00:12.049,133] <wrn> main: Interval: 6
[00:00:12.436,981] <wrn> main: Interval: 6
[00:00:12.445,037] <wrn> main: Interval: 6
Running...
TAGS:39,55,24,39,24,28,16,16,19,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/262
[00:00:13.740,722] <err> bt_scan: Prepare CTE conn IQ report failed -22
Running...
TAGS:30,44,27,39,32,34,19,19,20,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/277
[00:00:14.116,912] <wrn> main: Interval: 6
Scan is running...
Running...
TAGS:23,49,26,24,32,31,14,18,17,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/244
[00:00:15.251,892] <wrn> main: Interval: 6
[00:00:15.586,822] <wrn> main: Interval: 6
[00:00:16.003,265] <wrn> main: Interval: 6
Running...
TAGS:28,46,23,29,32,33,18,19,20,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 9 Num err: 0/260
[00:00:16.054,779] <wrn> main: Interval: 6
[00:00:16.424,407] <wrn> main: Interval: 6
[00:00:16.424,652] <dbg> main: scan_recv: Added tag from per adv list
[00:00:16.435,058] <wrn> main: Interval: 6
Scan is running...
Running...
TAGS:23,52,20,39,28,34,12,13,14,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 1 Synced: 9 Num err: 0/235
[00:00:17.270,812] <wrn> main: Interval: 6
[00:00:17.900,756] <wrn> main: Interval: 6
Running...
TAGS:25,64,23,41,31,36,14,20,21,sync...,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 1 Synced: 9 Num err: 0/280
[00:00:18.734,436] <wrn> main: Interval: 6
(18773) PER_ADV_SYNC[9]: [DEVICE]: 2B:DA:D3:2E:AE:89 (random) synced, Interval 0x0006 (7 ms), PHY LE 2M
[00:00:18.774,322] <dbg> main: sync_cb: Removed tag from per adv list
[00:00:18.774,353] <inf> main: Creating Sync for tags in list
Enable receiving of CTE...
(18775) success. CTE receive enabled.
[00:00:18.961,120] <wrn> main: Interval: 40
Running...
TAGS:23,45,27,31,26,31,14,14,17,2,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/237
Running...
----->After here no events arrive from controller, even powering of tags etc. won't generate sync term<-----
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
Running...
TAGS:0,0,0,0,0,0,0,0,0,0,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,-,
Queued: 0 Synced: 10 Num err: 0/0
```
**Environment (please complete the following information):**
- OS: Windows
- Toolchain: Zephyr
- Commit SHA or Version used: 894423e9aed34a394b97c995c33549513b0ac4a5
| priority | bluetooth controller syncing with devices with per adv int eventually causes events from bt controller stop arriving describe the bug when using periodic adv intervals then sometimes bt controller stops reporting any events including sync termination ctes scan events etc what target platform are you using zephyr is this a regression not checked to reproduce have tags sending per adv with interval with adv train of and cte enabled i have not found replicatable way to reproduce sometimes it s enough to just let all tags sync and then wait a minute and it happens sometimes it s easier to reproduce by randomly resetting some tags i have reproduced without cte sampling enabled however may or may not be easier to reproduce with it on i have also reproduced with only tags but i think it s easier with more application expected behavior bt controller keeps reporting events impact bt controller stops generating events application will be in strange state logs and console output the application prints main interval whenever a device sending per adv is seen as can be seen it s stopped reporting at the end booting zephyr os build zephyr starting connectionless locator demo bluetooth initialization success scan callbacks register success periodic advertising callbacks register success start scanning success main creating sync for tags in list waiting for periodic advertising scan is running main interval main scan recv added tag from per adv list main interval main scan recv added tag from per adv list main interval main scan recv added tag from per adv list scan is running main interval per adv sync dc random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled main interval main scan recv added tag from per adv list main interval main scan recv added tag from per adv list main interval main scan recv added tag from per adv list main interval per adv sync bd random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled running tags sync sync sync sync queued synced num err main interval main scan recv added tag from per adv list scan is running main interval per adv sync random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled main interval scan is running main interval main scan recv added tag from per adv list running tags sync sync sync sync sync queued synced num err main interval main interval scan is running main interval main interval running tags sync sync sync sync sync queued synced num err main interval main interval main scan recv added tag from per adv list main interval per adv sync cf random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled main interval per adv sync random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled scan is running main interval per adv sync random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled main interval running tags sync sync sync queued synced num err main interval main interval main interval main interval main interval running tags sync sync sync queued synced num err scan is running main interval running tags sync sync sync queued synced num err main interval main interval per adv sync random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled main interval main interval scan is running running tags sync sync queued synced num err main interval main interval main interval main interval running tags sync sync queued synced num err main interval per adv sync bb random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled scan is running main interval running tags sync queued synced num err main interval main interval per adv sync random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled scan is running main interval bt scan prepare cte conn iq report failed running tags queued synced num err main interval running tags queued synced num err main interval main interval scan is running bt scan prepare cte conn iq report failed running tags queued synced num err main interval main interval main interval running tags queued synced num err bt scan prepare cte conn iq report failed running tags queued synced num err main interval scan is running running tags queued synced num err main interval main interval main interval running tags queued synced num err main interval main interval main scan recv added tag from per adv list main interval scan is running running tags sync queued synced num err main interval main interval running tags sync queued synced num err main interval per adv sync da ae random synced interval ms phy le main sync cb removed tag from per adv list main creating sync for tags in list enable receiving of cte success cte receive enabled main interval running tags queued synced num err running after here no events arrive from controller even powering of tags etc won t generate sync term tags queued synced num err running tags queued synced num err running tags queued synced num err running tags queued synced num err running tags queued synced num err running tags queued synced num err environment please complete the following information os windows toolchain zephyr commit sha or version used | 1 |
25,658 | 2,683,911,299 | IssuesEvent | 2015-03-28 13:15:27 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | [ConEmu 2010.3.8] Biew - нет кейбара | 1 star bug imported Priority-Medium wontfix | _From [andrey.b...@gmail.com](https://code.google.com/u/117303310870772091797/) on March 17, 2010 14:45:25_
В biew (хекс едитор) не видно кейбара. Если отобразить реальную консоль, то
видно, что она километровой высоты. https://sourceforge.net/projects/beye/files/biew/6.1.0/biew-610-win32.zip/download
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=209_ | 1.0 | [ConEmu 2010.3.8] Biew - нет кейбара - _From [andrey.b...@gmail.com](https://code.google.com/u/117303310870772091797/) on March 17, 2010 14:45:25_
В biew (хекс едитор) не видно кейбара. Если отобразить реальную консоль, то
видно, что она километровой высоты. https://sourceforge.net/projects/beye/files/biew/6.1.0/biew-610-win32.zip/download
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=209_ | priority | biew нет кейбара from on march в biew хекс едитор не видно кейбара если отобразить реальную консоль то видно что она километровой высоты original issue | 1 |
774,781 | 27,209,902,997 | IssuesEvent | 2023-02-20 15:45:18 | Fiserv/Support | https://api.github.com/repos/Fiserv/Support | closed | Unable to reset Developer Studio Password | bug Priority - Medium Severity - Low BankingHub Login/Signup | # Reporting new issue for Fiser Developer Studio
I'm unable to reset the Developer Studio account password. Whenever I enter the One-time passcode (OTP) received via email to reset the password, it throws incorrect OTP error all the time.

| 1.0 | Unable to reset Developer Studio Password - # Reporting new issue for Fiser Developer Studio
I'm unable to reset the Developer Studio account password. Whenever I enter the One-time passcode (OTP) received via email to reset the password, it throws incorrect OTP error all the time.

| priority | unable to reset developer studio password reporting new issue for fiser developer studio i m unable to reset the developer studio account password whenever i enter the one time passcode otp received via email to reset the password it throws incorrect otp error all the time | 1 |
671,257 | 22,751,261,574 | IssuesEvent | 2022-07-07 13:18:57 | status-im/status-desktop | https://api.github.com/repos/status-im/status-desktop | closed | The notification banner 'Secure your seed phrase Back up now' is still shown after back up the seed phrase | bug priority 2: medium E:Bugfixes Settings S:2 E:Settings | # Bug Report
## Steps to reproduce
1. Create a new account
2. Go to Settings and click 'Secure your seed phrase Back up now'
3. Go step by step and complete successfully the process of backup
4. Check the Setting page and click on the notification banner 'Secure your seed phrase Back up now'
#### Expected behavior
The notification banner 'Secure your seed phrase Back up now' is NOT shown after backing up the seed phrase because the user successfully backed up the seed phrase
#### Actual behavior
The notification banner 'Secure your seed phrase Back up now' is still shown after backing up the seed phrase
### Additional Information
https://user-images.githubusercontent.com/14942081/176036318-d3f0112a-99fe-46cc-a04e-1878c1530a0f.mov
Status desktop version: https://ci.status.im/job/status-desktop/job/branches/job/macos/job/master/2246/
Operating System: macOS Monterey 12.3 Beta (21E5227a)
| 1.0 | The notification banner 'Secure your seed phrase Back up now' is still shown after back up the seed phrase - # Bug Report
## Steps to reproduce
1. Create a new account
2. Go to Settings and click 'Secure your seed phrase Back up now'
3. Go step by step and complete successfully the process of backup
4. Check the Setting page and click on the notification banner 'Secure your seed phrase Back up now'
#### Expected behavior
The notification banner 'Secure your seed phrase Back up now' is NOT shown after backing up the seed phrase because the user successfully backed up the seed phrase
#### Actual behavior
The notification banner 'Secure your seed phrase Back up now' is still shown after backing up the seed phrase
### Additional Information
https://user-images.githubusercontent.com/14942081/176036318-d3f0112a-99fe-46cc-a04e-1878c1530a0f.mov
Status desktop version: https://ci.status.im/job/status-desktop/job/branches/job/macos/job/master/2246/
Operating System: macOS Monterey 12.3 Beta (21E5227a)
| priority | the notification banner secure your seed phrase back up now is still shown after back up the seed phrase bug report steps to reproduce create a new account go to settings and click secure your seed phrase back up now go step by step and complete successfully the process of backup check the setting page and click on the notification banner secure your seed phrase back up now expected behavior the notification banner secure your seed phrase back up now is not shown after backing up the seed phrase because the user successfully backed up the seed phrase actual behavior the notification banner secure your seed phrase back up now is still shown after backing up the seed phrase additional information status desktop version operating system macos monterey beta | 1 |
552,205 | 16,218,309,014 | IssuesEvent | 2021-05-06 00:07:01 | MycroftAI/hardware-mycroft-mark-II | https://api.github.com/repos/MycroftAI/hardware-mycroft-mark-II | opened | Mycroft seems slower to respond on the Mark II | Priority: Medium bug | **Describe the bug**
Have had it raised by community members that responses to queries sometimes seem very slow on the Mark II where they are not on other instances of Mycroft. It seems that this may be caused by the listener silence detection not working correctly.
By default Mycroft will listen for up to 10 seconds before sending this to an STT service for transcription. However as many utterances are much less than this, if there is a long enough period of silence, Mycroft will consider the utterance finished and send the audio earlier than that 10 seconds.
I'm seeing that sometimes the listener remains open for the whole 10 seconds, even if the user is not speaking. This makes it seem like the processing is taking a very long time but it's actually listening for too long.
I presume this is related to the silence detection and may differ based on the audio input being received.
If that's the case, consideration should be given to making whatever fixes it configurable so it can be defined in `mycroft.conf` or the Hardware Abstraction Layer and hence differ from device to device.
**To Reproduce**
Steps to reproduce the behavior:
1. Open the mycroft-cli-client
2. Speak to Mycroft
3. Watch when the mic is activated and stopped. | 1.0 | Mycroft seems slower to respond on the Mark II - **Describe the bug**
Have had it raised by community members that responses to queries sometimes seem very slow on the Mark II where they are not on other instances of Mycroft. It seems that this may be caused by the listener silence detection not working correctly.
By default Mycroft will listen for up to 10 seconds before sending this to an STT service for transcription. However as many utterances are much less than this, if there is a long enough period of silence, Mycroft will consider the utterance finished and send the audio earlier than that 10 seconds.
I'm seeing that sometimes the listener remains open for the whole 10 seconds, even if the user is not speaking. This makes it seem like the processing is taking a very long time but it's actually listening for too long.
I presume this is related to the silence detection and may differ based on the audio input being received.
If that's the case, consideration should be given to making whatever fixes it configurable so it can be defined in `mycroft.conf` or the Hardware Abstraction Layer and hence differ from device to device.
**To Reproduce**
Steps to reproduce the behavior:
1. Open the mycroft-cli-client
2. Speak to Mycroft
3. Watch when the mic is activated and stopped. | priority | mycroft seems slower to respond on the mark ii describe the bug have had it raised by community members that responses to queries sometimes seem very slow on the mark ii where they are not on other instances of mycroft it seems that this may be caused by the listener silence detection not working correctly by default mycroft will listen for up to seconds before sending this to an stt service for transcription however as many utterances are much less than this if there is a long enough period of silence mycroft will consider the utterance finished and send the audio earlier than that seconds i m seeing that sometimes the listener remains open for the whole seconds even if the user is not speaking this makes it seem like the processing is taking a very long time but it s actually listening for too long i presume this is related to the silence detection and may differ based on the audio input being received if that s the case consideration should be given to making whatever fixes it configurable so it can be defined in mycroft conf or the hardware abstraction layer and hence differ from device to device to reproduce steps to reproduce the behavior open the mycroft cli client speak to mycroft watch when the mic is activated and stopped | 1 |
325,501 | 9,931,876,005 | IssuesEvent | 2019-07-02 08:33:55 | Code-Poets/sheetstorm | https://api.github.com/repos/Code-Poets/sheetstorm | closed | Unify validator usage in models | chore priority medium refactor | Right now there is no structured way of validators' usage. Part of the validation is done by passing `validator`s to models' fields arguments, part of validation is done in `clean` method.
Should be done:
------------
- Single field related validation should be done via `validator` passed as argument. ie.
```python
email = models.EmailField(
CustomUserModelText.EMAIL_ADDRESS,
max_length=constants.EMAIL_MAX_LENGTH,
unique=True,
validators=[custom_validate_email_function]
)
```
- Validation related to more fields (or related to inter-dependency between fields) should be done in `clean()` method, ie.
```python
def clean(self) -> None:
super().clean()
if (
hasattr(self, "author")
and isinstance(self.work_hours, Decimal)
and self.author.report_set.get_report_work_hours_sum_for_date(self.date, self.pk) + self.work_hours > 24
):
raise ValidationError(
message=ReportValidationStrings.WORK_HOURS_SUM_FOR_GIVEN_DATE_FOR_SINGLE_AUTHOR_EXCEEDED.value
)
```
| 1.0 | Unify validator usage in models - Right now there is no structured way of validators' usage. Part of the validation is done by passing `validator`s to models' fields arguments, part of validation is done in `clean` method.
Should be done:
------------
- Single field related validation should be done via `validator` passed as argument. ie.
```python
email = models.EmailField(
CustomUserModelText.EMAIL_ADDRESS,
max_length=constants.EMAIL_MAX_LENGTH,
unique=True,
validators=[custom_validate_email_function]
)
```
- Validation related to more fields (or related to inter-dependency between fields) should be done in `clean()` method, ie.
```python
def clean(self) -> None:
super().clean()
if (
hasattr(self, "author")
and isinstance(self.work_hours, Decimal)
and self.author.report_set.get_report_work_hours_sum_for_date(self.date, self.pk) + self.work_hours > 24
):
raise ValidationError(
message=ReportValidationStrings.WORK_HOURS_SUM_FOR_GIVEN_DATE_FOR_SINGLE_AUTHOR_EXCEEDED.value
)
```
| priority | unify validator usage in models right now there is no structured way of validators usage part of the validation is done by passing validator s to models fields arguments part of validation is done in clean method should be done single field related validation should be done via validator passed as argument ie python email models emailfield customusermodeltext email address max length constants email max length unique true validators validation related to more fields or related to inter dependency between fields should be done in clean method ie python def clean self none super clean if hasattr self author and isinstance self work hours decimal and self author report set get report work hours sum for date self date self pk self work hours raise validationerror message reportvalidationstrings work hours sum for given date for single author exceeded value | 1 |
295,136 | 9,082,303,810 | IssuesEvent | 2019-02-17 11:03:37 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | closed | [LOCALIZATION] | nick_the_godslayer | :beetle: bug - localization :scroll: :grey_exclamation: priority medium | **Mod Version**
Master branch
**Please explain your issue in as much detail as possible:**
"The godslayer" nickname don't have localisation.
**Upload screenshots of the problem localization:**
<details>
<summary>Click to expand</summary>


</details> | 1.0 | [LOCALIZATION] | nick_the_godslayer - **Mod Version**
Master branch
**Please explain your issue in as much detail as possible:**
"The godslayer" nickname don't have localisation.
**Upload screenshots of the problem localization:**
<details>
<summary>Click to expand</summary>


</details> | priority | nick the godslayer mod version master branch please explain your issue in as much detail as possible the godslayer nickname don t have localisation upload screenshots of the problem localization click to expand | 1 |
268,100 | 8,403,275,112 | IssuesEvent | 2018-10-11 09:19:42 | ePADD/epadd | https://api.github.com/repos/ePADD/epadd | opened | Autocomplete for Attachments in Advanced Search not working | Bug Medium priority | v7 10Oct; Bush full
An attachment with filename "A letter" exists in Blobs folder. System gives "no results" when user type "A" in the File name field in Attachment section in Advanced search. | 1.0 | Autocomplete for Attachments in Advanced Search not working - v7 10Oct; Bush full
An attachment with filename "A letter" exists in Blobs folder. System gives "no results" when user type "A" in the File name field in Attachment section in Advanced search. | priority | autocomplete for attachments in advanced search not working bush full an attachment with filename a letter exists in blobs folder system gives no results when user type a in the file name field in attachment section in advanced search | 1 |
629,890 | 20,070,134,812 | IssuesEvent | 2022-02-04 05:05:26 | sonia-auv/octopus-telemetry | https://api.github.com/repos/sonia-auv/octopus-telemetry | closed | Depth Indicator (SONIA) | Priority: Medium Type: Feature | **Warning :** Before creating an issue or task, make sure that it does not already exists in the [issue tracker](../). Thank you.
## Context
Implémenter l'indicateur de profondeur
## Changes
<!-- Give a brief description of the components that need to change and how -->
## Comments
<!-- Add further comments if needed -->
| 1.0 | Depth Indicator (SONIA) - **Warning :** Before creating an issue or task, make sure that it does not already exists in the [issue tracker](../). Thank you.
## Context
Implémenter l'indicateur de profondeur
## Changes
<!-- Give a brief description of the components that need to change and how -->
## Comments
<!-- Add further comments if needed -->
| priority | depth indicator sonia warning before creating an issue or task make sure that it does not already exists in the thank you context implémenter l indicateur de profondeur changes comments | 1 |
26,949 | 2,689,104,720 | IssuesEvent | 2015-03-31 07:49:14 | adobe/brackets | https://api.github.com/repos/adobe/brackets | opened | Drag & drop text preference (dragDropText) doesn't take effect properly on files already open | F Editor medium priority | 1. Start with the preference off
2. Open a few files
3. Open the preferences file, turn `dragDropText` on, and save
4. Try to drag the selected text in any of the already open editors
Result:
- on Mac, dragging the selection does nothing
- on Win, dragging the selection appears to drag an image of the entire editor area
Expected:
Selection should be dragged normally, as seen if the files are closed & reopened or if Brackets is restarted. | 1.0 | Drag & drop text preference (dragDropText) doesn't take effect properly on files already open - 1. Start with the preference off
2. Open a few files
3. Open the preferences file, turn `dragDropText` on, and save
4. Try to drag the selected text in any of the already open editors
Result:
- on Mac, dragging the selection does nothing
- on Win, dragging the selection appears to drag an image of the entire editor area
Expected:
Selection should be dragged normally, as seen if the files are closed & reopened or if Brackets is restarted. | priority | drag drop text preference dragdroptext doesn t take effect properly on files already open start with the preference off open a few files open the preferences file turn dragdroptext on and save try to drag the selected text in any of the already open editors result on mac dragging the selection does nothing on win dragging the selection appears to drag an image of the entire editor area expected selection should be dragged normally as seen if the files are closed reopened or if brackets is restarted | 1 |
545,053 | 15,935,106,733 | IssuesEvent | 2021-04-14 09:27:42 | ansible-collections/amazon.aws | https://api.github.com/repos/ansible-collections/amazon.aws | closed | When using keyed_groups with tags, an empty value introduces a trailing underscore | bug has_pr inventory plugins priority/medium | <!--- Verify first that your issue is not already reported on GitHub -- I searched, didn't find anything -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
I am using the `aws_ec2` inventory module to dynamically get inventory. I have configured it so that it creates groups of machines based on the tags that are added to the instance. With the old `ec2.py` inventory module I was using, a tag that did not have a value would result in a group named `tag_foo` where `foo` is the key name. Since the tag does not have a value, the tag simply includes the key and no value.
The `aws_ec2` inventory module has similar support when using `keyed_groups`. However, when a tag does not include a value, a trailing slash is appended to the group. In the example above, the name of the group ends up being: `tag_foo_`.
I feel that the trailing underscore is a bug and should not be appended if the tag's value is empty.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
aws_ec2 inventory module
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible 2.10.6
config file = /Users/rca/projects/ansible-playbooks/ansible.cfg
configured module search path = ['/Users/rca/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /Users/rca/.local/share/virtualenvs/ansible-playbooks-NHzTLeaW/lib/python3.9/site-packages/ansible
executable location = /Users/rca/.local/share/virtualenvs/ansible-playbooks-NHzTLeaW/bin/ansible
python version = 3.9.2 (v3.9.2:1a79785e3e, Feb 19 2021, 09:06:10) [Clang 6.0 (clang-600.0.57)]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
---
plugin: aws_ec2
cache: yes
regions:
- us-east-1
filters:
instance-state-name: running
hostnames:
- name: private-ip-address
prefix: tag:Name
keyed_groups:
- prefix: tag
key: tags
compose:
# Use the private IP address to connect to the host
# (note: this does not modify inventory_hostname, which is set via I(hostnames))
ansible_host: private_ip_address
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
mac os 10.15
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
create an ec2 instance and add a tag `foo` with no value. create the inventory and notice that the keyed group created for that tag has a trailing underscore.
<!--- Paste example playbooks or commands between quotes below -->
- clear out inventory cache before running this
```yaml
ansible-inventory --list | less
```
search for `tag_foo` and notice that it has a trailing slash
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
Expect to have a tag without a trailing slash.
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
The tag has a trailing slash:
```
"tag_foo_": {
"hosts": [
"machine-name_10.0.0.123"
]
},
```
<!--- Paste verbatim command output between quotes -->
```paste below
ansible tag_foo_ -m ping
```
that will ping the machine with that tag and confirm the group can be used with the trailing slash, but IMO the trailing slash should not be there.
Thank you! | 1.0 | When using keyed_groups with tags, an empty value introduces a trailing underscore - <!--- Verify first that your issue is not already reported on GitHub -- I searched, didn't find anything -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
I am using the `aws_ec2` inventory module to dynamically get inventory. I have configured it so that it creates groups of machines based on the tags that are added to the instance. With the old `ec2.py` inventory module I was using, a tag that did not have a value would result in a group named `tag_foo` where `foo` is the key name. Since the tag does not have a value, the tag simply includes the key and no value.
The `aws_ec2` inventory module has similar support when using `keyed_groups`. However, when a tag does not include a value, a trailing slash is appended to the group. In the example above, the name of the group ends up being: `tag_foo_`.
I feel that the trailing underscore is a bug and should not be appended if the tag's value is empty.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
aws_ec2 inventory module
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible 2.10.6
config file = /Users/rca/projects/ansible-playbooks/ansible.cfg
configured module search path = ['/Users/rca/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /Users/rca/.local/share/virtualenvs/ansible-playbooks-NHzTLeaW/lib/python3.9/site-packages/ansible
executable location = /Users/rca/.local/share/virtualenvs/ansible-playbooks-NHzTLeaW/bin/ansible
python version = 3.9.2 (v3.9.2:1a79785e3e, Feb 19 2021, 09:06:10) [Clang 6.0 (clang-600.0.57)]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
---
plugin: aws_ec2
cache: yes
regions:
- us-east-1
filters:
instance-state-name: running
hostnames:
- name: private-ip-address
prefix: tag:Name
keyed_groups:
- prefix: tag
key: tags
compose:
# Use the private IP address to connect to the host
# (note: this does not modify inventory_hostname, which is set via I(hostnames))
ansible_host: private_ip_address
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
mac os 10.15
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
create an ec2 instance and add a tag `foo` with no value. create the inventory and notice that the keyed group created for that tag has a trailing underscore.
<!--- Paste example playbooks or commands between quotes below -->
- clear out inventory cache before running this
```yaml
ansible-inventory --list | less
```
search for `tag_foo` and notice that it has a trailing slash
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
Expect to have a tag without a trailing slash.
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
The tag has a trailing slash:
```
"tag_foo_": {
"hosts": [
"machine-name_10.0.0.123"
]
},
```
<!--- Paste verbatim command output between quotes -->
```paste below
ansible tag_foo_ -m ping
```
that will ping the machine with that tag and confirm the group can be used with the trailing slash, but IMO the trailing slash should not be there.
Thank you! | priority | when using keyed groups with tags an empty value introduces a trailing underscore summary i am using the aws inventory module to dynamically get inventory i have configured it so that it creates groups of machines based on the tags that are added to the instance with the old py inventory module i was using a tag that did not have a value would result in a group named tag foo where foo is the key name since the tag does not have a value the tag simply includes the key and no value the aws inventory module has similar support when using keyed groups however when a tag does not include a value a trailing slash is appended to the group in the example above the name of the group ends up being tag foo i feel that the trailing underscore is a bug and should not be appended if the tag s value is empty issue type bug report component name aws inventory module ansible version paste below ansible config file users rca projects ansible playbooks ansible cfg configured module search path ansible python module location users rca local share virtualenvs ansible playbooks nhztleaw lib site packages ansible executable location users rca local share virtualenvs ansible playbooks nhztleaw bin ansible python version feb configuration paste below plugin aws cache yes regions us east filters instance state name running hostnames name private ip address prefix tag name keyed groups prefix tag key tags compose use the private ip address to connect to the host note this does not modify inventory hostname which is set via i hostnames ansible host private ip address os environment mac os steps to reproduce create an instance and add a tag foo with no value create the inventory and notice that the keyed group created for that tag has a trailing underscore clear out inventory cache before running this yaml ansible inventory list less search for tag foo and notice that it has a trailing slash expected results expect to have a tag without a trailing slash actual results the tag has a trailing slash tag foo hosts machine name paste below ansible tag foo m ping that will ping the machine with that tag and confirm the group can be used with the trailing slash but imo the trailing slash should not be there thank you | 1 |
404,815 | 11,863,329,366 | IssuesEvent | 2020-03-25 19:31:08 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | closed | Investigate where the duplicate appeal came from while there is an active supplemental claim | Feature: caseflow-decisions Priority: Medium Team: Foxtrot 🦊 Type: Investigation | Reference: https://dsva.slack.com/archives/C6E41RE92/p1558027963153500 | 1.0 | Investigate where the duplicate appeal came from while there is an active supplemental claim - Reference: https://dsva.slack.com/archives/C6E41RE92/p1558027963153500 | priority | investigate where the duplicate appeal came from while there is an active supplemental claim reference | 1 |
718,969 | 24,739,595,896 | IssuesEvent | 2022-10-21 03:06:06 | tensorchord/envd | https://api.github.com/repos/tensorchord/envd | closed | feat(CLI): Support positional args | help wanted 🆘 priority/2-medium 💛 type/feature 💡 | ## Description
Positional args like `docker exec <container> bash -c xxx` is user-friendly. The current design `envd run --comand "bash -c xxx"` is not great.
But the CLI framework we used does not support this.
- https://github.com/urfave/cli/issues/1074
- https://github.com/urfave/cli/issues/1237
---
<!-- Issue Author: Don't delete this message to encourage other users to support your issue! -->
**Message from the maintainers**:
Love this enhancement proposal? Give it a 👍. We prioritise the proposals with the most 👍.
| 1.0 | feat(CLI): Support positional args - ## Description
Positional args like `docker exec <container> bash -c xxx` is user-friendly. The current design `envd run --comand "bash -c xxx"` is not great.
But the CLI framework we used does not support this.
- https://github.com/urfave/cli/issues/1074
- https://github.com/urfave/cli/issues/1237
---
<!-- Issue Author: Don't delete this message to encourage other users to support your issue! -->
**Message from the maintainers**:
Love this enhancement proposal? Give it a 👍. We prioritise the proposals with the most 👍.
| priority | feat cli support positional args description positional args like docker exec bash c xxx is user friendly the current design envd run comand bash c xxx is not great but the cli framework we used does not support this message from the maintainers love this enhancement proposal give it a 👍 we prioritise the proposals with the most 👍 | 1 |
212,640 | 7,239,376,175 | IssuesEvent | 2018-02-13 17:17:32 | ansible/awx | https://api.github.com/repos/ansible/awx | opened | "Cleanup Fact Details" doesn't | component:api priority:medium type:bug | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
- API
##### SUMMARY
The 'cleanup facts' management job cleans up old system tracking facts... which is not something AWX even collects.
It should be updated to purge the fact cache.
##### ENVIRONMENT
* AWX version: current
##### STEPS TO REPRODUCE
run 'cleanup facts' with any timeframe
##### EXPECTED RESULTS
Facts are cleaned up.
##### ACTUAL RESULTS
Nothing.
##### ADDITIONAL INFORMATION
| 1.0 | "Cleanup Fact Details" doesn't - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
- API
##### SUMMARY
The 'cleanup facts' management job cleans up old system tracking facts... which is not something AWX even collects.
It should be updated to purge the fact cache.
##### ENVIRONMENT
* AWX version: current
##### STEPS TO REPRODUCE
run 'cleanup facts' with any timeframe
##### EXPECTED RESULTS
Facts are cleaned up.
##### ACTUAL RESULTS
Nothing.
##### ADDITIONAL INFORMATION
| priority | cleanup fact details doesn t issue type bug report component name api summary the cleanup facts management job cleans up old system tracking facts which is not something awx even collects it should be updated to purge the fact cache environment awx version current steps to reproduce run cleanup facts with any timeframe expected results facts are cleaned up actual results nothing additional information | 1 |
426,532 | 12,373,553,844 | IssuesEvent | 2020-05-18 22:52:20 | Matteas-Eden/roll-for-reaction | https://api.github.com/repos/Matteas-Eden/roll-for-reaction | closed | Load game crashes the game | Medium Priority bug | **Bug Summary**
<!-- A concise description of what the bug is -->
Loading a JSON file when starting the game causes it to crash and restart.
**Test Case(s)**
<!-- List the relevant unsuccessful test case(s) -->
---
**Expected behavior**
<!-- Describe what you expected to happen -->
Load the game to continue playing from saved point
**Observed Behaviour**
<!-- Describe the observed behaviour of the bug -->
Restarts the game
**Steps To Reproduce**
<!-- The steps performed to reproduce the bug -->
1. Save game at some point while playing
2. Load game when at main menu
**Environment**
- Version: 4c7bef8
- OS: All
- Browser: All
---
**Code Examples**
**Stack Trace**
**Screenshots**
**Error Report**
---
**Additional context**
<!-- Add any other context about the problem here -->
| 1.0 | Load game crashes the game - **Bug Summary**
<!-- A concise description of what the bug is -->
Loading a JSON file when starting the game causes it to crash and restart.
**Test Case(s)**
<!-- List the relevant unsuccessful test case(s) -->
---
**Expected behavior**
<!-- Describe what you expected to happen -->
Load the game to continue playing from saved point
**Observed Behaviour**
<!-- Describe the observed behaviour of the bug -->
Restarts the game
**Steps To Reproduce**
<!-- The steps performed to reproduce the bug -->
1. Save game at some point while playing
2. Load game when at main menu
**Environment**
- Version: 4c7bef8
- OS: All
- Browser: All
---
**Code Examples**
**Stack Trace**
**Screenshots**
**Error Report**
---
**Additional context**
<!-- Add any other context about the problem here -->
| priority | load game crashes the game bug summary loading a json file when starting the game causes it to crash and restart test case s expected behavior load the game to continue playing from saved point observed behaviour restarts the game steps to reproduce save game at some point while playing load game when at main menu environment version os all browser all code examples stack trace screenshots error report additional context | 1 |
199,480 | 6,989,798,025 | IssuesEvent | 2017-12-14 17:16:38 | salesagility/SuiteCRM | https://api.github.com/repos/salesagility/SuiteCRM | closed | SuiteP - Report PDF file - names of columns are missing | bug Fix Proposed Medium Priority Resolved: Next Release | <!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug -->
Reports for SuiteR are much nicer and have the names of culumns in PDF Report file

Look the diffrence between the SuiteP and SuiteR theme
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
- SuiteCRM Version used: 7.7.3
- Browser name and version : Chrome 53.xx m
- Environment name and version: MySQL 5.6, PHP 5.6
- Operating System and version: Linux - Hosting
| 1.0 | SuiteP - Report PDF file - names of columns are missing - <!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug -->
Reports for SuiteR are much nicer and have the names of culumns in PDF Report file

Look the diffrence between the SuiteP and SuiteR theme
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
- SuiteCRM Version used: 7.7.3
- Browser name and version : Chrome 53.xx m
- Environment name and version: MySQL 5.6, PHP 5.6
- Operating System and version: Linux - Hosting
| priority | suitep report pdf file names of columns are missing issue reports for suiter are much nicer and have the names of culumns in pdf report file look the diffrence between the suitep and suiter theme your environment suitecrm version used browser name and version chrome xx m environment name and version mysql php operating system and version linux hosting | 1 |
503,277 | 14,582,890,786 | IssuesEvent | 2020-12-18 13:08:51 | cesko-digital/web | https://api.github.com/repos/cesko-digital/web | closed | Implementace patičky webu | frontend medium priority rework | **Technologie:** React, TypeScript, Gatsby
V rámci #64
Vytvořit komponentu dle designu. Komponenta nemusí mít žádný vstup, ale je nutné počítat s překlady (#68).
Komponenta by měla využít komponentu tlačítka z #65.
**Figma:** [link](https://www.figma.com/file/8YdGIHXLPo8zt4ZN8KJ4Hy/%3EC.D-%2F-Web?node-id=40%3A6058)
V případě nejasností nebojte se na mne obrátit. 🙂
<img width="1120" alt="Screenshot 2020-10-22 at 20 45 17" src="https://user-images.githubusercontent.com/10332399/96916379-c741bc00-14a7-11eb-8611-4b88d2e1ed11.png">
| 1.0 | Implementace patičky webu - **Technologie:** React, TypeScript, Gatsby
V rámci #64
Vytvořit komponentu dle designu. Komponenta nemusí mít žádný vstup, ale je nutné počítat s překlady (#68).
Komponenta by měla využít komponentu tlačítka z #65.
**Figma:** [link](https://www.figma.com/file/8YdGIHXLPo8zt4ZN8KJ4Hy/%3EC.D-%2F-Web?node-id=40%3A6058)
V případě nejasností nebojte se na mne obrátit. 🙂
<img width="1120" alt="Screenshot 2020-10-22 at 20 45 17" src="https://user-images.githubusercontent.com/10332399/96916379-c741bc00-14a7-11eb-8611-4b88d2e1ed11.png">
| priority | implementace patičky webu technologie react typescript gatsby v rámci vytvořit komponentu dle designu komponenta nemusí mít žádný vstup ale je nutné počítat s překlady komponenta by měla využít komponentu tlačítka z figma v případě nejasností nebojte se na mne obrátit 🙂 img width alt screenshot at src | 1 |
769,171 | 26,995,100,762 | IssuesEvent | 2023-02-09 23:54:22 | dr3ams/Roguelike-Adventures-and-Dungeons-2 | https://api.github.com/repos/dr3ams/Roguelike-Adventures-and-Dungeons-2 | reopened | [Bug]: Alchemy skill not present | bug Unable To Fix Medium Priority Fix Incoming Known Issue | ### General Info
- [X] I didn't added any mods
- [X] I can reproduce this issue consistently in single-player
- [X] I can reproduce this issue consistently in multi-player
- [X] I have searched for this issue previously and it was either (1) not previously reported, or (2) previously fixed and I am having the same problem.
- [ ] I am crashing and can provide my crash report(s)
- [X] I am using the latest version of the modpack
### Your launcher
Curseforge
### Modpack version
RAD2 - 0.96a [MC1.16.5]
### Describe your issue
I'm not getting any Alchemy exp, it's not present in the skill list either. Can't progress healer tree.
### Steps to reproduce the issue
_No response_
### Additional Information
Pack is not crashing.. | 1.0 | [Bug]: Alchemy skill not present - ### General Info
- [X] I didn't added any mods
- [X] I can reproduce this issue consistently in single-player
- [X] I can reproduce this issue consistently in multi-player
- [X] I have searched for this issue previously and it was either (1) not previously reported, or (2) previously fixed and I am having the same problem.
- [ ] I am crashing and can provide my crash report(s)
- [X] I am using the latest version of the modpack
### Your launcher
Curseforge
### Modpack version
RAD2 - 0.96a [MC1.16.5]
### Describe your issue
I'm not getting any Alchemy exp, it's not present in the skill list either. Can't progress healer tree.
### Steps to reproduce the issue
_No response_
### Additional Information
Pack is not crashing.. | priority | alchemy skill not present general info i didn t added any mods i can reproduce this issue consistently in single player i can reproduce this issue consistently in multi player i have searched for this issue previously and it was either not previously reported or previously fixed and i am having the same problem i am crashing and can provide my crash report s i am using the latest version of the modpack your launcher curseforge modpack version describe your issue i m not getting any alchemy exp it s not present in the skill list either can t progress healer tree steps to reproduce the issue no response additional information pack is not crashing | 1 |
211,193 | 7,199,141,086 | IssuesEvent | 2018-02-05 15:07:18 | CherryPhil/iDiet | https://api.github.com/repos/CherryPhil/iDiet | closed | Slideshow Fun Fact (Fun Page) | Medium Priority Normal help wanted | As an inexperienced person, I want to learn more about dieting in a fun and engaging way so that I can start committing to my diet and hopefully become a healthy person. | 1.0 | Slideshow Fun Fact (Fun Page) - As an inexperienced person, I want to learn more about dieting in a fun and engaging way so that I can start committing to my diet and hopefully become a healthy person. | priority | slideshow fun fact fun page as an inexperienced person i want to learn more about dieting in a fun and engaging way so that i can start committing to my diet and hopefully become a healthy person | 1 |
30,708 | 2,724,927,533 | IssuesEvent | 2015-04-14 20:36:49 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | Lack of styles for widgets: ScrollBanner / gwtTextArea | bug imported Priority-Medium TargetVersion-5.1.1 | _From [claudio....@cruxframework.org](https://code.google.com/u/102254381191677355567/) on May 23, 2014 14:50:56_
Lack of styles for widgets: ScrollBanner / gwtTextArea
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=372_ | 1.0 | Lack of styles for widgets: ScrollBanner / gwtTextArea - _From [claudio....@cruxframework.org](https://code.google.com/u/102254381191677355567/) on May 23, 2014 14:50:56_
Lack of styles for widgets: ScrollBanner / gwtTextArea
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=372_ | priority | lack of styles for widgets scrollbanner gwttextarea from on may lack of styles for widgets scrollbanner gwttextarea original issue | 1 |
146,301 | 5,615,054,035 | IssuesEvent | 2017-04-03 13:50:37 | mPowering/django-orb | https://api.github.com/repos/mPowering/django-orb | opened | Collection model - make translatable | enhancement medium priority | Allow the title and description fields to be translatable | 1.0 | Collection model - make translatable - Allow the title and description fields to be translatable | priority | collection model make translatable allow the title and description fields to be translatable | 1 |
800,400 | 28,364,492,602 | IssuesEvent | 2023-04-12 13:04:38 | sebastien-d-me/SebBlog | https://api.github.com/repos/sebastien-d-me/SebBlog | closed | Comment article | Priority: Medium Status: Completed Type : Front-end Type : Back-end | #### Description:
Creation of a system to comment the articles
------------
###### Estimated time: 2 day(s)
###### Difficulty: ⭐⭐
| 1.0 | Comment article - #### Description:
Creation of a system to comment the articles
------------
###### Estimated time: 2 day(s)
###### Difficulty: ⭐⭐
| priority | comment article description creation of a system to comment the articles estimated time day s difficulty ⭐⭐ | 1 |
152,933 | 5,871,959,560 | IssuesEvent | 2017-05-15 10:04:45 | yuanxin1997/FYPJ | https://api.github.com/repos/yuanxin1997/FYPJ | closed | UX on navigation to the backview | Enhancement Medium Priority | Proper implementation:
Go back to the back view instead of going to a new state after adding items to cart | 1.0 | UX on navigation to the backview - Proper implementation:
Go back to the back view instead of going to a new state after adding items to cart | priority | ux on navigation to the backview proper implementation go back to the back view instead of going to a new state after adding items to cart | 1 |
66,840 | 3,263,095,489 | IssuesEvent | 2015-10-22 01:22:15 | cs2103aug2015-t16-4j/main | https://api.github.com/repos/cs2103aug2015-t16-4j/main | closed | As a user, Jim wants to be able to categorize tasks according to importance levels | priority.medium type.story | So that he can tell what is more important to do at a glance | 1.0 | As a user, Jim wants to be able to categorize tasks according to importance levels - So that he can tell what is more important to do at a glance | priority | as a user jim wants to be able to categorize tasks according to importance levels so that he can tell what is more important to do at a glance | 1 |
388,464 | 11,488,070,062 | IssuesEvent | 2020-02-11 13:15:12 | DigitalCampus/django-oppia | https://api.github.com/repos/DigitalCampus/django-oppia | closed | Upload Media - allow uploading of custom image | enhancement medium priority | In case none of the auto generated images are usable | 1.0 | Upload Media - allow uploading of custom image - In case none of the auto generated images are usable | priority | upload media allow uploading of custom image in case none of the auto generated images are usable | 1 |
274,796 | 8,567,969,927 | IssuesEvent | 2018-11-10 17:03:11 | CS2113-AY1819S1-W13-2/main | https://api.github.com/repos/CS2113-AY1819S1-W13-2/main | opened | edit command throws wrong exceptions | bug priority.medium severity.Medium | add ds/11112018 ts/0700 de/11112018 te/0705 ss/0702 d/16 l/Sentosa tz/+8
edit 1 d/43
edit 1 te/2300
edit 1 d/43 should say "dive is too long and too deep, max you can go is 42m"
edit 1 te/2300 should say "dive is too long and too deep"
but now both of them are throwing "Invalid time format! (Not 24Hr format) " | 1.0 | edit command throws wrong exceptions - add ds/11112018 ts/0700 de/11112018 te/0705 ss/0702 d/16 l/Sentosa tz/+8
edit 1 d/43
edit 1 te/2300
edit 1 d/43 should say "dive is too long and too deep, max you can go is 42m"
edit 1 te/2300 should say "dive is too long and too deep"
but now both of them are throwing "Invalid time format! (Not 24Hr format) " | priority | edit command throws wrong exceptions add ds ts de te ss d l sentosa tz edit d edit te edit d should say dive is too long and too deep max you can go is edit te should say dive is too long and too deep but now both of them are throwing invalid time format not format | 1 |
354,426 | 10,567,298,855 | IssuesEvent | 2019-10-06 02:39:28 | AY1920S1-CS2103T-T11-4/main | https://api.github.com/repos/AY1920S1-CS2103T-T11-4/main | closed | Start implementing a minimal version of your feature | priority.Medium type.Task | - Start implementing minimal version of features assigned as discussed if time permits | 1.0 | Start implementing a minimal version of your feature - - Start implementing minimal version of features assigned as discussed if time permits | priority | start implementing a minimal version of your feature start implementing minimal version of features assigned as discussed if time permits | 1 |
502,433 | 14,546,145,607 | IssuesEvent | 2020-12-15 20:46:04 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.9.1.8 beta release-130] Make government accounts into a pickable account to read for inclusion/exclusion of wealth related laws | Category: Gameplay Category: Laws Priority: Medium | Currently there is no possible way to read government accounts or Treasury in particular for pointing to wealth values or levels in law creation. | 1.0 | [0.9.1.8 beta release-130] Make government accounts into a pickable account to read for inclusion/exclusion of wealth related laws - Currently there is no possible way to read government accounts or Treasury in particular for pointing to wealth values or levels in law creation. | priority | make government accounts into a pickable account to read for inclusion exclusion of wealth related laws currently there is no possible way to read government accounts or treasury in particular for pointing to wealth values or levels in law creation | 1 |
128,502 | 5,067,960,663 | IssuesEvent | 2016-12-24 09:12:08 | buttercup-pw/buttercup-core | https://api.github.com/repos/buttercup-pw/buttercup-core | opened | Change Credentials to use default values + meta only | Priority: Medium Status: Available Type: Enhancement | Currently any number of properties can be added as 'data' to Credentials.. but they already support meta. Remove the support for the data blob and allow only default credential properties (username, password, keyfile). | 1.0 | Change Credentials to use default values + meta only - Currently any number of properties can be added as 'data' to Credentials.. but they already support meta. Remove the support for the data blob and allow only default credential properties (username, password, keyfile). | priority | change credentials to use default values meta only currently any number of properties can be added as data to credentials but they already support meta remove the support for the data blob and allow only default credential properties username password keyfile | 1 |
382,526 | 11,307,572,768 | IssuesEvent | 2020-01-18 21:54:22 | PREreview/prereview-standup | https://api.github.com/repos/PREreview/prereview-standup | closed | Making website responsive to display correctly in different devices | priority / medium scope / development type / bug type / enhancement | The website currently isn’t responsive and doesn’t display properly on mobile or screens of a narrow width. | 1.0 | Making website responsive to display correctly in different devices - The website currently isn’t responsive and doesn’t display properly on mobile or screens of a narrow width. | priority | making website responsive to display correctly in different devices the website currently isn’t responsive and doesn’t display properly on mobile or screens of a narrow width | 1 |
721,687 | 24,834,711,391 | IssuesEvent | 2022-10-26 07:55:17 | AY2223S1-CS2113-F11-4/tp | https://api.github.com/repos/AY2223S1-CS2113-F11-4/tp | closed | Modify Parser class to handle for editReason, deleteReason, viewVisit functions in VisitList | priority.Medium | 1. editReason - modified ver now only accepts ui, int index (referring to overall index of visit in visitList), and String reason
- enforce that reason cannot be ""
2. deleteReason - only accept ui and int index
3. viewVisit - modified version now only accepts ui, and int index | 1.0 | Modify Parser class to handle for editReason, deleteReason, viewVisit functions in VisitList - 1. editReason - modified ver now only accepts ui, int index (referring to overall index of visit in visitList), and String reason
- enforce that reason cannot be ""
2. deleteReason - only accept ui and int index
3. viewVisit - modified version now only accepts ui, and int index | priority | modify parser class to handle for editreason deletereason viewvisit functions in visitlist editreason modified ver now only accepts ui int index referring to overall index of visit in visitlist and string reason enforce that reason cannot be deletereason only accept ui and int index viewvisit modified version now only accepts ui and int index | 1 |
717,104 | 24,661,673,498 | IssuesEvent | 2022-10-18 07:13:34 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | closed | Container could not be removed | Type: Bug Priority: Medium Status: Available | ### Describe the bug
Error when trying to upgrade container.
### Steps to reproduce
docker run --rm --name watchtower2 -v /var/run/docker.sock:/var/run/docker.sock containrrr/watchtower --debug --run-once
### Expected behavior
It should upgrade the container without error.
### Screenshots
_No response_
### Environment
- Architecture: Linux iot 5.10.0-16-amd64 #1 SMP Debian 5.10.127-2 (2022-07-23) x86_64 GNU/Linux
- Docker Version: Docker version 20.10.5+dfsg1, build 55c4c88
### Your logs
```text
time="2022-10-18T05:53:14Z" level=info msg="Stopping /brian-node-rust (573b8d6e5cb4) with SIGTERM"
time="2022-10-18T05:53:24Z" level=debug msg="AutoRemove container 573b8d6e5cb4, skipping ContainerRemove call."
time="2022-10-18T05:53:34Z" level=error msg="container /brian-node-rust (573b8d6e5cb4) could not be removed"
```
### Additional context
Have looked at https://github.com/containrrr/watchtower/issues/1268, tried increasing the timeout, it didn't help. In any case the container should stop immediately. | 1.0 | Container could not be removed - ### Describe the bug
Error when trying to upgrade container.
### Steps to reproduce
docker run --rm --name watchtower2 -v /var/run/docker.sock:/var/run/docker.sock containrrr/watchtower --debug --run-once
### Expected behavior
It should upgrade the container without error.
### Screenshots
_No response_
### Environment
- Architecture: Linux iot 5.10.0-16-amd64 #1 SMP Debian 5.10.127-2 (2022-07-23) x86_64 GNU/Linux
- Docker Version: Docker version 20.10.5+dfsg1, build 55c4c88
### Your logs
```text
time="2022-10-18T05:53:14Z" level=info msg="Stopping /brian-node-rust (573b8d6e5cb4) with SIGTERM"
time="2022-10-18T05:53:24Z" level=debug msg="AutoRemove container 573b8d6e5cb4, skipping ContainerRemove call."
time="2022-10-18T05:53:34Z" level=error msg="container /brian-node-rust (573b8d6e5cb4) could not be removed"
```
### Additional context
Have looked at https://github.com/containrrr/watchtower/issues/1268, tried increasing the timeout, it didn't help. In any case the container should stop immediately. | priority | container could not be removed describe the bug error when trying to upgrade container steps to reproduce docker run rm name v var run docker sock var run docker sock containrrr watchtower debug run once expected behavior it should upgrade the container without error screenshots no response environment architecture linux iot smp debian gnu linux docker version docker version build your logs text time level info msg stopping brian node rust with sigterm time level debug msg autoremove container skipping containerremove call time level error msg container brian node rust could not be removed additional context have looked at tried increasing the timeout it didn t help in any case the container should stop immediately | 1 |
466,280 | 13,399,332,442 | IssuesEvent | 2020-09-03 14:21:12 | strapi/strapi | https://api.github.com/repos/strapi/strapi | closed | Relation field shows "No permission to view this field" for super-admin | priority: medium status: confirmed type: bug | **Describe the bug**
After upgrading to strapi 3.1 when trying to create a new entry with a relation field inside a repetable component. The field shown with a "No permission to view this field". Even tho i'm a super admin. Checked the settings page and the super admin seem to have access to the relation table.
**Steps to reproduce the behavior**
1. Create content list with a repeatable component that has a relation field
2. Uograde to strapi 3.1
3. try to create a new entry, the relation field will be disabled
**Expected behavior**
I was expectin to be able to modify the relation field as usual.
**Screenshots**
<img width="1217" alt="Screen Shot 2020-07-28 at 17 07 14" src="https://user-images.githubusercontent.com/8872447/88677428-effb2500-d0f5-11ea-995c-49c1cb3e4697.png">
**System**
- Node.js version: v12.18.0
- NPM version: 6.14.4
- Strapi version: 3.1.1
- Database: Mongo DB
- Operating system: Mac OS
| 1.0 | Relation field shows "No permission to view this field" for super-admin - **Describe the bug**
After upgrading to strapi 3.1 when trying to create a new entry with a relation field inside a repetable component. The field shown with a "No permission to view this field". Even tho i'm a super admin. Checked the settings page and the super admin seem to have access to the relation table.
**Steps to reproduce the behavior**
1. Create content list with a repeatable component that has a relation field
2. Uograde to strapi 3.1
3. try to create a new entry, the relation field will be disabled
**Expected behavior**
I was expectin to be able to modify the relation field as usual.
**Screenshots**
<img width="1217" alt="Screen Shot 2020-07-28 at 17 07 14" src="https://user-images.githubusercontent.com/8872447/88677428-effb2500-d0f5-11ea-995c-49c1cb3e4697.png">
**System**
- Node.js version: v12.18.0
- NPM version: 6.14.4
- Strapi version: 3.1.1
- Database: Mongo DB
- Operating system: Mac OS
| priority | relation field shows no permission to view this field for super admin describe the bug after upgrading to strapi when trying to create a new entry with a relation field inside a repetable component the field shown with a no permission to view this field even tho i m a super admin checked the settings page and the super admin seem to have access to the relation table steps to reproduce the behavior create content list with a repeatable component that has a relation field uograde to strapi try to create a new entry the relation field will be disabled expected behavior i was expectin to be able to modify the relation field as usual screenshots img width alt screen shot at src system node js version npm version strapi version database mongo db operating system mac os | 1 |
43,588 | 2,889,868,367 | IssuesEvent | 2015-06-13 20:51:43 | damonkohler/android-scripting | https://api.github.com/repos/damonkohler/android-scripting | closed | Include pySerial | auto-migrated Priority-Medium Type-Enhancement | ```
Currently, only some rooted phones allow for serial commands out of the USB
port. On those it is non-trivial to get serial commands in and out of the
device using Python and ASE.
More devices may support serial and using Bluetooth serial might make it
handy to have pySerial in ASE.
```
Original issue reported on code.google.com by `rhickman` on 2 Mar 2010 at 9:44 | 1.0 | Include pySerial - ```
Currently, only some rooted phones allow for serial commands out of the USB
port. On those it is non-trivial to get serial commands in and out of the
device using Python and ASE.
More devices may support serial and using Bluetooth serial might make it
handy to have pySerial in ASE.
```
Original issue reported on code.google.com by `rhickman` on 2 Mar 2010 at 9:44 | priority | include pyserial currently only some rooted phones allow for serial commands out of the usb port on those it is non trivial to get serial commands in and out of the device using python and ase more devices may support serial and using bluetooth serial might make it handy to have pyserial in ase original issue reported on code google com by rhickman on mar at | 1 |
147,373 | 5,638,535,421 | IssuesEvent | 2017-04-06 12:14:51 | DiCarloLab-Delft/PycQED_py3 | https://api.github.com/repos/DiCarloLab-Delft/PycQED_py3 | opened | Analysis for motzoi for experiment with no calibration poitns | enhancement priority: should/medium | This will happen with experiments that use lookuptables. | 1.0 | Analysis for motzoi for experiment with no calibration poitns - This will happen with experiments that use lookuptables. | priority | analysis for motzoi for experiment with no calibration poitns this will happen with experiments that use lookuptables | 1 |
404,217 | 11,853,901,371 | IssuesEvent | 2020-03-24 23:10:10 | isi-vista/adam | https://api.github.com/repos/isi-vista/adam | closed | Automated curriculum-renderer | priority-0-high size-medium | Create a script (and make whatever necessary alterations to `make_scenes.py` to run through all of the scenes in the phase 1 curricula and output renderings for each frame.
- Ensure that the final outcome of the positioning system is what is rendered
- Have a mechanism for excluding scenes with features that aren't fully implemented or are currently buggy. | 1.0 | Automated curriculum-renderer - Create a script (and make whatever necessary alterations to `make_scenes.py` to run through all of the scenes in the phase 1 curricula and output renderings for each frame.
- Ensure that the final outcome of the positioning system is what is rendered
- Have a mechanism for excluding scenes with features that aren't fully implemented or are currently buggy. | priority | automated curriculum renderer create a script and make whatever necessary alterations to make scenes py to run through all of the scenes in the phase curricula and output renderings for each frame ensure that the final outcome of the positioning system is what is rendered have a mechanism for excluding scenes with features that aren t fully implemented or are currently buggy | 1 |
292,994 | 8,971,522,684 | IssuesEvent | 2019-01-29 16:05:48 | campus-bites/campus-bites-vue | https://api.github.com/repos/campus-bites/campus-bites-vue | opened | Incorporate new logo and font across the website | Priority: Medium Status: Available Type: Enhancement | We now have a new logo and font for Campus Bites which we can use across all of our platforms.

We should use the logo on the website and change the font for heading elements to match this logo.
| 1.0 | Incorporate new logo and font across the website - We now have a new logo and font for Campus Bites which we can use across all of our platforms.

We should use the logo on the website and change the font for heading elements to match this logo.
| priority | incorporate new logo and font across the website we now have a new logo and font for campus bites which we can use across all of our platforms we should use the logo on the website and change the font for heading elements to match this logo | 1 |
384,236 | 11,385,885,573 | IssuesEvent | 2020-01-29 12:05:45 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.0 staging-1350] JungleStew & JungleCampfireStew duplicate | Priority: Medium Status: Fixed | We have duplicate recipes. If we keep naming consistent, JungleStew should be deleted. JungleStew has the icon we need so we should assign it to JungleCampfireStew before deleting. | 1.0 | [0.9.0 staging-1350] JungleStew & JungleCampfireStew duplicate - We have duplicate recipes. If we keep naming consistent, JungleStew should be deleted. JungleStew has the icon we need so we should assign it to JungleCampfireStew before deleting. | priority | junglestew junglecampfirestew duplicate we have duplicate recipes if we keep naming consistent junglestew should be deleted junglestew has the icon we need so we should assign it to junglecampfirestew before deleting | 1 |
632,118 | 20,173,210,377 | IssuesEvent | 2022-02-10 12:20:12 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Assertion in nordic's BLE controller lll.c:352 | bug priority: medium area: Bluetooth platform: nRF area: Bluetooth Controller | **Describe the bug**
Time after time I catch next assert (after periodic public configuration and next once an hour)
```
00> ASSERTION FAIL @ WEST_TOPDIR/zephyr/subsys/bluetooth/controller/ll_sw/nordic/lll/lll.c:352
00> E: r0/a1: 0x00000003 r1/a2: 0x20009cf0 r2/a3: 0x00000001
00> E: r3/a4: 0x20009d46 r12/ip: 0xa0000000 r14/lr: 0x0002ecbb
00> E: xpsr: 0x61000011
00>...Code
```
**To Reproduce**
Assert occurs most often when I use 3+ mesh sensors in one sensor server (from NCS) with periodic publication configured.
**Expected behavior**
Expected periodic publishing without resets
**Impact**
Pre-release stopper
**Logs and console output**
RTT Log
```
00> I: 2 Sectors of 4096 bytes
00> I: alloc wra: 0, fb8
00> I: data wra: 0, 34
00> I: HW Platform: Nordic Semiconductor (0x0002)
00> I: HW Variant: nRF52x (0x0002)
00> I: Firmware: Standard Bluetooth controller (0x00) Version 2.6 Build 99
00> I: No ID address. App must call settings_load()
00> I: Bluetooth ready
00> D: Sensor 0x0052
00> D: Sensor 0x0054
00> D: Sensor 0x0057
00> D: Sensor 0x0059
00> D: Sensor 0x00a3
00> I: Identity: C9:C8:96:33:17:06 (random)
00> I: HCI: version 5.2 (0x0b) revision 0x0000, manufacturer 0x01db
00> I: LMP: version 5.2 (0x0b) subver 0xffff
00> I: Device UUID: 00000000-0000-0080-b149-e8a3dc72e362
00> I: Bluetooth mesh initialized
00> D: ctrl1=0x47 @tick=352898
00> I: Primary Element: 0x0008
00> D: net_idx 0x0000 flags 0x00 iv_index 0x0000
00> D: addr 0x0008 elem_count 1
00> D: addr 0x0008 mod_count 10 vnd_mod_count 0
00> D:
00> D: Sensor 0x0052
00> D: Sensor 0x0054
00> D: Sensor 0x0057
00> D: Sensor 0x0059
00> D: Sensor 0x00a3
00> I: Identity: C9:C8:96:33:17:06 (random)
00> I: HCI: version 5.2 (0x0b) revision 0x0000, manufacturer 0x01db
00> I: LMP: version 5.2 (0x0b) subver 0xffff
00> I: Device UUID: 00000000-0000-0080-b149-e8a3dc72e362
00> I: Bluetooth mesh initialized
00> D: ctrl1=0x47 @tick=352898
00> I: Primary Element: 0x0008
00> D: net_idx 0x0000 flags 0x00 iv_index 0x0000
00> D: addr 0x0008 elem_count 1
00> D: addr 0x0008 mod_count 10 vnd_mod_count 0
00> D:
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 3: 8008ff
00> D: OpCode 0x00008008
00> D: net_idx 0x0000 app_idx 0xfffe src 0x0001 len 1: ff
00> D: Composition page 255 not available
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 36: 0200db0110c1203e0a00070001000a00000002000c100010041006100710011000110111
00> W: No matching TX context for ack
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 2: 800c
00> D: OpCode 0x0000800c
00> D: net_idx 0x0000 app_idx 0xfffe src 0x0001 len 0:
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 3: 800e64
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 20: 000000003fe6b1c56ae39fff7f3d5062c9addef5
00> D: OpCode 0x00000000
00> D: AppIdx 0x0000 NetIdx 0x0000
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 6: 800300000000
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 3: 80240a
00> D: OpCode 0x00008024
00> D: net_idx 0x0000 app_idx 0xfffe src 0x0001 len 1: 0a
00> D: Transmit 0x0a (count 2 interval 20ms)
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 3: 80250a
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 8: 803d080000000011
00> D: OpCode 0x0000803d
00> D: ID 0x1100 addr 0x0008
00> D: model 0x20000c18 key_idx 0x000
00> D: status 0x00
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 9: 803e00080000000011
00> D: model key 0x0000
00> D: Stored bt/mesh/s/8/bind value
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 12: 03080000c00000ff4a090011
00> D: OpCode 0x00000003
00> D: elem_addr 0x0008 pub_addr 0xc000 cred_flag 0
00> D: pub_app_idx 0x000, pub_ttl 255 pub_period 0x4a
00> D: retransmit 0x09 (count 1 interval 100ms)
00> D: ID 0x1100 addr 0x0008
00> D: period 10000 ms
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 14: 801900080000c00000ff4a090011
00> W: No matching TX context for ack
00> D: Stored bt/mesh/s/8/pub value
00> D:
00> D: #1 Period: 10000 ms Divisor: 0 (normal)
00> W: Device INA219 is not ready
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 28 (28 - 0) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 4 (4 - 0) thrsh: 0
00> D: Accel value #0: 0,-38304
00> D: Accel value #1: -1,-264032
00> D: Accel value #2: 9,499392
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3205ms
00> D: Publishing next time in 6795ms
00> D:
00> D: #2 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-114912
00> D: Accel value #1: -1,-225728
00> D: Accel value #2: 9,537696
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3204ms
00> D: Publishing next time in 6796ms
00> D:
00> D: #3 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-153216
00> D: Accel value #1: -1,-225728
00> D: Accel value #2: 9,499392
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3204ms
00> D: Publishing next time in 6796ms
00> D:
00> D: #4 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-76608
00> D: Accel value #1: -1,-187424
00> D: Accel value #2: 9,576000
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3204ms
00> D: Publishing next time in 6796ms
00> D:
00> D: #5 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-38304
00> D: Accel value #1: -1,-302336
00> D: Accel value #2: 9,461088
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3204ms
00> D: Publishing next time in 6796ms
00> D:
00> D: #6 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 113
00> D: Temperature:28,250000
00> D: Temp sensor value: 28,250000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-38304
00> D: Accel value #1: -1,-225728
00> D: Accel value #2: 9,537696
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> ASSERTION FAIL @ WEST_TOPDIR/zephyr/subsys/bluetooth/controller/ll_sw/nordic/lll/lll.c:352
00> E: r0/a1: 0x00000003 r1/a2: 0x20009be0 r2/a3: 0x00000001
00> E: r3/a4: 0x20009c3e r12/ip: 0x00000000 r14/lr: 0x0002fd57
00> E: xpsr: 0x61000011
00> E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000
00> E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000
00> E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000
00> E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000
00> E: fpscr: 0x00031f06
00> E: Faulting instruction address (r15/pc): 0x0002fd62
00> E: >>> ZEPHYR FATAL ERROR 3: Kernel oops on CPU 0
00> E: Fault during interrupt handling
00>
00> E: Current thread: 0x20002bc8 (unknown)
00> E: Resetting system
00>
00>
```
**Environment (please complete the following information):**
* Board: Custom board based on nRF52840
* Zephyr: build: v2.6.99-ncs1-1
* NCS: 1.7.1
* Toolchain: gcc-arm-none-eabi-9-2019-q4-major \ gcc-arm-none-eabi-10-2020-q4-major
* CMAKE_VERSION 3.20.5
* West: version "0.10.1"
**Additional context**
Add any other context that could be relevant to your issue, such as pin setting,
target configuration, ...
| 1.0 | Assertion in nordic's BLE controller lll.c:352 - **Describe the bug**
Time after time I catch next assert (after periodic public configuration and next once an hour)
```
00> ASSERTION FAIL @ WEST_TOPDIR/zephyr/subsys/bluetooth/controller/ll_sw/nordic/lll/lll.c:352
00> E: r0/a1: 0x00000003 r1/a2: 0x20009cf0 r2/a3: 0x00000001
00> E: r3/a4: 0x20009d46 r12/ip: 0xa0000000 r14/lr: 0x0002ecbb
00> E: xpsr: 0x61000011
00>...Code
```
**To Reproduce**
Assert occurs most often when I use 3+ mesh sensors in one sensor server (from NCS) with periodic publication configured.
**Expected behavior**
Expected periodic publishing without resets
**Impact**
Pre-release stopper
**Logs and console output**
RTT Log
```
00> I: 2 Sectors of 4096 bytes
00> I: alloc wra: 0, fb8
00> I: data wra: 0, 34
00> I: HW Platform: Nordic Semiconductor (0x0002)
00> I: HW Variant: nRF52x (0x0002)
00> I: Firmware: Standard Bluetooth controller (0x00) Version 2.6 Build 99
00> I: No ID address. App must call settings_load()
00> I: Bluetooth ready
00> D: Sensor 0x0052
00> D: Sensor 0x0054
00> D: Sensor 0x0057
00> D: Sensor 0x0059
00> D: Sensor 0x00a3
00> I: Identity: C9:C8:96:33:17:06 (random)
00> I: HCI: version 5.2 (0x0b) revision 0x0000, manufacturer 0x01db
00> I: LMP: version 5.2 (0x0b) subver 0xffff
00> I: Device UUID: 00000000-0000-0080-b149-e8a3dc72e362
00> I: Bluetooth mesh initialized
00> D: ctrl1=0x47 @tick=352898
00> I: Primary Element: 0x0008
00> D: net_idx 0x0000 flags 0x00 iv_index 0x0000
00> D: addr 0x0008 elem_count 1
00> D: addr 0x0008 mod_count 10 vnd_mod_count 0
00> D:
00> D: Sensor 0x0052
00> D: Sensor 0x0054
00> D: Sensor 0x0057
00> D: Sensor 0x0059
00> D: Sensor 0x00a3
00> I: Identity: C9:C8:96:33:17:06 (random)
00> I: HCI: version 5.2 (0x0b) revision 0x0000, manufacturer 0x01db
00> I: LMP: version 5.2 (0x0b) subver 0xffff
00> I: Device UUID: 00000000-0000-0080-b149-e8a3dc72e362
00> I: Bluetooth mesh initialized
00> D: ctrl1=0x47 @tick=352898
00> I: Primary Element: 0x0008
00> D: net_idx 0x0000 flags 0x00 iv_index 0x0000
00> D: addr 0x0008 elem_count 1
00> D: addr 0x0008 mod_count 10 vnd_mod_count 0
00> D:
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 3: 8008ff
00> D: OpCode 0x00008008
00> D: net_idx 0x0000 app_idx 0xfffe src 0x0001 len 1: ff
00> D: Composition page 255 not available
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 36: 0200db0110c1203e0a00070001000a00000002000c100010041006100710011000110111
00> W: No matching TX context for ack
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 2: 800c
00> D: OpCode 0x0000800c
00> D: net_idx 0x0000 app_idx 0xfffe src 0x0001 len 0:
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 3: 800e64
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 20: 000000003fe6b1c56ae39fff7f3d5062c9addef5
00> D: OpCode 0x00000000
00> D: AppIdx 0x0000 NetIdx 0x0000
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 6: 800300000000
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 3: 80240a
00> D: OpCode 0x00008024
00> D: net_idx 0x0000 app_idx 0xfffe src 0x0001 len 1: 0a
00> D: Transmit 0x0a (count 2 interval 20ms)
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 3: 80250a
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 8: 803d080000000011
00> D: OpCode 0x0000803d
00> D: ID 0x1100 addr 0x0008
00> D: model 0x20000c18 key_idx 0x000
00> D: status 0x00
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 9: 803e00080000000011
00> D: model key 0x0000
00> D: Stored bt/mesh/s/8/bind value
00> D: app_idx 0xfffe src 0x0001 dst 0x0008
00> D: len 12: 03080000c00000ff4a090011
00> D: OpCode 0x00000003
00> D: elem_addr 0x0008 pub_addr 0xc000 cred_flag 0
00> D: pub_app_idx 0x000, pub_ttl 255 pub_period 0x4a
00> D: retransmit 0x09 (count 1 interval 100ms)
00> D: ID 0x1100 addr 0x0008
00> D: period 10000 ms
00> D: net_idx 0x0000 app_idx 0xfffe dst 0x0001
00> D: len 14: 801900080000c00000ff4a090011
00> W: No matching TX context for ack
00> D: Stored bt/mesh/s/8/pub value
00> D:
00> D: #1 Period: 10000 ms Divisor: 0 (normal)
00> W: Device INA219 is not ready
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 28 (28 - 0) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 4 (4 - 0) thrsh: 0
00> D: Accel value #0: 0,-38304
00> D: Accel value #1: -1,-264032
00> D: Accel value #2: 9,499392
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3205ms
00> D: Publishing next time in 6795ms
00> D:
00> D: #2 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-114912
00> D: Accel value #1: -1,-225728
00> D: Accel value #2: 9,537696
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3204ms
00> D: Publishing next time in 6796ms
00> D:
00> D: #3 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-153216
00> D: Accel value #1: -1,-225728
00> D: Accel value #2: 9,499392
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3204ms
00> D: Publishing next time in 6796ms
00> D:
00> D: #4 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-76608
00> D: Accel value #1: -1,-187424
00> D: Accel value #2: 9,576000
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3204ms
00> D: Publishing next time in 6796ms
00> D:
00> D: #5 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 114
00> D: Temperature:28,500000
00> D: Temp sensor value: 28,500000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-38304
00> D: Accel value #1: -1,-302336
00> D: Accel value #2: 9,461088
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> D: err 0
00> D: Publishing took 3204ms
00> D: Publishing next time in 6796ms
00> D:
00> D: #6 Period: 10000 ms Divisor: 0 (normal)
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: sample: 113
00> D: Temperature:28,250000
00> D: Temp sensor value: 28,250000
00> D: Delta: 0 (28 - 28) thrsh: 0
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: Delta: 0 (4 - 4) thrsh: 0
00> D: Accel value #0: 0,-38304
00> D: Accel value #1: -1,-225728
00> D: Accel value #2: 9,537696
00> D: Delta: 0 (0 - 0) thrsh: 0
00> D: err 0
00> D: Publishing next time in 100ms
00> D:
00> ASSERTION FAIL @ WEST_TOPDIR/zephyr/subsys/bluetooth/controller/ll_sw/nordic/lll/lll.c:352
00> E: r0/a1: 0x00000003 r1/a2: 0x20009be0 r2/a3: 0x00000001
00> E: r3/a4: 0x20009c3e r12/ip: 0x00000000 r14/lr: 0x0002fd57
00> E: xpsr: 0x61000011
00> E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000
00> E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000
00> E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000
00> E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000
00> E: fpscr: 0x00031f06
00> E: Faulting instruction address (r15/pc): 0x0002fd62
00> E: >>> ZEPHYR FATAL ERROR 3: Kernel oops on CPU 0
00> E: Fault during interrupt handling
00>
00> E: Current thread: 0x20002bc8 (unknown)
00> E: Resetting system
00>
00>
```
**Environment (please complete the following information):**
* Board: Custom board based on nRF52840
* Zephyr: build: v2.6.99-ncs1-1
* NCS: 1.7.1
* Toolchain: gcc-arm-none-eabi-9-2019-q4-major \ gcc-arm-none-eabi-10-2020-q4-major
* CMAKE_VERSION 3.20.5
* West: version "0.10.1"
**Additional context**
Add any other context that could be relevant to your issue, such as pin setting,
target configuration, ...
| priority | assertion in nordic s ble controller lll c describe the bug time after time i catch next assert after periodic public configuration and next once an hour assertion fail west topdir zephyr subsys bluetooth controller ll sw nordic lll lll c e e ip lr e xpsr code to reproduce assert occurs most often when i use mesh sensors in one sensor server from ncs with periodic publication configured expected behavior expected periodic publishing without resets impact pre release stopper logs and console output rtt log i sectors of bytes i alloc wra i data wra i hw platform nordic semiconductor i hw variant i firmware standard bluetooth controller version build i no id address app must call settings load i bluetooth ready d sensor d sensor d sensor d sensor d sensor i identity random i hci version revision manufacturer i lmp version subver i device uuid i bluetooth mesh initialized d tick i primary element d net idx flags iv index d addr elem count d addr mod count vnd mod count d d sensor d sensor d sensor d sensor d sensor i identity random i hci version revision manufacturer i lmp version subver i device uuid i bluetooth mesh initialized d tick i primary element d net idx flags iv index d addr elem count d addr mod count vnd mod count d d app idx src dst d len d opcode d net idx app idx src len ff d composition page not available d net idx app idx dst d len w no matching tx context for ack d app idx src dst d len d opcode d net idx app idx src len d net idx app idx dst d len d app idx src dst d len d opcode d appidx netidx d net idx app idx dst d len d app idx src dst d len d opcode d net idx app idx src len d transmit count interval d net idx app idx dst d len d app idx src dst d len d opcode d id addr d model key idx d status d net idx app idx dst d len d model key d stored bt mesh s bind value d app idx src dst d len d opcode d elem addr pub addr cred flag d pub app idx pub ttl pub period d retransmit count interval d id addr d period ms d net idx app idx dst d len w no matching tx context for ack d stored bt mesh s pub value d d period ms divisor normal w device is not ready d delta thrsh d sample d temperature d temp sensor value d delta thrsh d delta thrsh d delta thrsh d accel value d accel value d accel value d delta thrsh d err d publishing next time in d d err d publishing took d publishing next time in d d period ms divisor normal d delta thrsh d sample d temperature d temp sensor value d delta thrsh d delta thrsh d delta thrsh d accel value d accel value d accel value d delta thrsh d err d publishing next time in d d err d publishing took d publishing next time in d d period ms divisor normal d delta thrsh d sample d temperature d temp sensor value d delta thrsh d delta thrsh d delta thrsh d accel value d accel value d accel value d delta thrsh d err d publishing next time in d d err d publishing took d publishing next time in d d period ms divisor normal d delta thrsh d sample d temperature d temp sensor value d delta thrsh d delta thrsh d delta thrsh d accel value d accel value d accel value d delta thrsh d err d publishing next time in d d err d publishing took d publishing next time in d d period ms divisor normal d delta thrsh d sample d temperature d temp sensor value d delta thrsh d delta thrsh d delta thrsh d accel value d accel value d accel value d delta thrsh d err d publishing next time in d d err d publishing took d publishing next time in d d period ms divisor normal d delta thrsh d sample d temperature d temp sensor value d delta thrsh d delta thrsh d delta thrsh d accel value d accel value d accel value d delta thrsh d err d publishing next time in d assertion fail west topdir zephyr subsys bluetooth controller ll sw nordic lll lll c e e ip lr e xpsr e s s s s e s s s s e s s s s e s s s s e fpscr e faulting instruction address pc e zephyr fatal error kernel oops on cpu e fault during interrupt handling e current thread unknown e resetting system environment please complete the following information board custom board based on zephyr build ncs toolchain gcc arm none eabi major gcc arm none eabi major cmake version west version additional context add any other context that could be relevant to your issue such as pin setting target configuration | 1 |
222,394 | 7,431,859,941 | IssuesEvent | 2018-03-25 18:49:26 | ropensci/rrricanes | https://api.github.com/repos/ropensci/rrricanes | opened | Getting list of cyclones for a given year | Features Medium Priority | Currently, `rrricanes` uses [get_storms](https://github.com/ropensci/rrricanes/blob/master/R/get_storms.R#L43) to retrieve a list of storms for one or more years and one or more basins. The return is a dataframe object that includes a link for each cyclone to the cyclone's respective archive page. This value is passed to product functions that return data for the specific storm.
Per #113, in a move to implement the FTP server, this functionality would change quite a bit.
Depending on the year(s) requested, the function would access different URLs. It appears at this moment the "default" URL would be for the current or most recent year of tropical cyclones: ftp://ftp.nhc.noaa.gov/atcf/index/.
Several files depending on basin exist in this directory. For the time being, the concern would only be collecting AL and EP cyclones. Each basin has several text files depending on the weather office issuing the text products.
For example, at this moment, there are three text files for AL:
* AL_storms.txt.hpc - a list of storms (real cyclones and test cyclones) as issued by the Hydrometeorological Prediction Center. `rrricanes` does not access HPC products anyway so these can likely be ignored.
* AL_storms.txt.nhc - a list of storms (real and test) as issued by the NHC. These, we would want.
* AL_storms.txt.wpc - storms with advisories issued from the Weather Prediction Center.
Additional stations for other basins are:
* CPHC - Central Pacific Hurricane Center (may cover both EP and CP cyclones)
A master list exists that appears to list each cyclone's `Key` and `Name`, but no additional details contained in the other text files.
However, storm_list.txt exists which appears to be a consolidated text file of all cyclones going back to 1851, for all basins.
The caveat of all of these text files is that TEST, INVEST, and GENESIS cyclones are listed (TEST are not true cyclones and only for testing purposes. INVEST and GENESIS would not have the text products currently covered in `rrricanes`).
All text files are comma-delimited, so could easily be imported into rrricanes using `readr::read_csv()`.
According to [NRL](https://www.nrlmry.navy.mil/atcf_web/docs/database/new/descriptive.html), the columns are:
STORM NAME, RE, X, R2, R3,R4, R5, CY, YYYY, TY, I, YYY1MMDDHH, YYY2MMDDHH, SIZE, GENESIS_NUM, PAR1, PAR2, PRIORITY, STORM_STATE, WT_NUMBER, STORMID
The definitions are:
* STORM NAME = Literal storm name, "INVEST", or "GENESISxxx" where xxx is a number/P>
* RE = Region (basin) code: WP, IO, SH, CP, EP, AL, LS. (See [4. Data Format](https://www.nrlmry.navy.mil/atcf_web/docs/database/new/database.html#dataoverview))
* X = Subregion code: W, A, B, S, P, C, E, L, Q. In [Storm History Record Format](https://www.nrlmry.navy.mil/atcf_web/docs/database/new/abdeck.txt), these are listed as:
+ A - Arabian Sea
+ B - Bay of Bengal
+ C - Central Pacific
+ E - Eastern Pacific
+ L - Atlantic
+ P - South Pacific (135E - 120W)
+ Q - South Atlantic
+ S - South IO (20E - 135E)
+ W - Western Pacific
* R2 = Region 2 code: WP, IO, SH, CP, or EP. This and R3-R5 are codes for basins entered subsequent to the original basin where the storm was generated.
* R3 = Region 3 code: WP, IO, SH, CP, or EP.
* R4 = Region 4 code: WP, IO, SH, CP, or EP.
* R5 = Region 5 code: WP, IO, SH, CP, or EP.
* CY = Annual cyclone number: 01 through 99.
* YYYY = Cyclone Year: 0000 through 9999.
* TY = Highest level of tc development: TD, TS, TY, ST, TC, HU, SH, XX (unknown).
* I = S, R, O; straight mover, recurver, odd mover.
* YYY1MMDDHH = Starting DTG: 0000010100 through 9999123123.
* YYY2MMDDHH = Ending DTG: 0000010100 through 9999123123.
* SIZE = Storm Size (MIDG (midget) , GIAN (giant), etc.).
* GENESIS_NUM = Annual genesis number: 001 through 999.
* PAR1 = UNUSED.
* PAR2 = UNUSED.
* PRIORITY = Priority for model runs (e.g., GFDN, GFDL, COAMPS-TC, H-WRF): 1-9.
* STORM_STATE = Storm state: METWATCH,TCFA,WARNING or ARCHIVE
* WT_NUMBER = Minute of warning or TCFA (00-59)
* STORMID (or `Key`) = Storm ID composed of basin designator and annual cyclone number (e.g. wp081993)
For example, the current AL_storms.txt.nhc contains the row
```
ARLENE, AL, L, , , , , 01, 2017, EX, S, 2017041606, 2017042112, , , , , 8, , 1, AL012017
```
* STORM NAME: ARLENE
* RE: AL Atlantic
* X: L Atlantic
* R2: NA
* R3: NA
* R4: NA
* R5: NA
* CY: 01
* YYYY: 2017
* TY: EX
* I: S
* YYY1MMDDHH: 2017041601
* YYY2MMDDHH: 2017042112
* SIZE: NA
* GENESIS_NUM: NA
* PAR1: NA
* PAR2: NA
* PRIORITY: 8
* STORM_STATE: NA
* WT_NUMBER: 1
* STORMID: AL012017
| 1.0 | Getting list of cyclones for a given year - Currently, `rrricanes` uses [get_storms](https://github.com/ropensci/rrricanes/blob/master/R/get_storms.R#L43) to retrieve a list of storms for one or more years and one or more basins. The return is a dataframe object that includes a link for each cyclone to the cyclone's respective archive page. This value is passed to product functions that return data for the specific storm.
Per #113, in a move to implement the FTP server, this functionality would change quite a bit.
Depending on the year(s) requested, the function would access different URLs. It appears at this moment the "default" URL would be for the current or most recent year of tropical cyclones: ftp://ftp.nhc.noaa.gov/atcf/index/.
Several files depending on basin exist in this directory. For the time being, the concern would only be collecting AL and EP cyclones. Each basin has several text files depending on the weather office issuing the text products.
For example, at this moment, there are three text files for AL:
* AL_storms.txt.hpc - a list of storms (real cyclones and test cyclones) as issued by the Hydrometeorological Prediction Center. `rrricanes` does not access HPC products anyway so these can likely be ignored.
* AL_storms.txt.nhc - a list of storms (real and test) as issued by the NHC. These, we would want.
* AL_storms.txt.wpc - storms with advisories issued from the Weather Prediction Center.
Additional stations for other basins are:
* CPHC - Central Pacific Hurricane Center (may cover both EP and CP cyclones)
A master list exists that appears to list each cyclone's `Key` and `Name`, but no additional details contained in the other text files.
However, storm_list.txt exists which appears to be a consolidated text file of all cyclones going back to 1851, for all basins.
The caveat of all of these text files is that TEST, INVEST, and GENESIS cyclones are listed (TEST are not true cyclones and only for testing purposes. INVEST and GENESIS would not have the text products currently covered in `rrricanes`).
All text files are comma-delimited, so could easily be imported into rrricanes using `readr::read_csv()`.
According to [NRL](https://www.nrlmry.navy.mil/atcf_web/docs/database/new/descriptive.html), the columns are:
STORM NAME, RE, X, R2, R3,R4, R5, CY, YYYY, TY, I, YYY1MMDDHH, YYY2MMDDHH, SIZE, GENESIS_NUM, PAR1, PAR2, PRIORITY, STORM_STATE, WT_NUMBER, STORMID
The definitions are:
* STORM NAME = Literal storm name, "INVEST", or "GENESISxxx" where xxx is a number/P>
* RE = Region (basin) code: WP, IO, SH, CP, EP, AL, LS. (See [4. Data Format](https://www.nrlmry.navy.mil/atcf_web/docs/database/new/database.html#dataoverview))
* X = Subregion code: W, A, B, S, P, C, E, L, Q. In [Storm History Record Format](https://www.nrlmry.navy.mil/atcf_web/docs/database/new/abdeck.txt), these are listed as:
+ A - Arabian Sea
+ B - Bay of Bengal
+ C - Central Pacific
+ E - Eastern Pacific
+ L - Atlantic
+ P - South Pacific (135E - 120W)
+ Q - South Atlantic
+ S - South IO (20E - 135E)
+ W - Western Pacific
* R2 = Region 2 code: WP, IO, SH, CP, or EP. This and R3-R5 are codes for basins entered subsequent to the original basin where the storm was generated.
* R3 = Region 3 code: WP, IO, SH, CP, or EP.
* R4 = Region 4 code: WP, IO, SH, CP, or EP.
* R5 = Region 5 code: WP, IO, SH, CP, or EP.
* CY = Annual cyclone number: 01 through 99.
* YYYY = Cyclone Year: 0000 through 9999.
* TY = Highest level of tc development: TD, TS, TY, ST, TC, HU, SH, XX (unknown).
* I = S, R, O; straight mover, recurver, odd mover.
* YYY1MMDDHH = Starting DTG: 0000010100 through 9999123123.
* YYY2MMDDHH = Ending DTG: 0000010100 through 9999123123.
* SIZE = Storm Size (MIDG (midget) , GIAN (giant), etc.).
* GENESIS_NUM = Annual genesis number: 001 through 999.
* PAR1 = UNUSED.
* PAR2 = UNUSED.
* PRIORITY = Priority for model runs (e.g., GFDN, GFDL, COAMPS-TC, H-WRF): 1-9.
* STORM_STATE = Storm state: METWATCH,TCFA,WARNING or ARCHIVE
* WT_NUMBER = Minute of warning or TCFA (00-59)
* STORMID (or `Key`) = Storm ID composed of basin designator and annual cyclone number (e.g. wp081993)
For example, the current AL_storms.txt.nhc contains the row
```
ARLENE, AL, L, , , , , 01, 2017, EX, S, 2017041606, 2017042112, , , , , 8, , 1, AL012017
```
* STORM NAME: ARLENE
* RE: AL Atlantic
* X: L Atlantic
* R2: NA
* R3: NA
* R4: NA
* R5: NA
* CY: 01
* YYYY: 2017
* TY: EX
* I: S
* YYY1MMDDHH: 2017041601
* YYY2MMDDHH: 2017042112
* SIZE: NA
* GENESIS_NUM: NA
* PAR1: NA
* PAR2: NA
* PRIORITY: 8
* STORM_STATE: NA
* WT_NUMBER: 1
* STORMID: AL012017
| priority | getting list of cyclones for a given year currently rrricanes uses to retrieve a list of storms for one or more years and one or more basins the return is a dataframe object that includes a link for each cyclone to the cyclone s respective archive page this value is passed to product functions that return data for the specific storm per in a move to implement the ftp server this functionality would change quite a bit depending on the year s requested the function would access different urls it appears at this moment the default url would be for the current or most recent year of tropical cyclones ftp ftp nhc noaa gov atcf index several files depending on basin exist in this directory for the time being the concern would only be collecting al and ep cyclones each basin has several text files depending on the weather office issuing the text products for example at this moment there are three text files for al al storms txt hpc a list of storms real cyclones and test cyclones as issued by the hydrometeorological prediction center rrricanes does not access hpc products anyway so these can likely be ignored al storms txt nhc a list of storms real and test as issued by the nhc these we would want al storms txt wpc storms with advisories issued from the weather prediction center additional stations for other basins are cphc central pacific hurricane center may cover both ep and cp cyclones a master list exists that appears to list each cyclone s key and name but no additional details contained in the other text files however storm list txt exists which appears to be a consolidated text file of all cyclones going back to for all basins the caveat of all of these text files is that test invest and genesis cyclones are listed test are not true cyclones and only for testing purposes invest and genesis would not have the text products currently covered in rrricanes all text files are comma delimited so could easily be imported into rrricanes using readr read csv according to the columns are storm name re x cy yyyy ty i size genesis num priority storm state wt number stormid the definitions are storm name literal storm name invest or genesisxxx where xxx is a number p re region basin code wp io sh cp ep al ls see x subregion code w a b s p c e l q in these are listed as a arabian sea b bay of bengal c central pacific e eastern pacific l atlantic p south pacific q south atlantic s south io w western pacific region code wp io sh cp or ep this and are codes for basins entered subsequent to the original basin where the storm was generated region code wp io sh cp or ep region code wp io sh cp or ep region code wp io sh cp or ep cy annual cyclone number through yyyy cyclone year through ty highest level of tc development td ts ty st tc hu sh xx unknown i s r o straight mover recurver odd mover starting dtg through ending dtg through size storm size midg midget gian giant etc genesis num annual genesis number through unused unused priority priority for model runs e g gfdn gfdl coamps tc h wrf storm state storm state metwatch tcfa warning or archive wt number minute of warning or tcfa stormid or key storm id composed of basin designator and annual cyclone number e g for example the current al storms txt nhc contains the row arlene al l ex s storm name arlene re al atlantic x l atlantic na na na na cy yyyy ty ex i s size na genesis num na na na priority storm state na wt number stormid | 1 |
823,042 | 30,925,086,680 | IssuesEvent | 2023-08-06 11:41:51 | NotVeryAsh/IdeaHub | https://api.github.com/repos/NotVeryAsh/IdeaHub | opened | Dashboard functionality | new feature medium priority | ## Description
Add more dashboard functionality including changing the view and adding tests
## What value does this add?
A better dashboard view / ui
## Technical details
Mostly focusing on the tests in this ticket - adding tests and adding basic functionality for the dashboard view since the [Tailwind UI Redesign Ticket](https://github.com/NotVeryAsh/IdeaHub/issues/11) is about redesigning all of the data
- Add tests
- Add relevant data in the dashboard
| 1.0 | Dashboard functionality - ## Description
Add more dashboard functionality including changing the view and adding tests
## What value does this add?
A better dashboard view / ui
## Technical details
Mostly focusing on the tests in this ticket - adding tests and adding basic functionality for the dashboard view since the [Tailwind UI Redesign Ticket](https://github.com/NotVeryAsh/IdeaHub/issues/11) is about redesigning all of the data
- Add tests
- Add relevant data in the dashboard
| priority | dashboard functionality description add more dashboard functionality including changing the view and adding tests what value does this add a better dashboard view ui technical details mostly focusing on the tests in this ticket adding tests and adding basic functionality for the dashboard view since the is about redesigning all of the data add tests add relevant data in the dashboard | 1 |
466,340 | 13,400,297,071 | IssuesEvent | 2020-09-03 15:37:28 | HYPERNETS/hypernets_processor | https://api.github.com/repos/HYPERNETS/hypernets_processor | closed | Define what config information is required by processor | operational priority: medium | Should go into one of the files:
* [jobs.list](https://github.com/HYPERNETS/hypernets_processor/blob/master/hypernets_processor/cli/tests/config_files/jobs.list)
* [processor.config](https://github.com/HYPERNETS/hypernets_processor/blob/master/hypernets_processor/cli/tests/config_files/processor.config)
* [job.config](https://github.com/HYPERNETS/hypernets_processor/blob/master/hypernets_processor/cli/tests/config_files/job1.config)
* [scheduler.config](https://github.com/HYPERNETS/hypernets_processor/blob/master/hypernets_processor/cli/tests/config_files/scheduler.config)
Items to include should be things like file paths and processing options.
Once this is done the code should be updated to reflect this (e.g. config file readers). | 1.0 | Define what config information is required by processor - Should go into one of the files:
* [jobs.list](https://github.com/HYPERNETS/hypernets_processor/blob/master/hypernets_processor/cli/tests/config_files/jobs.list)
* [processor.config](https://github.com/HYPERNETS/hypernets_processor/blob/master/hypernets_processor/cli/tests/config_files/processor.config)
* [job.config](https://github.com/HYPERNETS/hypernets_processor/blob/master/hypernets_processor/cli/tests/config_files/job1.config)
* [scheduler.config](https://github.com/HYPERNETS/hypernets_processor/blob/master/hypernets_processor/cli/tests/config_files/scheduler.config)
Items to include should be things like file paths and processing options.
Once this is done the code should be updated to reflect this (e.g. config file readers). | priority | define what config information is required by processor should go into one of the files items to include should be things like file paths and processing options once this is done the code should be updated to reflect this e g config file readers | 1 |
48,471 | 2,998,193,399 | IssuesEvent | 2015-07-23 12:49:41 | jayway/powermock | https://api.github.com/repos/jayway/powermock | opened | MockGateway is loaded by the wrong classloader | bug imported Milestone-Release2.0 Priority-Medium | _From [johan.ha...@gmail.com](https://code.google.com/u/105676376875942041029/) on January 24, 2009 20:00:36_
It should be loaded by the mock test classloader. However
@PrepareEverythingForTest will fail if this is done. Investigate why.
_Original issue: http://code.google.com/p/powermock/issues/detail?id=91_ | 1.0 | MockGateway is loaded by the wrong classloader - _From [johan.ha...@gmail.com](https://code.google.com/u/105676376875942041029/) on January 24, 2009 20:00:36_
It should be loaded by the mock test classloader. However
@PrepareEverythingForTest will fail if this is done. Investigate why.
_Original issue: http://code.google.com/p/powermock/issues/detail?id=91_ | priority | mockgateway is loaded by the wrong classloader from on january it should be loaded by the mock test classloader however prepareeverythingfortest will fail if this is done investigate why original issue | 1 |
758,613 | 26,562,047,119 | IssuesEvent | 2023-01-20 16:39:21 | stats4sd/ccrp-soils | https://api.github.com/repos/stats4sd/ccrp-soils | closed | Make sample downloads more useful | Priority: High❕ Type:Enhancement Priority: Medium | For split downloads, add farmer_id and other relavent identifiers into each worksheet, to minimise the amount of merging / lookups that need to be added after the fact.
For wide downloads, investigate if we can get all samples repeated across the wide format. | 2.0 | Make sample downloads more useful - For split downloads, add farmer_id and other relavent identifiers into each worksheet, to minimise the amount of merging / lookups that need to be added after the fact.
For wide downloads, investigate if we can get all samples repeated across the wide format. | priority | make sample downloads more useful for split downloads add farmer id and other relavent identifiers into each worksheet to minimise the amount of merging lookups that need to be added after the fact for wide downloads investigate if we can get all samples repeated across the wide format | 1 |
682,649 | 23,351,487,776 | IssuesEvent | 2022-08-10 00:47:29 | space-syndicate/space-station-14 | https://api.github.com/repos/space-syndicate/space-station-14 | opened | Убрать пустые суффиксы из файлов локализации | Priority: 3-Medium | ### Описание
Изменить код скрипта на питоне в `/Tools/ss14-ru` на то, что бы он удалял теперь пустые суффиксы а не проставлял.
Так как в [#3114](https://github.com/space-wizards/RobustToolbox/pull/3114) проблема с суффиксами была исправлена.
### Шаги воспроизведения
_No response_
### Скриншоты
_No response_ | 1.0 | Убрать пустые суффиксы из файлов локализации - ### Описание
Изменить код скрипта на питоне в `/Tools/ss14-ru` на то, что бы он удалял теперь пустые суффиксы а не проставлял.
Так как в [#3114](https://github.com/space-wizards/RobustToolbox/pull/3114) проблема с суффиксами была исправлена.
### Шаги воспроизведения
_No response_
### Скриншоты
_No response_ | priority | убрать пустые суффиксы из файлов локализации описание изменить код скрипта на питоне в tools ru на то что бы он удалял теперь пустые суффиксы а не проставлял так как в проблема с суффиксами была исправлена шаги воспроизведения no response скриншоты no response | 1 |
30,084 | 2,722,220,009 | IssuesEvent | 2015-04-14 01:11:07 | CruxFramework/crux-smart-faces | https://api.github.com/repos/CruxFramework/crux-smart-faces | closed | Decrease the label size of the FileUploader component in mobile | bug imported Milestone-M14-C4 Module-CruxWidgets Priority-Medium TargetVersion-5.3.0 | _From [flavia.jesus@triggolabs.com](https://code.google.com/u/flavia.jesus@triggolabs.com/) on March 25, 2015 19:21:02_
Decrease the label size of the FileUploader component in mobile.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=664_ | 1.0 | Decrease the label size of the FileUploader component in mobile - _From [flavia.jesus@triggolabs.com](https://code.google.com/u/flavia.jesus@triggolabs.com/) on March 25, 2015 19:21:02_
Decrease the label size of the FileUploader component in mobile.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=664_ | priority | decrease the label size of the fileuploader component in mobile from on march decrease the label size of the fileuploader component in mobile original issue | 1 |
381,821 | 11,288,659,494 | IssuesEvent | 2020-01-16 08:27:03 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | USER ISSUE: Stuck in Blocks when placing stone/ore under yourself. | Priority: Medium | **Version:** 0.7.2.0 beta staging-f6bf50db
**Steps to Reproduce:**
1. Stand under a block (can be a ceiling or anything)
2. This block must be 3 blocks high so you can barely fit under it.
3. Pick up some stone or ores
4. Place it under you
**Expected behavior:**
Being pushed to the side to make room for the stone/ore
**Actual behavior:**
Pushed into the ceiling/block above without being able to get out. (if you don't have the right tools)
P.s. if you are pushed inside a block near a stockpile you can't remove the block you are in, when you click you go into the stockpile instead. | 1.0 | USER ISSUE: Stuck in Blocks when placing stone/ore under yourself. - **Version:** 0.7.2.0 beta staging-f6bf50db
**Steps to Reproduce:**
1. Stand under a block (can be a ceiling or anything)
2. This block must be 3 blocks high so you can barely fit under it.
3. Pick up some stone or ores
4. Place it under you
**Expected behavior:**
Being pushed to the side to make room for the stone/ore
**Actual behavior:**
Pushed into the ceiling/block above without being able to get out. (if you don't have the right tools)
P.s. if you are pushed inside a block near a stockpile you can't remove the block you are in, when you click you go into the stockpile instead. | priority | user issue stuck in blocks when placing stone ore under yourself version beta staging steps to reproduce stand under a block can be a ceiling or anything this block must be blocks high so you can barely fit under it pick up some stone or ores place it under you expected behavior being pushed to the side to make room for the stone ore actual behavior pushed into the ceiling block above without being able to get out if you don t have the right tools p s if you are pushed inside a block near a stockpile you can t remove the block you are in when you click you go into the stockpile instead | 1 |
226,380 | 7,518,589,981 | IssuesEvent | 2018-04-12 08:54:05 | 7ninjas/scss-mixins | https://api.github.com/repos/7ninjas/scss-mixins | closed | Incorrect docs description for triangle mixin | Priority: Medium Status: Closed | change docs example for triangle, right now it's css-triangle and in the code we have triangle | 1.0 | Incorrect docs description for triangle mixin - change docs example for triangle, right now it's css-triangle and in the code we have triangle | priority | incorrect docs description for triangle mixin change docs example for triangle right now it s css triangle and in the code we have triangle | 1 |
170,789 | 6,471,567,914 | IssuesEvent | 2017-08-17 11:59:59 | systers/powerup-android | https://api.github.com/repos/systers/powerup-android | closed | Developer needs UI elements for the School scenario | Priority: MEDIUM Program: GSoC17 Type: Enhancement | ## Description
As a developer,
I want to see the UIs of School scenario,
so that I can code the stage School scenario.
## Mocks
N/A
## Acceptance Criteria
-Background
-Redesign existing UI elements for avatar and Marcello.
-Redesign UI element for School
## Definition of Done
-Passes user testing per acceptance criteria items.
-Featured in design meeting.
-Approved by UI team. (Dil & May)
## Estimation
| 1.0 | Developer needs UI elements for the School scenario - ## Description
As a developer,
I want to see the UIs of School scenario,
so that I can code the stage School scenario.
## Mocks
N/A
## Acceptance Criteria
-Background
-Redesign existing UI elements for avatar and Marcello.
-Redesign UI element for School
## Definition of Done
-Passes user testing per acceptance criteria items.
-Featured in design meeting.
-Approved by UI team. (Dil & May)
## Estimation
| priority | developer needs ui elements for the school scenario description as a developer i want to see the uis of school scenario so that i can code the stage school scenario mocks n a acceptance criteria background redesign existing ui elements for avatar and marcello redesign ui element for school definition of done passes user testing per acceptance criteria items featured in design meeting approved by ui team dil may estimation | 1 |
40,385 | 2,868,916,388 | IssuesEvent | 2015-06-05 21:56:39 | dart-lang/dart_style | https://api.github.com/repos/dart-lang/dart_style | closed | Formatter problem with initializer list | AssumedStale bug Priority-Medium | <a href="https://github.com/Fox32"><img src="https://avatars.githubusercontent.com/u/648527?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [Fox32](https://github.com/Fox32)**
_Originally opened as dart-lang/sdk#17514_
----
Dart formatter is unable to format this code correctly:
Vector3(double x_, double y_, double z_): storage = new Float32List(3) {
setValues(x_, y_, z_);
}
If I give a hint by adding a line break before the ":", everything works fine. If I add a space instead, the space is removed.
Dart Editor version 1.3.0.dev_03_02 (DEV)
Dart SDK version 1.3.0-dev.3.2 | 1.0 | Formatter problem with initializer list - <a href="https://github.com/Fox32"><img src="https://avatars.githubusercontent.com/u/648527?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [Fox32](https://github.com/Fox32)**
_Originally opened as dart-lang/sdk#17514_
----
Dart formatter is unable to format this code correctly:
Vector3(double x_, double y_, double z_): storage = new Float32List(3) {
setValues(x_, y_, z_);
}
If I give a hint by adding a line break before the ":", everything works fine. If I add a space instead, the space is removed.
Dart Editor version 1.3.0.dev_03_02 (DEV)
Dart SDK version 1.3.0-dev.3.2 | priority | formatter problem with initializer list issue by originally opened as dart lang sdk dart formatter is unable to format this code correctly nbsp nbsp double x double y double z storage new nbsp nbsp nbsp nbsp setvalues x y z nbsp nbsp if i give a hint by adding a line break before the quot quot everything works fine if i add a space instead the space is removed dart editor version dev dev dart sdk version dev | 1 |
437,466 | 12,598,060,862 | IssuesEvent | 2020-06-11 01:50:00 | JuezUN/INGInious | https://api.github.com/repos/JuezUN/INGInious | closed | Upload multiple files - Task configuration | Change request Medium Priority Plugins Task | In the section to upload multiple files for a task, allow the user to specify the upload path. | 1.0 | Upload multiple files - Task configuration - In the section to upload multiple files for a task, allow the user to specify the upload path. | priority | upload multiple files task configuration in the section to upload multiple files for a task allow the user to specify the upload path | 1 |
533,462 | 15,591,414,325 | IssuesEvent | 2021-03-18 10:25:55 | visual-framework/vf-core | https://api.github.com/repos/visual-framework/vf-core | closed | Bug: vf-card vertically centres content | Priority: Medium Type: Bug | When several vf-card have `vf-card__content` of differing heights, the grid vertically centres.
As seen at https://stable.visual-framework.dev/

This is currently the probable solution, but needs testing (lest it introduce its own bugs):
```css
.vf-card {
align-content: start;
``` | 1.0 | Bug: vf-card vertically centres content - When several vf-card have `vf-card__content` of differing heights, the grid vertically centres.
As seen at https://stable.visual-framework.dev/

This is currently the probable solution, but needs testing (lest it introduce its own bugs):
```css
.vf-card {
align-content: start;
``` | priority | bug vf card vertically centres content when several vf card have vf card content of differing heights the grid vertically centres as seen at this is currently the probable solution but needs testing lest it introduce its own bugs css vf card align content start | 1 |
339,275 | 10,245,458,230 | IssuesEvent | 2019-08-20 12:54:38 | AbsaOSS/enceladus | https://api.github.com/repos/AbsaOSS/enceladus | closed | Allow Negation Rule of Boolean columns | Conformance feature priority: medium | ## Background
Currently, we only allow negation of Numeric columns. I would expect the negation to stretch to Booleans as well.
## Feature
Allow negation rule to be applied to boolean columns as well.
| 1.0 | Allow Negation Rule of Boolean columns - ## Background
Currently, we only allow negation of Numeric columns. I would expect the negation to stretch to Booleans as well.
## Feature
Allow negation rule to be applied to boolean columns as well.
| priority | allow negation rule of boolean columns background currently we only allow negation of numeric columns i would expect the negation to stretch to booleans as well feature allow negation rule to be applied to boolean columns as well | 1 |
439,086 | 12,677,087,796 | IssuesEvent | 2020-06-19 07:00:28 | visual-framework/vf-core | https://api.github.com/repos/visual-framework/vf-core | opened | FEATURE - Track network in vf-analytics-google | Priority: Medium Type: Feature | As of Feb 2020 Google Analytics no longer tracks the network name of visitors, this was previously very helpful in identifying internal traffic.
- [ ] It is possible to add a similar functionality with this plugin: https://ipmeta.io/
- [ ] Make it an optional feature (should be enabled by default for EMBL.org pages)
- [ ] Add some way to pass a `Custom Dimensions` index where it differs from a default
- Question: perhaps this should be a embl-analytics-ipmeta
https://ipmeta.io/instructions
```js
ga('require', 'ipMeta', {
serviceProvider: 'dimension1',
networkDomain: 'dimension2',
networkType: 'dimension3',
});
ga('ipMeta:loadNetworkFields');
ga('send', 'pageview');
```
There's other work planned for vf-anlatyics-google in #963 | 1.0 | FEATURE - Track network in vf-analytics-google - As of Feb 2020 Google Analytics no longer tracks the network name of visitors, this was previously very helpful in identifying internal traffic.
- [ ] It is possible to add a similar functionality with this plugin: https://ipmeta.io/
- [ ] Make it an optional feature (should be enabled by default for EMBL.org pages)
- [ ] Add some way to pass a `Custom Dimensions` index where it differs from a default
- Question: perhaps this should be a embl-analytics-ipmeta
https://ipmeta.io/instructions
```js
ga('require', 'ipMeta', {
serviceProvider: 'dimension1',
networkDomain: 'dimension2',
networkType: 'dimension3',
});
ga('ipMeta:loadNetworkFields');
ga('send', 'pageview');
```
There's other work planned for vf-anlatyics-google in #963 | priority | feature track network in vf analytics google as of feb google analytics no longer tracks the network name of visitors this was previously very helpful in identifying internal traffic it is possible to add a similar functionality with this plugin make it an optional feature should be enabled by default for embl org pages add some way to pass a custom dimensions index where it differs from a default question perhaps this should be a embl analytics ipmeta js ga require ipmeta serviceprovider networkdomain networktype ga ipmeta loadnetworkfields ga send pageview there s other work planned for vf anlatyics google in | 1 |
169,177 | 6,396,106,670 | IssuesEvent | 2017-08-04 14:46:58 | ruany/LiteBans | https://api.github.com/repos/ruany/LiteBans | opened | TODO: AdvancedBan import option | Enhancement Medium priority | https://www.spigotmc.org/resources/advancedban.8695/
This would be high priority if there was more demand, but I've only had a few requests so far.
Importing was originally intended for plugins that didn't have UUID support at the time, like MaxBans and BanManagerv4 (the original MaxBans still doesn't have actual UUID support despite claiming to have it).
AdvancedBan was released after LiteBans was first released, unlike all of the other plugins which can currently be imported from. | 1.0 | TODO: AdvancedBan import option - https://www.spigotmc.org/resources/advancedban.8695/
This would be high priority if there was more demand, but I've only had a few requests so far.
Importing was originally intended for plugins that didn't have UUID support at the time, like MaxBans and BanManagerv4 (the original MaxBans still doesn't have actual UUID support despite claiming to have it).
AdvancedBan was released after LiteBans was first released, unlike all of the other plugins which can currently be imported from. | priority | todo advancedban import option this would be high priority if there was more demand but i ve only had a few requests so far importing was originally intended for plugins that didn t have uuid support at the time like maxbans and the original maxbans still doesn t have actual uuid support despite claiming to have it advancedban was released after litebans was first released unlike all of the other plugins which can currently be imported from | 1 |
18,783 | 2,616,005,422 | IssuesEvent | 2015-03-02 00:49:51 | jasonhall/bwapi | https://api.github.com/repos/jasonhall/bwapi | closed | Remove ChaosDir | auto-migrated Priority-Medium Type-Other | ```
I think it definitely prevents people from joining the project team.
```
Original issue reported on code.google.com by `AHeinerm` on 3 Jun 2009 at 1:58 | 1.0 | Remove ChaosDir - ```
I think it definitely prevents people from joining the project team.
```
Original issue reported on code.google.com by `AHeinerm` on 3 Jun 2009 at 1:58 | priority | remove chaosdir i think it definitely prevents people from joining the project team original issue reported on code google com by aheinerm on jun at | 1 |
30,017 | 2,722,130,244 | IssuesEvent | 2015-04-14 00:14:19 | CruxFramework/crux-smart-faces | https://api.github.com/repos/CruxFramework/crux-smart-faces | closed | wrong XSD for topMenuDisposal | bug imported Milestone-M14-C4 Priority-Medium TargetVersion-5.3.0 | _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on February 11, 2015 11:08:32_
generated XSD for topMenuDisposal does not validate the expected content
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=595_ | 1.0 | wrong XSD for topMenuDisposal - _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on February 11, 2015 11:08:32_
generated XSD for topMenuDisposal does not validate the expected content
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=595_ | priority | wrong xsd for topmenudisposal from on february generated xsd for topmenudisposal does not validate the expected content original issue | 1 |
274,655 | 8,563,842,175 | IssuesEvent | 2018-11-09 15:10:53 | NamelessMC/Nameless | https://api.github.com/repos/NamelessMC/Nameless | closed | Rank sync with playerInfo API method | enhancement priority: medium v2 | Use the serverInfo API method to synchronise ranks between Minecraft server(s) and the website.
- Multiple servers might send conflicting data. How do we handle that?
- People might want to exclude specific ranks from being synced. Create a page in AdminCP where people can blacklist specific ranks or servers. This page should also have a short explanation of how rank sync works and how to enable it in the plugin.
- The interval at which servers send data is configurable. It might vary from several times per second to a few times a day. The website should be able to handle this | 1.0 | Rank sync with playerInfo API method - Use the serverInfo API method to synchronise ranks between Minecraft server(s) and the website.
- Multiple servers might send conflicting data. How do we handle that?
- People might want to exclude specific ranks from being synced. Create a page in AdminCP where people can blacklist specific ranks or servers. This page should also have a short explanation of how rank sync works and how to enable it in the plugin.
- The interval at which servers send data is configurable. It might vary from several times per second to a few times a day. The website should be able to handle this | priority | rank sync with playerinfo api method use the serverinfo api method to synchronise ranks between minecraft server s and the website multiple servers might send conflicting data how do we handle that people might want to exclude specific ranks from being synced create a page in admincp where people can blacklist specific ranks or servers this page should also have a short explanation of how rank sync works and how to enable it in the plugin the interval at which servers send data is configurable it might vary from several times per second to a few times a day the website should be able to handle this | 1 |
369,945 | 10,920,371,716 | IssuesEvent | 2019-11-21 21:06:47 | townhallproject/admin | https://api.github.com/repos/townhallproject/admin | opened | Update restriction to self-approval of events | Priority: Medium | Allow all THP staff to approve their own events if the event occurs in 48 hours or less. | 1.0 | Update restriction to self-approval of events - Allow all THP staff to approve their own events if the event occurs in 48 hours or less. | priority | update restriction to self approval of events allow all thp staff to approve their own events if the event occurs in hours or less | 1 |
820,027 | 30,758,394,875 | IssuesEvent | 2023-07-29 11:13:33 | code4romania/asistent-medical-comunitar | https://api.github.com/repos/code4romania/asistent-medical-comunitar | closed | [Beneficiari/Gospodării] change ''Beneficiaries'' field label | medium-priority | **Describe the bug**
Wrong field label.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Listă beneficiari section
2. Go to Gospodării tab
3. Click on the ''Adaugă household'' button
4. Notice the field label ''Beneficiaries''

**Expected behavior**
Change field label from ''Beneficiaries'' to ''Beneficiari''
| 1.0 | [Beneficiari/Gospodării] change ''Beneficiaries'' field label - **Describe the bug**
Wrong field label.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Listă beneficiari section
2. Go to Gospodării tab
3. Click on the ''Adaugă household'' button
4. Notice the field label ''Beneficiaries''

**Expected behavior**
Change field label from ''Beneficiaries'' to ''Beneficiari''
| priority | change beneficiaries field label describe the bug wrong field label to reproduce steps to reproduce the behavior go to listă beneficiari section go to gospodării tab click on the adaugă household button notice the field label beneficiaries expected behavior change field label from beneficiaries to beneficiari | 1 |
194,206 | 6,892,424,620 | IssuesEvent | 2017-11-22 21:00:18 | CCAFS/MARLO | https://api.github.com/repos/CCAFS/MARLO | closed | CIAT-MARLO - Get HR info from OCS to be used in CapDev module | Priority - Medium Specificity-CIAT Type - Enhancement | Develop a module to get information from OCS about CIAT's visiting researchers and students, so it can be used in the CapDev module that is being developed in CIAT-MARLO. | 1.0 | CIAT-MARLO - Get HR info from OCS to be used in CapDev module - Develop a module to get information from OCS about CIAT's visiting researchers and students, so it can be used in the CapDev module that is being developed in CIAT-MARLO. | priority | ciat marlo get hr info from ocs to be used in capdev module develop a module to get information from ocs about ciat s visiting researchers and students so it can be used in the capdev module that is being developed in ciat marlo | 1 |
412,444 | 12,042,628,972 | IssuesEvent | 2020-04-14 10:55:37 | ooni/probe | https://api.github.com/repos/ooni/probe | closed | NDT: intermittent broken stats | bug discuss priority/medium | As mentioned in #877 the migration from ndt4 to ndt5 is impacting our NDT. We've mitigated the most sever issues but we need still to deal with https://github.com/m-lab/ndt-server/issues/161 and https://github.com/m-lab/ndt-server/issues/163. To this end, the MVP is to disable printing advanced stats in the apps. When we'll have properly dealt with the above mentioned issues of ndt-server, we'll then re-enable the stats.
I will lead this process as discussed in today's dev meeting. Paging also all the other devs that are likely to collaborate on fixing this. | 1.0 | NDT: intermittent broken stats - As mentioned in #877 the migration from ndt4 to ndt5 is impacting our NDT. We've mitigated the most sever issues but we need still to deal with https://github.com/m-lab/ndt-server/issues/161 and https://github.com/m-lab/ndt-server/issues/163. To this end, the MVP is to disable printing advanced stats in the apps. When we'll have properly dealt with the above mentioned issues of ndt-server, we'll then re-enable the stats.
I will lead this process as discussed in today's dev meeting. Paging also all the other devs that are likely to collaborate on fixing this. | priority | ndt intermittent broken stats as mentioned in the migration from to is impacting our ndt we ve mitigated the most sever issues but we need still to deal with and to this end the mvp is to disable printing advanced stats in the apps when we ll have properly dealt with the above mentioned issues of ndt server we ll then re enable the stats i will lead this process as discussed in today s dev meeting paging also all the other devs that are likely to collaborate on fixing this | 1 |
567,540 | 16,870,269,005 | IssuesEvent | 2021-06-22 02:56:05 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.4 staging-2020] A small delay when shot with bow | Priority: Medium Regression Squad: Otter Status: Fixed Type: Bug | Step to reproduce:
- take any bow and arrows:
- hold LMB and wait 5 sec:

- release LMB, you have a small delay to shot.
- pretty hard to hit animal now. | 1.0 | [0.9.4 staging-2020] A small delay when shot with bow - Step to reproduce:
- take any bow and arrows:
- hold LMB and wait 5 sec:

- release LMB, you have a small delay to shot.
- pretty hard to hit animal now. | priority | a small delay when shot with bow step to reproduce take any bow and arrows hold lmb and wait sec release lmb you have a small delay to shot pretty hard to hit animal now | 1 |
55,916 | 3,075,268,503 | IssuesEvent | 2015-08-20 12:47:32 | bireme/proethos | https://api.github.com/repos/bireme/proethos | closed | Al subir una propuesta, cuando uno sube un adjunto se borra lo que había ingresado en la página | bug priority 2 (medium) severity 3 (normal/minor impact) | Debe poder haber la opción de "grabar" las Referencias bibliográficas, contacto, etc. antes de subir el adjunto, o alternativamente se debe subir los adjuntos ("Presentación de archivos") en una página separada. | 1.0 | Al subir una propuesta, cuando uno sube un adjunto se borra lo que había ingresado en la página - Debe poder haber la opción de "grabar" las Referencias bibliográficas, contacto, etc. antes de subir el adjunto, o alternativamente se debe subir los adjuntos ("Presentación de archivos") en una página separada. | priority | al subir una propuesta cuando uno sube un adjunto se borra lo que había ingresado en la página debe poder haber la opción de grabar las referencias bibliográficas contacto etc antes de subir el adjunto o alternativamente se debe subir los adjuntos presentación de archivos en una página separada | 1 |
414,336 | 12,102,356,959 | IssuesEvent | 2020-04-20 16:33:58 | ESGF/esgf-compute-wps | https://api.github.com/repos/ESGF/esgf-compute-wps | closed | Add provenance handling | area/backend area/frontend kind/feature priority/high size/medium | We just need to store two things.
- [ ] Version of the containers
- [ ] Instances like EDASK and Ophidia we'll need to retrieve the version somehow
- [ ] Original URL | 1.0 | Add provenance handling - We just need to store two things.
- [ ] Version of the containers
- [ ] Instances like EDASK and Ophidia we'll need to retrieve the version somehow
- [ ] Original URL | priority | add provenance handling we just need to store two things version of the containers instances like edask and ophidia we ll need to retrieve the version somehow original url | 1 |
289,698 | 8,875,287,287 | IssuesEvent | 2019-01-12 01:59:25 | masari-project/masari | https://api.github.com/repos/masari-project/masari | closed | Remaining core tests for SECOR before mainnet | Priority: Medium Status: Completed Type: Task | Tests to do for mainnet:
* Orphaned uncle transaction outs don't get included (void) in the transaction db and aren't spendable - add tests to ensure that outs are void and aren't spendable somehow
* Current re-org tests need assertions on expected behaviours (i.e. weight is as expected)
Tests already done for testnet:
* uncles mined properly and reorg well in all different cases
* a parent can't be an uncle
* an uncle must be at the same height as the parent
* an uncle must be the same version as parent
* an uncle must have a timestamp within range of the parent (T / 2)
* an uncle must have common ancestry (uncleparent == grandparent or uncleparent's parent == grandparent's parent)
* a nephew must mine the correct out tx to uncle
* a nephew must mine the corret out tx amount to uncle
* a nephew cannot overflow amounts in miner tx | 1.0 | Remaining core tests for SECOR before mainnet - Tests to do for mainnet:
* Orphaned uncle transaction outs don't get included (void) in the transaction db and aren't spendable - add tests to ensure that outs are void and aren't spendable somehow
* Current re-org tests need assertions on expected behaviours (i.e. weight is as expected)
Tests already done for testnet:
* uncles mined properly and reorg well in all different cases
* a parent can't be an uncle
* an uncle must be at the same height as the parent
* an uncle must be the same version as parent
* an uncle must have a timestamp within range of the parent (T / 2)
* an uncle must have common ancestry (uncleparent == grandparent or uncleparent's parent == grandparent's parent)
* a nephew must mine the correct out tx to uncle
* a nephew must mine the corret out tx amount to uncle
* a nephew cannot overflow amounts in miner tx | priority | remaining core tests for secor before mainnet tests to do for mainnet orphaned uncle transaction outs don t get included void in the transaction db and aren t spendable add tests to ensure that outs are void and aren t spendable somehow current re org tests need assertions on expected behaviours i e weight is as expected tests already done for testnet uncles mined properly and reorg well in all different cases a parent can t be an uncle an uncle must be at the same height as the parent an uncle must be the same version as parent an uncle must have a timestamp within range of the parent t an uncle must have common ancestry uncleparent grandparent or uncleparent s parent grandparent s parent a nephew must mine the correct out tx to uncle a nephew must mine the corret out tx amount to uncle a nephew cannot overflow amounts in miner tx | 1 |
813,001 | 30,441,814,862 | IssuesEvent | 2023-07-15 06:28:22 | Memmy-App/memmy | https://api.github.com/repos/Memmy-App/memmy | closed | Memmy don’t ask permission to save image | bug medium priority image viewer | **Describe the bug**
Memmy don’t ask for photo permission when trying to save an image with new image viewer when pressing floppy disk icon. The haptic feedback is provided tho as if something happened.
**To Reproduce**
Steps to reproduce the behavior:
1. Ensure memmy has no photo permission
2. open an image in memmy
3. Try to save by pressing floppy disk icon
**Expected behavior**
Memmy asks for permission otherwise it won’t be able to save the photo.
**Smartphone (please complete the following information):**
- Device: iPhone 11 Pro
- OS: iOS16.5.1
- Version : 0.3
| 1.0 | Memmy don’t ask permission to save image - **Describe the bug**
Memmy don’t ask for photo permission when trying to save an image with new image viewer when pressing floppy disk icon. The haptic feedback is provided tho as if something happened.
**To Reproduce**
Steps to reproduce the behavior:
1. Ensure memmy has no photo permission
2. open an image in memmy
3. Try to save by pressing floppy disk icon
**Expected behavior**
Memmy asks for permission otherwise it won’t be able to save the photo.
**Smartphone (please complete the following information):**
- Device: iPhone 11 Pro
- OS: iOS16.5.1
- Version : 0.3
| priority | memmy don’t ask permission to save image describe the bug memmy don’t ask for photo permission when trying to save an image with new image viewer when pressing floppy disk icon the haptic feedback is provided tho as if something happened to reproduce steps to reproduce the behavior ensure memmy has no photo permission open an image in memmy try to save by pressing floppy disk icon expected behavior memmy asks for permission otherwise it won’t be able to save the photo smartphone please complete the following information device iphone pro os version | 1 |
757,201 | 26,499,961,839 | IssuesEvent | 2023-01-18 09:32:10 | polyflix/frontend | https://api.github.com/repos/polyflix/frontend | closed | [ENHANCEMENT]: When watching a video, the creator of the video must be clickable (to go to his profile page) | priority: medium action: triage type: fix type: style | ### Description
When watching a video, the creator of the video must be clickable (to go to his profile page)
fast example: https://polyflix.dopolytech.fr/videos/trust-by-verify-nicolas-muller-nicolas-tournier-polycloud-2022-4rx7M

### Additional Information
_No response_ | 1.0 | [ENHANCEMENT]: When watching a video, the creator of the video must be clickable (to go to his profile page) - ### Description
When watching a video, the creator of the video must be clickable (to go to his profile page)
fast example: https://polyflix.dopolytech.fr/videos/trust-by-verify-nicolas-muller-nicolas-tournier-polycloud-2022-4rx7M

### Additional Information
_No response_ | priority | when watching a video the creator of the video must be clickable to go to his profile page description when watching a video the creator of the video must be clickable to go to his profile page fast example additional information no response | 1 |
35,897 | 2,793,819,830 | IssuesEvent | 2015-05-11 13:37:16 | elecoest/allevents-3-2 | https://api.github.com/repos/elecoest/allevents-3-2 | closed | Frontend\Liste_Events\Visualisation\Bouton Traitement\Toggle-Dropdown | auto-migrated Priority-Medium Type-Enhancement | ```
joomla 3.4 AllEvents version 3.2.13
Afin de simplifier la visualisation d'un événement dans la liste des
événements et afin de gagner de la place en hauteur pour chacun des
événements,ne pourrait-on pas:
-Supprimer le logo indiquant qu'il y a ou non une photo du lieu (info
subsidiaire)
-Remplacer le logo flèche grise en bas à droite par un bouton "visualiser"
-Ne pas faire apparaître le bouton traitement/Toggle-Dropdown si
l'utilisateur est un simple enregistré (cas général)
-Remonter le tout vers le haut
voir ma suggestion en pièce jointe
```
Original issue reported on code.google.com by `jeanluc...@gmail.com` on 9 Mar 2015 at 9:18
Attachments:
* [liste_événements1.jpg](https://storage.googleapis.com/google-code-attachments/allevents-3-2/issue-476/comment-0/liste_événements1.jpg)
| 1.0 | Frontend\Liste_Events\Visualisation\Bouton Traitement\Toggle-Dropdown - ```
joomla 3.4 AllEvents version 3.2.13
Afin de simplifier la visualisation d'un événement dans la liste des
événements et afin de gagner de la place en hauteur pour chacun des
événements,ne pourrait-on pas:
-Supprimer le logo indiquant qu'il y a ou non une photo du lieu (info
subsidiaire)
-Remplacer le logo flèche grise en bas à droite par un bouton "visualiser"
-Ne pas faire apparaître le bouton traitement/Toggle-Dropdown si
l'utilisateur est un simple enregistré (cas général)
-Remonter le tout vers le haut
voir ma suggestion en pièce jointe
```
Original issue reported on code.google.com by `jeanluc...@gmail.com` on 9 Mar 2015 at 9:18
Attachments:
* [liste_événements1.jpg](https://storage.googleapis.com/google-code-attachments/allevents-3-2/issue-476/comment-0/liste_événements1.jpg)
| priority | frontend liste events visualisation bouton traitement toggle dropdown joomla allevents version afin de simplifier la visualisation d un événement dans la liste des événements et afin de gagner de la place en hauteur pour chacun des événements ne pourrait on pas supprimer le logo indiquant qu il y a ou non une photo du lieu info subsidiaire remplacer le logo flèche grise en bas à droite par un bouton visualiser ne pas faire apparaître le bouton traitement toggle dropdown si l utilisateur est un simple enregistré cas général remonter le tout vers le haut voir ma suggestion en pièce jointe original issue reported on code google com by jeanluc gmail com on mar at attachments | 1 |
457,817 | 13,162,866,076 | IssuesEvent | 2020-08-10 22:39:16 | forestgeo/allodb | https://api.github.com/repos/forestgeo/allodb | closed | Determine how to interact with BIOMASS package | priority high priority medium | Determine how our package interacts with/ sources the [BIOMASS package](https://cran.r-project.org/web/packages/BIOMASS/index.html).
Requirements/ Goals:
1- Our code should give the same answer as BIOMASS, and we need assurance that this is maintained as both packages change
2- fgeo users should experience a simple, smoothly integrated, unified process for computing biomass. They should not have to separately load and interact with BIOMASS or give separate treatment to sites using allodb equations tables vs BIOMASS package.
Two main options:
1- Put fgeo wrapper around BIOMASS
2- Use pieces of BIOMASS code in our package
| 2.0 | Determine how to interact with BIOMASS package - Determine how our package interacts with/ sources the [BIOMASS package](https://cran.r-project.org/web/packages/BIOMASS/index.html).
Requirements/ Goals:
1- Our code should give the same answer as BIOMASS, and we need assurance that this is maintained as both packages change
2- fgeo users should experience a simple, smoothly integrated, unified process for computing biomass. They should not have to separately load and interact with BIOMASS or give separate treatment to sites using allodb equations tables vs BIOMASS package.
Two main options:
1- Put fgeo wrapper around BIOMASS
2- Use pieces of BIOMASS code in our package
| priority | determine how to interact with biomass package determine how our package interacts with sources the requirements goals our code should give the same answer as biomass and we need assurance that this is maintained as both packages change fgeo users should experience a simple smoothly integrated unified process for computing biomass they should not have to separately load and interact with biomass or give separate treatment to sites using allodb equations tables vs biomass package two main options put fgeo wrapper around biomass use pieces of biomass code in our package | 1 |
517,482 | 15,014,678,344 | IssuesEvent | 2021-02-01 07:03:57 | teamforus/general | https://api.github.com/repos/teamforus/general | closed | Rename "acties" to "aanbod" | Approval: Granted Impact: Moderate Priority: Must have Status: Planned Urgency: Medium - Next 3 sprints project-100 | @maxvisser commented on [Thu Sep 10 2020](https://github.com/teamforus/forus/issues/1758)
## Main asssignee: @
## Context/goal:
<img width="675" alt="Screenshot 2020-12-30 at 10 45 46" src="https://user-images.githubusercontent.com/30194799/103343397-42b58000-4a8c-11eb-9af6-f8606f8c8984.png">
---
@maxvisser commented on [Mon Sep 21 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-696098909)
Fixed
---
@maxvisser commented on [Wed Nov 25 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-733783579)
@SjoerdHilhorst I see you changed the page subsidies to 'Aanbod' and the navbar item to 'Aanbod'; seems that this was not in the text fix issue as a planned task. Could you revert these changes?
---
@SjoerdHilhorst commented on [Mon Nov 30 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-735639951)
This mentions changing it to "aanbod":
https://docs.google.com/spreadsheets/d/1B1rSQjzejkIMThG2bAIW27Xg6Ol2GLY-hLts7b5vcsI/edit#gid=433994186&range=C26
@jamalv what do you think
---
@maxvisser commented on [Mon Nov 30 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-735670355)
We already reversed the change in release 0.16.3
---
@maxvisser commented on [Mon Nov 30 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-735670565)
Did you implement any other textual fixes from the spreadsheet?
---
@maartenfv commented on [Tue Dec 22 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-749616268)
@jamalv Can we change it to aanbod? As requested by Groningen?
---
@maxvisser commented on [Wed Dec 23 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-750081915)
Please open a new CR instead of opening old forus tasks.
| 1.0 | Rename "acties" to "aanbod" - @maxvisser commented on [Thu Sep 10 2020](https://github.com/teamforus/forus/issues/1758)
## Main asssignee: @
## Context/goal:
<img width="675" alt="Screenshot 2020-12-30 at 10 45 46" src="https://user-images.githubusercontent.com/30194799/103343397-42b58000-4a8c-11eb-9af6-f8606f8c8984.png">
---
@maxvisser commented on [Mon Sep 21 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-696098909)
Fixed
---
@maxvisser commented on [Wed Nov 25 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-733783579)
@SjoerdHilhorst I see you changed the page subsidies to 'Aanbod' and the navbar item to 'Aanbod'; seems that this was not in the text fix issue as a planned task. Could you revert these changes?
---
@SjoerdHilhorst commented on [Mon Nov 30 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-735639951)
This mentions changing it to "aanbod":
https://docs.google.com/spreadsheets/d/1B1rSQjzejkIMThG2bAIW27Xg6Ol2GLY-hLts7b5vcsI/edit#gid=433994186&range=C26
@jamalv what do you think
---
@maxvisser commented on [Mon Nov 30 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-735670355)
We already reversed the change in release 0.16.3
---
@maxvisser commented on [Mon Nov 30 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-735670565)
Did you implement any other textual fixes from the spreadsheet?
---
@maartenfv commented on [Tue Dec 22 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-749616268)
@jamalv Can we change it to aanbod? As requested by Groningen?
---
@maxvisser commented on [Wed Dec 23 2020](https://github.com/teamforus/forus/issues/1758#issuecomment-750081915)
Please open a new CR instead of opening old forus tasks.
| priority | rename acties to aanbod maxvisser commented on main asssignee context goal img width alt screenshot at src maxvisser commented on fixed maxvisser commented on sjoerdhilhorst i see you changed the page subsidies to aanbod and the navbar item to aanbod seems that this was not in the text fix issue as a planned task could you revert these changes sjoerdhilhorst commented on this mentions changing it to aanbod jamalv what do you think maxvisser commented on we already reversed the change in release maxvisser commented on did you implement any other textual fixes from the spreadsheet maartenfv commented on jamalv can we change it to aanbod as requested by groningen maxvisser commented on please open a new cr instead of opening old forus tasks | 1 |
68,516 | 3,289,011,395 | IssuesEvent | 2015-10-29 17:14:40 | softdevteam/krun | https://api.github.com/repos/softdevteam/krun | closed | Tickless operation on linux | enhancement medium priority (a clear improvement but not a blocker for publication) | As raised by ltratt, we may need to use tickless mode on linux:
http://lwn.net/Articles/549580/
Our debian 8 systems currently run idle tickless mode:
```
vext01@bencher3:~$ cat /boot/config-`uname -r` | grep HZ
CONFIG_NO_HZ_COMMON=y
# CONFIG_HZ_PERIODIC is not set
CONFIG_NO_HZ_IDLE=y
# CONFIG_NO_HZ_FULL is not set <--------------------- this would need to be on!
# CONFIG_NO_HZ is not set
CONFIG_RCU_FAST_NO_HZ=y
# CONFIG_HZ_100 is not set
CONFIG_HZ_250=y
# CONFIG_HZ_300 is not set
# CONFIG_HZ_1000 is not set
CONFIG_HZ=250
CONFIG_MACHZ_WDT=m
```
To fix this, we would need to build a custom kernel. Another impractical step unfortunately.
Opening bug for discussion. @ltratt @snim2 @cfbolz | 1.0 | Tickless operation on linux - As raised by ltratt, we may need to use tickless mode on linux:
http://lwn.net/Articles/549580/
Our debian 8 systems currently run idle tickless mode:
```
vext01@bencher3:~$ cat /boot/config-`uname -r` | grep HZ
CONFIG_NO_HZ_COMMON=y
# CONFIG_HZ_PERIODIC is not set
CONFIG_NO_HZ_IDLE=y
# CONFIG_NO_HZ_FULL is not set <--------------------- this would need to be on!
# CONFIG_NO_HZ is not set
CONFIG_RCU_FAST_NO_HZ=y
# CONFIG_HZ_100 is not set
CONFIG_HZ_250=y
# CONFIG_HZ_300 is not set
# CONFIG_HZ_1000 is not set
CONFIG_HZ=250
CONFIG_MACHZ_WDT=m
```
To fix this, we would need to build a custom kernel. Another impractical step unfortunately.
Opening bug for discussion. @ltratt @snim2 @cfbolz | priority | tickless operation on linux as raised by ltratt we may need to use tickless mode on linux our debian systems currently run idle tickless mode cat boot config uname r grep hz config no hz common y config hz periodic is not set config no hz idle y config no hz full is not set this would need to be on config no hz is not set config rcu fast no hz y config hz is not set config hz y config hz is not set config hz is not set config hz config machz wdt m to fix this we would need to build a custom kernel another impractical step unfortunately opening bug for discussion ltratt cfbolz | 1 |
363,308 | 10,741,053,307 | IssuesEvent | 2019-10-29 19:25:53 | eriq-augustine/test-issue-copy | https://api.github.com/repos/eriq-augustine/test-issue-copy | opened | [CLOSED] Java Interface | Difficulty - Medium Interfaces Priority - Normal Type - Investigation | <a href="https://github.com/eriq-augustine"><img src="https://avatars0.githubusercontent.com/u/337857?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [eriq-augustine](https://github.com/eriq-augustine)**
_Monday Jan 30, 2017 at 07:24 GMT_
_Originally opened as https://github.com/eriq-augustine/psl/issues/44_
----
With the addition of the string parsing of rules, I don't think there is a hard requirement for the Groovy interface anymore.
See what it would take to have a Java interface.
| 1.0 | [CLOSED] Java Interface - <a href="https://github.com/eriq-augustine"><img src="https://avatars0.githubusercontent.com/u/337857?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [eriq-augustine](https://github.com/eriq-augustine)**
_Monday Jan 30, 2017 at 07:24 GMT_
_Originally opened as https://github.com/eriq-augustine/psl/issues/44_
----
With the addition of the string parsing of rules, I don't think there is a hard requirement for the Groovy interface anymore.
See what it would take to have a Java interface.
| priority | java interface issue by monday jan at gmt originally opened as with the addition of the string parsing of rules i don t think there is a hard requirement for the groovy interface anymore see what it would take to have a java interface | 1 |
167,096 | 6,332,096,212 | IssuesEvent | 2017-07-26 11:46:13 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio] The featured articles does not appear in the Home page when you first create the site in editorial bp | bug Priority: Medium | Create a new site using the website_editorial bp.
Notice that the featured articles section of the Home page is empty. To see the articles, you need to refresh.

| 1.0 | [studio] The featured articles does not appear in the Home page when you first create the site in editorial bp - Create a new site using the website_editorial bp.
Notice that the featured articles section of the Home page is empty. To see the articles, you need to refresh.

| priority | the featured articles does not appear in the home page when you first create the site in editorial bp create a new site using the website editorial bp notice that the featured articles section of the home page is empty to see the articles you need to refresh | 1 |
58,715 | 3,090,670,824 | IssuesEvent | 2015-08-26 08:25:02 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Хэширование файлов "спотыкается" о "файл", который является символической ссылкой на другой файл, и стоит на месте (никогда не заканчивается) | bug imported Priority-Medium | _From [ilyakane...@yandex.ru](https://code.google.com/u/112743903078435663629/) on May 21, 2015 04:53:49_
Win7 x64 SP1 Pro, NTFS
FlylinkDC++ r502 -x64 build 18654
По пути "D:\Музыка\Транс\Menno De Jong\Judge Jules - Weekend Warmup (Menno De Jong & Lange Guests) (24.02.2012).mp3" у меня лежит символическая ссылка на "D:\Музыка\Транс\Menno De Jong\Judge Jules - Weekend Warmup (Menno De Jong & Lange Guests) (24.02.2012).mp3".
Ссылка создана в FAR2 путем нажатия Alt-F6 и далее выбрано "Тип ссылки" = "символическая ссылка (файл)".
В окне Flylink'а "Хеширую файлы..." (вызывается по Ctrl-H) видно, что скорость по нулям, осталось еще 2000+ файлов, и вверху отображается "D:\Музыка\Транс\Menno De Jong\Judge Jules - Weekend Warmup (Menno De Jong & Lange Guests) (24.02.2012).mp3". И дальше ничего не происходит.
Скриншот прикрепить не могу, так как здесь внизу написано: "Issue attachment storage quota exceeded."
Могу выполнить любую дополнительную диагностику, только скажите, какую :-).
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1573_ | 1.0 | Хэширование файлов "спотыкается" о "файл", который является символической ссылкой на другой файл, и стоит на месте (никогда не заканчивается) - _From [ilyakane...@yandex.ru](https://code.google.com/u/112743903078435663629/) on May 21, 2015 04:53:49_
Win7 x64 SP1 Pro, NTFS
FlylinkDC++ r502 -x64 build 18654
По пути "D:\Музыка\Транс\Menno De Jong\Judge Jules - Weekend Warmup (Menno De Jong & Lange Guests) (24.02.2012).mp3" у меня лежит символическая ссылка на "D:\Музыка\Транс\Menno De Jong\Judge Jules - Weekend Warmup (Menno De Jong & Lange Guests) (24.02.2012).mp3".
Ссылка создана в FAR2 путем нажатия Alt-F6 и далее выбрано "Тип ссылки" = "символическая ссылка (файл)".
В окне Flylink'а "Хеширую файлы..." (вызывается по Ctrl-H) видно, что скорость по нулям, осталось еще 2000+ файлов, и вверху отображается "D:\Музыка\Транс\Menno De Jong\Judge Jules - Weekend Warmup (Menno De Jong & Lange Guests) (24.02.2012).mp3". И дальше ничего не происходит.
Скриншот прикрепить не могу, так как здесь внизу написано: "Issue attachment storage quota exceeded."
Могу выполнить любую дополнительную диагностику, только скажите, какую :-).
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1573_ | priority | хэширование файлов спотыкается о файл который является символической ссылкой на другой файл и стоит на месте никогда не заканчивается from on may pro ntfs flylinkdc build по пути d музыка транс menno de jong judge jules weekend warmup menno de jong lange guests у меня лежит символическая ссылка на d музыка транс menno de jong judge jules weekend warmup menno de jong lange guests ссылка создана в путем нажатия alt и далее выбрано тип ссылки символическая ссылка файл в окне flylink а хеширую файлы вызывается по ctrl h видно что скорость по нулям осталось еще файлов и вверху отображается d музыка транс menno de jong judge jules weekend warmup menno de jong lange guests и дальше ничего не происходит скриншот прикрепить не могу так как здесь внизу написано issue attachment storage quota exceeded могу выполнить любую дополнительную диагностику только скажите какую original issue | 1 |
340,542 | 10,273,165,842 | IssuesEvent | 2019-08-23 18:29:54 | fgpv-vpgf/fgpv-vpgf | https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf | closed | Structure legend symbology stack images | inactive: fixed priority: medium problem: bug type: corrective | When we replace icons (unique value renderer) for a layer inside the structure legend, the unique value visibility selector doesn't work anymore.

Config file snippet
```
"symbologyStack": [
{
"image": "https://geoappext.nrcan.gc.ca/fgpv/images/saving.png",
"text": "A"
},
{
"image": "https://geoappext.nrcan.gc.ca/fgpv/images/powering.png",
"text": "B"
},
{
"image": "https://geoappext.nrcan.gc.ca/fgpv/images/reusable.png",
"text": "C"
},
{
"image": "https://geoappext.nrcan.gc.ca/fgpv/images/world.png",
"text": "D"
}
],
"symbologyRenderStyle": "icons",
``` | 1.0 | Structure legend symbology stack images - When we replace icons (unique value renderer) for a layer inside the structure legend, the unique value visibility selector doesn't work anymore.

Config file snippet
```
"symbologyStack": [
{
"image": "https://geoappext.nrcan.gc.ca/fgpv/images/saving.png",
"text": "A"
},
{
"image": "https://geoappext.nrcan.gc.ca/fgpv/images/powering.png",
"text": "B"
},
{
"image": "https://geoappext.nrcan.gc.ca/fgpv/images/reusable.png",
"text": "C"
},
{
"image": "https://geoappext.nrcan.gc.ca/fgpv/images/world.png",
"text": "D"
}
],
"symbologyRenderStyle": "icons",
``` | priority | structure legend symbology stack images when we replace icons unique value renderer for a layer inside the structure legend the unique value visibility selector doesn t work anymore config file snippet symbologystack image text a image text b image text c image text d symbologyrenderstyle icons | 1 |
523,254 | 15,176,473,618 | IssuesEvent | 2021-02-14 05:35:04 | reberhardt7/cplayground | https://api.github.com/repos/reberhardt7/cplayground | closed | Show zombie processes in debugger | difficulty:medium good first issue priority:medium type:enhancement work:backend work:kernel-module | The debugger has a pane that lists running processes:

Say the user forks off a child process; the child process exits, but the parent process is still running and has not yet called `waitpid()` on the child. The child process is then a *zombie process*; it still exists, even though it has terminated. As such, we should list that zombie process in the process listing. (This is important because cplayground is used in an educational setting where we help students new to multiprocessing learn how to reap processes and avoid resource leaks.)
The data for the process listing comes from two places: `/proc/cplayground` (a file generated by our kernel module) and gdb. I am pretty sure `/proc/cplayground` does not currently include zombie processes, so we'll have to change that. (I believe [this](https://github.com/reberhardt7/cplayground/blob/aa07b66af975df940979d77f1cd7ee9fb5a065c2/src/server/kernel-mod/cplayground.c#L184) skips over zombies so that we can get the process namespace. If true, we need some other way to get the process's PID namespace. But this code is rusty in my head.) Then, we'll need to change the [code that integrates the procfile data with gdb data](https://github.com/reberhardt7/cplayground/blob/3bba190888f1f06f48de5af4b0583dfe1297294b/src/server/debugging.ts#L508). Right now, it omits processes that don't show up in gdb, but we wouldn't want to do that because we want to keep the zombie processes (even though they're no longer gdb inferiors). | 1.0 | Show zombie processes in debugger - The debugger has a pane that lists running processes:

Say the user forks off a child process; the child process exits, but the parent process is still running and has not yet called `waitpid()` on the child. The child process is then a *zombie process*; it still exists, even though it has terminated. As such, we should list that zombie process in the process listing. (This is important because cplayground is used in an educational setting where we help students new to multiprocessing learn how to reap processes and avoid resource leaks.)
The data for the process listing comes from two places: `/proc/cplayground` (a file generated by our kernel module) and gdb. I am pretty sure `/proc/cplayground` does not currently include zombie processes, so we'll have to change that. (I believe [this](https://github.com/reberhardt7/cplayground/blob/aa07b66af975df940979d77f1cd7ee9fb5a065c2/src/server/kernel-mod/cplayground.c#L184) skips over zombies so that we can get the process namespace. If true, we need some other way to get the process's PID namespace. But this code is rusty in my head.) Then, we'll need to change the [code that integrates the procfile data with gdb data](https://github.com/reberhardt7/cplayground/blob/3bba190888f1f06f48de5af4b0583dfe1297294b/src/server/debugging.ts#L508). Right now, it omits processes that don't show up in gdb, but we wouldn't want to do that because we want to keep the zombie processes (even though they're no longer gdb inferiors). | priority | show zombie processes in debugger the debugger has a pane that lists running processes say the user forks off a child process the child process exits but the parent process is still running and has not yet called waitpid on the child the child process is then a zombie process it still exists even though it has terminated as such we should list that zombie process in the process listing this is important because cplayground is used in an educational setting where we help students new to multiprocessing learn how to reap processes and avoid resource leaks the data for the process listing comes from two places proc cplayground a file generated by our kernel module and gdb i am pretty sure proc cplayground does not currently include zombie processes so we ll have to change that i believe skips over zombies so that we can get the process namespace if true we need some other way to get the process s pid namespace but this code is rusty in my head then we ll need to change the right now it omits processes that don t show up in gdb but we wouldn t want to do that because we want to keep the zombie processes even though they re no longer gdb inferiors | 1 |
145,187 | 5,560,078,152 | IssuesEvent | 2017-03-24 18:30:19 | vanilla-framework/vanilla-framework | https://api.github.com/repos/vanilla-framework/vanilla-framework | closed | Links wrong in the Readme | Priority: Medium Status: Review Type: Maintenance | Some broken links in the Readme
- Mailing list
- Homepage link by hotlink | 1.0 | Links wrong in the Readme - Some broken links in the Readme
- Mailing list
- Homepage link by hotlink | priority | links wrong in the readme some broken links in the readme mailing list homepage link by hotlink | 1 |
693,126 | 23,764,144,047 | IssuesEvent | 2022-09-01 11:23:31 | MegEngine/MegEngine | https://api.github.com/repos/MegEngine/MegEngine | closed | 默认使用 uuid 作为随机种子生成器 | type: enhancement status: in progress priority: medium | 使用 uuid 作为随机种子生成器
## 背景
在多机训练的过程中,对于模型并行非共享部分的参数,我希望它们默认为不同的初始化值。
## 需求描述
具体地,我希望以下代码中设置 random_seed 的部分,使用 uuid.uuid4().int,而不是原本的实现。
https://github.com/MegEngine/MegEngine/blob/f91881ffdc051ab49314b1bd12c4a07a862dc9c6/python_module/megengine/random/rng.py#L20
在原本的实现中,相邻编号的两个 worker 有极高的可能相差 1s 启动,此时它们的 random_seed 就是相同的,所初始化的参数也是相同的。我认为这在许多情况下会导致恶性 bug 且难以 debug | 1.0 | 默认使用 uuid 作为随机种子生成器 - 使用 uuid 作为随机种子生成器
## 背景
在多机训练的过程中,对于模型并行非共享部分的参数,我希望它们默认为不同的初始化值。
## 需求描述
具体地,我希望以下代码中设置 random_seed 的部分,使用 uuid.uuid4().int,而不是原本的实现。
https://github.com/MegEngine/MegEngine/blob/f91881ffdc051ab49314b1bd12c4a07a862dc9c6/python_module/megengine/random/rng.py#L20
在原本的实现中,相邻编号的两个 worker 有极高的可能相差 1s 启动,此时它们的 random_seed 就是相同的,所初始化的参数也是相同的。我认为这在许多情况下会导致恶性 bug 且难以 debug | priority | 默认使用 uuid 作为随机种子生成器 使用 uuid 作为随机种子生成器 背景 在多机训练的过程中,对于模型并行非共享部分的参数,我希望它们默认为不同的初始化值。 需求描述 具体地,我希望以下代码中设置 random seed 的部分,使用 uuid int,而不是原本的实现。 在原本的实现中,相邻编号的两个 worker 有极高的可能相差 启动,此时它们的 random seed 就是相同的,所初始化的参数也是相同的。我认为这在许多情况下会导致恶性 bug 且难以 debug | 1 |
405,534 | 11,874,357,934 | IssuesEvent | 2020-03-26 18:51:31 | aol/moloch | https://api.github.com/repos/aol/moloch | opened | Search expression changes removed when selecting new page size | bug medium priority viewer | Selecting a new page size from the pagination control removes changes in search expression if the search wasn't issued. This can be fixed in one of two ways:
1. The page size should be updated, then if the search expression has changed, it should be applied and a new query should be run (this mimics the behavior of page and time changes).
2. The page size should be updated and the new search expression should be preserved. But a new query should not run until the user issues a request to do so. | 1.0 | Search expression changes removed when selecting new page size - Selecting a new page size from the pagination control removes changes in search expression if the search wasn't issued. This can be fixed in one of two ways:
1. The page size should be updated, then if the search expression has changed, it should be applied and a new query should be run (this mimics the behavior of page and time changes).
2. The page size should be updated and the new search expression should be preserved. But a new query should not run until the user issues a request to do so. | priority | search expression changes removed when selecting new page size selecting a new page size from the pagination control removes changes in search expression if the search wasn t issued this can be fixed in one of two ways the page size should be updated then if the search expression has changed it should be applied and a new query should be run this mimics the behavior of page and time changes the page size should be updated and the new search expression should be preserved but a new query should not run until the user issues a request to do so | 1 |
265,105 | 8,337,317,913 | IssuesEvent | 2018-09-28 10:40:46 | CS2103-AY1819S1-W14-1/main | https://api.github.com/repos/CS2103-AY1819S1-W14-1/main | opened | As a medical clinic receptionist, I want to update doctor of patient’s medical details. | priority.Medium type.Story | So that doctor will know how to better treat patient. | 1.0 | As a medical clinic receptionist, I want to update doctor of patient’s medical details. - So that doctor will know how to better treat patient. | priority | as a medical clinic receptionist i want to update doctor of patient’s medical details so that doctor will know how to better treat patient | 1 |
234,823 | 7,726,656,570 | IssuesEvent | 2018-05-24 22:03:20 | photonstorm/phaser | https://api.github.com/repos/photonstorm/phaser | closed | Camera shake broken for Tilemaps in WebGL since 3.7.1 | 🐛 Bug 🤨 Difficulty: Medium 🥇 Priority: High | The following commit broke the camera shake effect when using tilemaps on WebGL
https://github.com/photonstorm/phaser/commit/1ce2311465b5e82bc0f3585b8e6dbf5b77401c3c
Can be reproduced by adding:
```
this.cameras.main.shake(2000)
```
to any of the labs using tilemaps.
This one for example:
http://labs.phaser.io/edit.html?src=src\game%20objects\tilemap\dynamic\randomize%20tiles.js
Changing it back to canvas works fine.
Reverting the above commit makes it work again.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/58469251-camera-shake-broken-for-tilemaps-in-webgl-since-3-7-1?utm_campaign=plugin&utm_content=tracker%2F283654&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F283654&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Camera shake broken for Tilemaps in WebGL since 3.7.1 - The following commit broke the camera shake effect when using tilemaps on WebGL
https://github.com/photonstorm/phaser/commit/1ce2311465b5e82bc0f3585b8e6dbf5b77401c3c
Can be reproduced by adding:
```
this.cameras.main.shake(2000)
```
to any of the labs using tilemaps.
This one for example:
http://labs.phaser.io/edit.html?src=src\game%20objects\tilemap\dynamic\randomize%20tiles.js
Changing it back to canvas works fine.
Reverting the above commit makes it work again.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/58469251-camera-shake-broken-for-tilemaps-in-webgl-since-3-7-1?utm_campaign=plugin&utm_content=tracker%2F283654&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F283654&utm_medium=issues&utm_source=github).
</bountysource-plugin> | priority | camera shake broken for tilemaps in webgl since the following commit broke the camera shake effect when using tilemaps on webgl can be reproduced by adding this cameras main shake to any of the labs using tilemaps this one for example changing it back to canvas works fine reverting the above commit makes it work again want to back this issue we accept bounties via | 1 |
6,660 | 2,590,685,174 | IssuesEvent | 2015-02-18 20:24:20 | olga-jane/prizm | https://api.github.com/repos/olga-jane/prizm | closed | External files one-transaction in all other places | Attachments bug bug - functional Coding MEDIUM priority Refactoring to_share_students | related tasks:
Saving files should be one transaction with saving the entity #854
Exception while adding second attachment #1290
It is impossible to create a New Component #1241
Please, implement one-transaction for external files, in other places
- [x] pipe
- [x] joint
- [x] spool
- [x] release note | 1.0 | External files one-transaction in all other places - related tasks:
Saving files should be one transaction with saving the entity #854
Exception while adding second attachment #1290
It is impossible to create a New Component #1241
Please, implement one-transaction for external files, in other places
- [x] pipe
- [x] joint
- [x] spool
- [x] release note | priority | external files one transaction in all other places related tasks saving files should be one transaction with saving the entity exception while adding second attachment it is impossible to create a new component please implement one transaction for external files in other places pipe joint spool release note | 1 |
354,256 | 10,564,562,758 | IssuesEvent | 2019-10-05 03:04:01 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | closed | Diff info for output of two ScoreCmd's | Category: Core Priority: Medium Type: Feature | Its very difficult sometimes to tell which conflated features were affected when making alg changes. The outputs of two `score-matches` command runs aren't always easy to compare to each other. This would provide a utility to make the comparison easier.
It would be nice if you could run a command against the conflated output from two 'hoot score-matches' runs and add a custom tag to the element id's in the output file that changed their match/miss/review status between the two conflations runs. This would make it much easier to determine the affect a change in a conflation algorithm has on output, going beyond just the change in correctness score. Also, maybe a text blurb in the command line about how the different types of match/miss/review counts changed between the two conflation runs.
This may need some more thinking, but:
- add a score-matches-diff command, which takes in two input osm files which were the output files of separate conflation runs with the score-matches command.
- the output of score-matches-diff is a single file containing only the elements whose match/miss/review status changed between the two conflations runs
- for each element whose status changed add two tags:
- hoot:status:before = Match/Miss/Review
- hoot:status:after = Match/Miss/Review
- hoot:status:before and hoot:status:after should always have different values for each element
- write to command line output a short summary of the changes. e.g.: "X elements changed from Match to Review, Y elements changed from Miss to Review..." etc.
| 1.0 | Diff info for output of two ScoreCmd's - Its very difficult sometimes to tell which conflated features were affected when making alg changes. The outputs of two `score-matches` command runs aren't always easy to compare to each other. This would provide a utility to make the comparison easier.
It would be nice if you could run a command against the conflated output from two 'hoot score-matches' runs and add a custom tag to the element id's in the output file that changed their match/miss/review status between the two conflations runs. This would make it much easier to determine the affect a change in a conflation algorithm has on output, going beyond just the change in correctness score. Also, maybe a text blurb in the command line about how the different types of match/miss/review counts changed between the two conflation runs.
This may need some more thinking, but:
- add a score-matches-diff command, which takes in two input osm files which were the output files of separate conflation runs with the score-matches command.
- the output of score-matches-diff is a single file containing only the elements whose match/miss/review status changed between the two conflations runs
- for each element whose status changed add two tags:
- hoot:status:before = Match/Miss/Review
- hoot:status:after = Match/Miss/Review
- hoot:status:before and hoot:status:after should always have different values for each element
- write to command line output a short summary of the changes. e.g.: "X elements changed from Match to Review, Y elements changed from Miss to Review..." etc.
| priority | diff info for output of two scorecmd s its very difficult sometimes to tell which conflated features were affected when making alg changes the outputs of two score matches command runs aren t always easy to compare to each other this would provide a utility to make the comparison easier it would be nice if you could run a command against the conflated output from two hoot score matches runs and add a custom tag to the element id s in the output file that changed their match miss review status between the two conflations runs this would make it much easier to determine the affect a change in a conflation algorithm has on output going beyond just the change in correctness score also maybe a text blurb in the command line about how the different types of match miss review counts changed between the two conflation runs this may need some more thinking but add a score matches diff command which takes in two input osm files which were the output files of separate conflation runs with the score matches command the output of score matches diff is a single file containing only the elements whose match miss review status changed between the two conflations runs for each element whose status changed add two tags hoot status before match miss review hoot status after match miss review hoot status before and hoot status after should always have different values for each element write to command line output a short summary of the changes e g x elements changed from match to review y elements changed from miss to review etc | 1 |
719,556 | 24,763,879,662 | IssuesEvent | 2022-10-22 08:47:55 | bounswe/bounswe2022group2 | https://api.github.com/repos/bounswe/bounswe2022group2 | closed | Mobile: Back-end Connection of Sign Up Endpoint | enhancement priority-medium status-inprogress mobile | ### Issue Description
As we have determined in our weekly meeting 3, we will implement signup, login, forgot password, and verify email screens for this week. We also divided the tasks among the mobile team. I took responsibility for the signup screen and I implemented the signup screen. You can see the details of this process under [this issue](https://github.com/bounswe/bounswe2022group2/pull/368). Now, I will integrate the signup endpoint developed by the back-end team into the app. However, this issue depends on the finish date of the [sign up endpoint issue](https://github.com/bounswe/bounswe2022group2/pull/366).
### Step Details
Steps that will be performed:
- [ ] Read the documentation of the signup endpoint
- [ ] Create the request and response models
- [ ] Create the network manager
- [ ] Create authentication service abstract class
- [ ] Create authentication service
- [ ] Create network request to signup endpoint
- [ ] Integrate endpoint request to the action button
- [ ] Check error cases and success functionality
### Final Actions
I will create a branch on our repo and will be committing and pushing the changes into that branch. I will also share the development details under this issue. Finally, I will create a PR and after at least 2 people from the mobile team approved the pull request, it will be merged to the main branch.
### Deadline of the Issue
25.10.2022 - Tuesday - 23:59
### Reviewer
Onur Kömürcü
### Deadline for the Review
26.10.2022 - Wednesday - 16:59 | 1.0 | Mobile: Back-end Connection of Sign Up Endpoint - ### Issue Description
As we have determined in our weekly meeting 3, we will implement signup, login, forgot password, and verify email screens for this week. We also divided the tasks among the mobile team. I took responsibility for the signup screen and I implemented the signup screen. You can see the details of this process under [this issue](https://github.com/bounswe/bounswe2022group2/pull/368). Now, I will integrate the signup endpoint developed by the back-end team into the app. However, this issue depends on the finish date of the [sign up endpoint issue](https://github.com/bounswe/bounswe2022group2/pull/366).
### Step Details
Steps that will be performed:
- [ ] Read the documentation of the signup endpoint
- [ ] Create the request and response models
- [ ] Create the network manager
- [ ] Create authentication service abstract class
- [ ] Create authentication service
- [ ] Create network request to signup endpoint
- [ ] Integrate endpoint request to the action button
- [ ] Check error cases and success functionality
### Final Actions
I will create a branch on our repo and will be committing and pushing the changes into that branch. I will also share the development details under this issue. Finally, I will create a PR and after at least 2 people from the mobile team approved the pull request, it will be merged to the main branch.
### Deadline of the Issue
25.10.2022 - Tuesday - 23:59
### Reviewer
Onur Kömürcü
### Deadline for the Review
26.10.2022 - Wednesday - 16:59 | priority | mobile back end connection of sign up endpoint issue description as we have determined in our weekly meeting we will implement signup login forgot password and verify email screens for this week we also divided the tasks among the mobile team i took responsibility for the signup screen and i implemented the signup screen you can see the details of this process under now i will integrate the signup endpoint developed by the back end team into the app however this issue depends on the finish date of the step details steps that will be performed read the documentation of the signup endpoint create the request and response models create the network manager create authentication service abstract class create authentication service create network request to signup endpoint integrate endpoint request to the action button check error cases and success functionality final actions i will create a branch on our repo and will be committing and pushing the changes into that branch i will also share the development details under this issue finally i will create a pr and after at least people from the mobile team approved the pull request it will be merged to the main branch deadline of the issue tuesday reviewer onur kömürcü deadline for the review wednesday | 1 |
249,193 | 7,954,057,537 | IssuesEvent | 2018-07-12 05:43:30 | minio/minio | https://api.github.com/repos/minio/minio | closed | Minio s3 ListMultipartUploads | priority: medium | Hello, i'm trying to list multipart uploads with python boto library (boto==2.48.0) and all time getting response without multipart uploads while in storage directory of minio i can see my not completed uploads (.minio.sys/multipart/test_bucket/file1/<UUID>/ there are fs.json, object1, object2 as I uploaded two parts), the same behavior with postman.
## Expected Behavior
https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadListMPUpload.html
Upload with file1 key
## Current Behavior
No Uploads tag at all
## Steps to Reproduce (for bugs)
1. Initiate multipart uploads (https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadInitiate.html)
2. Upload part (https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadUploadPart.html)
3. List multipart uploads (https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadListMPUpload.html)
No complete multipart was called.
Boto example:
```
import os.path
import boto.s3
from boto.s3.connection import S3Connection
conn = S3Connection("<my-access-key>", "<my-secret-key>", is_secure=false, port=9000, host=localhost,
calling_format='boto.s3.connection.OrdinaryCallingFormat')
bucket = conn.get_bucket('test_bucket')
mp = bucket.initiate_multipart_upload('file1')
with open('/tmp/largefile.zip', 'rb') as f:
part1 = mp.upload_part_from_file(f, 1, size=10*1024*1024)
part2 = mp.upload_part_from_file(f, 2, size=10*1024*1024)
print(bucket.list_multipart_uploads(upload_id_marker=mp.id)
mp.complete_upload()
```
## Context
I've got integration tests which fails when run them on minio. Probably, can be also an issue for different s3 browsers.
## Your Environment
* Version used (`minio version`): minio.RELEASE.2018-01-18T20-33-21Z
* Environment name and version (e.g. nginx 1.9.1): direct connection but inside docker image
* Operating System and version (`uname -a`): Darwin data_race 17.4.0 Darwin Kernel Version 17.4.0: Sun Dec 17 09:19:54 PST 2017; root:xnu-4570.41.2~1/RELEASE_X86_64 x86_64
| 1.0 | Minio s3 ListMultipartUploads - Hello, i'm trying to list multipart uploads with python boto library (boto==2.48.0) and all time getting response without multipart uploads while in storage directory of minio i can see my not completed uploads (.minio.sys/multipart/test_bucket/file1/<UUID>/ there are fs.json, object1, object2 as I uploaded two parts), the same behavior with postman.
## Expected Behavior
https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadListMPUpload.html
Upload with file1 key
## Current Behavior
No Uploads tag at all
## Steps to Reproduce (for bugs)
1. Initiate multipart uploads (https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadInitiate.html)
2. Upload part (https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadUploadPart.html)
3. List multipart uploads (https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadListMPUpload.html)
No complete multipart was called.
Boto example:
```
import os.path
import boto.s3
from boto.s3.connection import S3Connection
conn = S3Connection("<my-access-key>", "<my-secret-key>", is_secure=false, port=9000, host=localhost,
calling_format='boto.s3.connection.OrdinaryCallingFormat')
bucket = conn.get_bucket('test_bucket')
mp = bucket.initiate_multipart_upload('file1')
with open('/tmp/largefile.zip', 'rb') as f:
part1 = mp.upload_part_from_file(f, 1, size=10*1024*1024)
part2 = mp.upload_part_from_file(f, 2, size=10*1024*1024)
print(bucket.list_multipart_uploads(upload_id_marker=mp.id)
mp.complete_upload()
```
## Context
I've got integration tests which fails when run them on minio. Probably, can be also an issue for different s3 browsers.
## Your Environment
* Version used (`minio version`): minio.RELEASE.2018-01-18T20-33-21Z
* Environment name and version (e.g. nginx 1.9.1): direct connection but inside docker image
* Operating System and version (`uname -a`): Darwin data_race 17.4.0 Darwin Kernel Version 17.4.0: Sun Dec 17 09:19:54 PST 2017; root:xnu-4570.41.2~1/RELEASE_X86_64 x86_64
| priority | minio listmultipartuploads hello i m trying to list multipart uploads with python boto library boto and all time getting response without multipart uploads while in storage directory of minio i can see my not completed uploads minio sys multipart test bucket there are fs json as i uploaded two parts the same behavior with postman expected behavior upload with key current behavior no uploads tag at all steps to reproduce for bugs initiate multipart uploads upload part list multipart uploads no complete multipart was called boto example import os path import boto from boto connection import conn is secure false port host localhost calling format boto connection ordinarycallingformat bucket conn get bucket test bucket mp bucket initiate multipart upload with open tmp largefile zip rb as f mp upload part from file f size mp upload part from file f size print bucket list multipart uploads upload id marker mp id mp complete upload context i ve got integration tests which fails when run them on minio probably can be also an issue for different browsers your environment version used minio version minio release environment name and version e g nginx direct connection but inside docker image operating system and version uname a darwin data race darwin kernel version sun dec pst root xnu release | 1 |
213,496 | 7,254,209,973 | IssuesEvent | 2018-02-16 10:02:54 | salesagility/SuiteCRM | https://api.github.com/repos/salesagility/SuiteCRM | closed | TinyMCE or SuiteCRM strips barcode tags in PDF-Templates | Fix Proposed Medium Priority Resolved: Next Release bug | #### Issue
When editing a PDF Template in HTML mode, to add a` <barcode>` element it gets stripped.
#### Expected Behavior
Barcode tags are saved in the PDF templates
#### Actual Behavior
TinyMCE strips these tags, so that a barcode will generated when you use generate a PDF.
#### Possible Fix
Fix settings of TinyMCE to allow those tags to be saved. When you add the tags directly in the database, thus side stepping TinyMCE it works without problems. http://support.sugarcrm.com/Documentation/Sugar_Developer/Sugar_Developer_Guide_6.5/Application_Framework/TinyMCE/Modifying_the_TinyMCE_Editor/
Maybe it's not related to TinyMCE at all since there is similiar build in behavior in SuiteCRM itself:
see #3672
#### Steps to Reproduce
1. Create or Edit a PDF Template
2. Open the HTML view of the document body
3. Add a barcode tag to the source `<barcode code="978-0-9542246-0" type="ISBN" height="0.66" text="1" />`
4. Click Update
#### Your Environment
* SuiteCRM Version used: Version 7.9.5
* Browser name and version: Firefox Nightly (58.0a1)
* Environment name and version: PHP Version 5.6.30-0+deb8u1, Mysql 5.5.55
* Operating System and version: Debian 9
| 1.0 | TinyMCE or SuiteCRM strips barcode tags in PDF-Templates - #### Issue
When editing a PDF Template in HTML mode, to add a` <barcode>` element it gets stripped.
#### Expected Behavior
Barcode tags are saved in the PDF templates
#### Actual Behavior
TinyMCE strips these tags, so that a barcode will generated when you use generate a PDF.
#### Possible Fix
Fix settings of TinyMCE to allow those tags to be saved. When you add the tags directly in the database, thus side stepping TinyMCE it works without problems. http://support.sugarcrm.com/Documentation/Sugar_Developer/Sugar_Developer_Guide_6.5/Application_Framework/TinyMCE/Modifying_the_TinyMCE_Editor/
Maybe it's not related to TinyMCE at all since there is similiar build in behavior in SuiteCRM itself:
see #3672
#### Steps to Reproduce
1. Create or Edit a PDF Template
2. Open the HTML view of the document body
3. Add a barcode tag to the source `<barcode code="978-0-9542246-0" type="ISBN" height="0.66" text="1" />`
4. Click Update
#### Your Environment
* SuiteCRM Version used: Version 7.9.5
* Browser name and version: Firefox Nightly (58.0a1)
* Environment name and version: PHP Version 5.6.30-0+deb8u1, Mysql 5.5.55
* Operating System and version: Debian 9
| priority | tinymce or suitecrm strips barcode tags in pdf templates issue when editing a pdf template in html mode to add a element it gets stripped expected behavior barcode tags are saved in the pdf templates actual behavior tinymce strips these tags so that a barcode will generated when you use generate a pdf possible fix fix settings of tinymce to allow those tags to be saved when you add the tags directly in the database thus side stepping tinymce it works without problems maybe it s not related to tinymce at all since there is similiar build in behavior in suitecrm itself see steps to reproduce create or edit a pdf template open the html view of the document body add a barcode tag to the source click update your environment suitecrm version used version browser name and version firefox nightly environment name and version php version mysql operating system and version debian | 1 |
752,702 | 26,321,297,487 | IssuesEvent | 2023-01-10 00:07:55 | Alluxio/alluxio | https://api.github.com/repos/Alluxio/alluxio | closed | The policy LocalFirstAvoidEvictionPolicy doesn't not work for Alluxio Fuse | type-docs priority-medium area-k8s | **Alluxio Version:**
2.1.0-SNAPSHOT
**Describe the bug**
I'd like to use LocalFirstAvoidEvictionPolicy for alluxio Fuse(client). The purpose is that the data is loaded to the worker in the same node first, and when there are not enough space in the same node, it won't evict the current node. Instead it will turn to another available node(with enough space) . But the result is the eviction still occurs in the same node.
1. Make the fuse client with `LocalFirstAvoidEvictionPolicy`
```
/usr/bin/java\
-cp :/opt/alluxio/integration/fuse/bin/../alluxio-fuse-2.1.0-SNAPSHOT.jar
-server\
-Xms1G\
-Xmx1G\
-XX:MaxDirectMemorySize=4g\
-Dalluxio.master.hostname=alluxio-master\
-Dalluxio.user.hostname=192.168.0.128\
-Dalluxio.worker.hostname=192.168.0.128\
-XX:+UnlockExperimentalVMOptions\
-XX:+UseCGroupMemoryLimitForHeap\
-XX:MaxRAMFraction=2\
-XX:MaxDirectMemorySize=8g\
-Dalluxio.worker.tieredstore.levels=1\
-Dalluxio.worker.tieredstore.level0.alias=MEM\
-Dalluxio.worker.tieredstore.level0.dirs.mediumtype=MEM,SSD\
-Dalluxio.worker.tieredstore.level0.dirs.path=/dev/shm,/alluxio-ssd\
-Dalluxio.worker.tieredstore.level0.dirs.quota=10GB,50GB\
-Dalluxio.worker.tieredstore.level0.watermark.high.ratio=0.95\
-Dalluxio.worker.tieredstore.level0.watermark.low.ratio=0.7\
-Dalluxio.master.journal.folder=/journal\
-Dalluxio.master.journal.type=UFS\
-Dalluxio.security.stale.channel.purge.interval=365d\
-Dalluxio.master.journal.folder=/journal\
-Dalluxio.master.journal.type=UFS\
-Dalluxio.security.stale.channel.purge.interval=365d\
-Dalluxio.user.file.passive.cache.enabled=false\
-Dalluxio.user.ufs.block.read.location.policy=alluxio.client.block.policy.LocalFirstAvoidEvictionPolicy\
-Dalluxio.home=/opt/alluxio-2.1.0-SNAPSHOT\
-Dalluxio.conf.dir=/opt/alluxio-2.1.0-SNAPSHOT/conf\
-Dalluxio.logs.dir=/opt/alluxio-2.1.0-SNAPSHOT/logs\
-Dalluxio.user.logs.dir=/opt/alluxio-2.1.0-SNAPSHOT/logs/user\
-Dlog4j.configuration=file:/opt/alluxio-2.1.0-SNAPSHOT/conf/log4j.properties\
-Dorg.apache.jasper.compiler.disablejsr199=true\
-Djava.net.preferIPv4Stack=true\
-Dalluxio.logger.type=FUSE_LOGGER,Console alluxio.fuse.AlluxioFuse\
-o big_writes,allow_other,kernel_cache,auto_cache,entry_timeout=120,attr_timeout=120\
-m /alluxio-fuse -r /
```
You can see `-Dalluxio.user.ufs.block.read.location.policy=alluxio.client.block.policy.LocalFirstAvoidEviction
`
2. Go to the node `192.168.0.128`, and copy the data from alluxio fuse to local directory
```
# du -sh /alluxio-fuse/imagenet-nas/
144G /alluxio-fuse/imagenet-nas/
# time cp -r /alluxio-fuse/imagenet-nas/training-data/images /var/lib/docker/images/
real 29m7.499s
user 0m1.231s
sys 2m36.773s
```
3. Check the capacity, I can see only the node `192.168.0.128` has cache data, other nodes have no cache data.
```
# /opt/alluxio/bin/alluxio fsadmin report capacity
Capacity information for all workers:
Total Capacity: 240.00GB
Tier: MEM Size: 240.00GB
Used Capacity: 44.47GB
Tier: MEM Size: 44.47GB
Used Percentage: 18%
Free Percentage: 82%
Worker Name Last Heartbeat Storage MEM
192.168.0.125 0 capacity 60.00GB
used 0B (0%)
192.168.0.128 0 capacity 60.00GB
used 44.47GB (74%)
192.168.0.126 0 capacity 60.00GB
used 0B (0%)
192.168.0.127 0 capacity 60.00GB
used 0B (0%)
```
4. Check the master logs, I can see the node 192.168.0.128 has been evicted.
```
2019-09-14 07:07:50,514 INFO DefaultBlockMaster - getWorkerId(): WorkerNetAddress: WorkerNetAddress{host=192.168.0.128, rpcPort=29999, dataPort=29999, webPort=30000, domainSocketPath=, tieredIdentity=TieredIdentity(node=192.168.0.128, rack=null)} id: 7259287311378747917
2019-09-14 07:07:50,566 INFO DefaultBlockMaster - registerWorker(): MasterWorkerInfo{id=7259287311378747917, workerAddress=WorkerNetAddress{host=192.168.0.128, rpcPort=29999, dataPort=29999, webPort=30000, domainSocketPath=, tieredIdentity=TieredIdentity(node=192.168.0.128, rack=null)}, capacityBytes=64424509440, usedBytes=0, lastUpdatedTimeMs=1568444870566, blocks=[], lostStorage={}}
2019-09-14 07:29:58,698 INFO DefaultBlockMaster - Block 9244246016 is removed on worker 7259287311378747917.
2019-09-14 07:29:58,699 INFO DefaultBlockMaster - Block 9244246017 is removed on worker 7259287311378747917.
2019-09-14 07:29:58,699 INFO DefaultBlockMaster - Block 9244246018 is removed on worker 7259287311378747917.
2019-09-14 07:29:58,699 INFO DefaultBlockMaster - Block 15720251392 is removed on worker 7259287311378747917.
2019-09-14 07:29:58,699 INFO DefaultBlockMaster - Block 15720251393 is removed on worker 7259287311378747917.
```
Please let me know if there is any thing wrong in my configuration?
**To Reproduce**
Steps to reproduce the behavior (as minimally and precisely as possible)
**Expected behavior**
A clear and concise description of what you expected to happen.
**Urgency**
Describe the impact and urgency of the bug.
**Additional context**
Add any other context about the problem here.
| 1.0 | The policy LocalFirstAvoidEvictionPolicy doesn't not work for Alluxio Fuse - **Alluxio Version:**
2.1.0-SNAPSHOT
**Describe the bug**
I'd like to use LocalFirstAvoidEvictionPolicy for alluxio Fuse(client). The purpose is that the data is loaded to the worker in the same node first, and when there are not enough space in the same node, it won't evict the current node. Instead it will turn to another available node(with enough space) . But the result is the eviction still occurs in the same node.
1. Make the fuse client with `LocalFirstAvoidEvictionPolicy`
```
/usr/bin/java\
-cp :/opt/alluxio/integration/fuse/bin/../alluxio-fuse-2.1.0-SNAPSHOT.jar
-server\
-Xms1G\
-Xmx1G\
-XX:MaxDirectMemorySize=4g\
-Dalluxio.master.hostname=alluxio-master\
-Dalluxio.user.hostname=192.168.0.128\
-Dalluxio.worker.hostname=192.168.0.128\
-XX:+UnlockExperimentalVMOptions\
-XX:+UseCGroupMemoryLimitForHeap\
-XX:MaxRAMFraction=2\
-XX:MaxDirectMemorySize=8g\
-Dalluxio.worker.tieredstore.levels=1\
-Dalluxio.worker.tieredstore.level0.alias=MEM\
-Dalluxio.worker.tieredstore.level0.dirs.mediumtype=MEM,SSD\
-Dalluxio.worker.tieredstore.level0.dirs.path=/dev/shm,/alluxio-ssd\
-Dalluxio.worker.tieredstore.level0.dirs.quota=10GB,50GB\
-Dalluxio.worker.tieredstore.level0.watermark.high.ratio=0.95\
-Dalluxio.worker.tieredstore.level0.watermark.low.ratio=0.7\
-Dalluxio.master.journal.folder=/journal\
-Dalluxio.master.journal.type=UFS\
-Dalluxio.security.stale.channel.purge.interval=365d\
-Dalluxio.master.journal.folder=/journal\
-Dalluxio.master.journal.type=UFS\
-Dalluxio.security.stale.channel.purge.interval=365d\
-Dalluxio.user.file.passive.cache.enabled=false\
-Dalluxio.user.ufs.block.read.location.policy=alluxio.client.block.policy.LocalFirstAvoidEvictionPolicy\
-Dalluxio.home=/opt/alluxio-2.1.0-SNAPSHOT\
-Dalluxio.conf.dir=/opt/alluxio-2.1.0-SNAPSHOT/conf\
-Dalluxio.logs.dir=/opt/alluxio-2.1.0-SNAPSHOT/logs\
-Dalluxio.user.logs.dir=/opt/alluxio-2.1.0-SNAPSHOT/logs/user\
-Dlog4j.configuration=file:/opt/alluxio-2.1.0-SNAPSHOT/conf/log4j.properties\
-Dorg.apache.jasper.compiler.disablejsr199=true\
-Djava.net.preferIPv4Stack=true\
-Dalluxio.logger.type=FUSE_LOGGER,Console alluxio.fuse.AlluxioFuse\
-o big_writes,allow_other,kernel_cache,auto_cache,entry_timeout=120,attr_timeout=120\
-m /alluxio-fuse -r /
```
You can see `-Dalluxio.user.ufs.block.read.location.policy=alluxio.client.block.policy.LocalFirstAvoidEviction
`
2. Go to the node `192.168.0.128`, and copy the data from alluxio fuse to local directory
```
# du -sh /alluxio-fuse/imagenet-nas/
144G /alluxio-fuse/imagenet-nas/
# time cp -r /alluxio-fuse/imagenet-nas/training-data/images /var/lib/docker/images/
real 29m7.499s
user 0m1.231s
sys 2m36.773s
```
3. Check the capacity, I can see only the node `192.168.0.128` has cache data, other nodes have no cache data.
```
# /opt/alluxio/bin/alluxio fsadmin report capacity
Capacity information for all workers:
Total Capacity: 240.00GB
Tier: MEM Size: 240.00GB
Used Capacity: 44.47GB
Tier: MEM Size: 44.47GB
Used Percentage: 18%
Free Percentage: 82%
Worker Name Last Heartbeat Storage MEM
192.168.0.125 0 capacity 60.00GB
used 0B (0%)
192.168.0.128 0 capacity 60.00GB
used 44.47GB (74%)
192.168.0.126 0 capacity 60.00GB
used 0B (0%)
192.168.0.127 0 capacity 60.00GB
used 0B (0%)
```
4. Check the master logs, I can see the node 192.168.0.128 has been evicted.
```
2019-09-14 07:07:50,514 INFO DefaultBlockMaster - getWorkerId(): WorkerNetAddress: WorkerNetAddress{host=192.168.0.128, rpcPort=29999, dataPort=29999, webPort=30000, domainSocketPath=, tieredIdentity=TieredIdentity(node=192.168.0.128, rack=null)} id: 7259287311378747917
2019-09-14 07:07:50,566 INFO DefaultBlockMaster - registerWorker(): MasterWorkerInfo{id=7259287311378747917, workerAddress=WorkerNetAddress{host=192.168.0.128, rpcPort=29999, dataPort=29999, webPort=30000, domainSocketPath=, tieredIdentity=TieredIdentity(node=192.168.0.128, rack=null)}, capacityBytes=64424509440, usedBytes=0, lastUpdatedTimeMs=1568444870566, blocks=[], lostStorage={}}
2019-09-14 07:29:58,698 INFO DefaultBlockMaster - Block 9244246016 is removed on worker 7259287311378747917.
2019-09-14 07:29:58,699 INFO DefaultBlockMaster - Block 9244246017 is removed on worker 7259287311378747917.
2019-09-14 07:29:58,699 INFO DefaultBlockMaster - Block 9244246018 is removed on worker 7259287311378747917.
2019-09-14 07:29:58,699 INFO DefaultBlockMaster - Block 15720251392 is removed on worker 7259287311378747917.
2019-09-14 07:29:58,699 INFO DefaultBlockMaster - Block 15720251393 is removed on worker 7259287311378747917.
```
Please let me know if there is any thing wrong in my configuration?
**To Reproduce**
Steps to reproduce the behavior (as minimally and precisely as possible)
**Expected behavior**
A clear and concise description of what you expected to happen.
**Urgency**
Describe the impact and urgency of the bug.
**Additional context**
Add any other context about the problem here.
| priority | the policy localfirstavoidevictionpolicy doesn t not work for alluxio fuse alluxio version snapshot describe the bug i d like to use localfirstavoidevictionpolicy for alluxio fuse client the purpose is that the data is loaded to the worker in the same node first and when there are not enough space in the same node it won t evict the current node instead it will turn to another available node with enough space but the result is the eviction still occurs in the same node make the fuse client with localfirstavoidevictionpolicy usr bin java cp opt alluxio integration fuse bin alluxio fuse snapshot jar server xx maxdirectmemorysize dalluxio master hostname alluxio master dalluxio user hostname dalluxio worker hostname xx unlockexperimentalvmoptions xx usecgroupmemorylimitforheap xx maxramfraction xx maxdirectmemorysize dalluxio worker tieredstore levels dalluxio worker tieredstore alias mem dalluxio worker tieredstore dirs mediumtype mem ssd dalluxio worker tieredstore dirs path dev shm alluxio ssd dalluxio worker tieredstore dirs quota dalluxio worker tieredstore watermark high ratio dalluxio worker tieredstore watermark low ratio dalluxio master journal folder journal dalluxio master journal type ufs dalluxio security stale channel purge interval dalluxio master journal folder journal dalluxio master journal type ufs dalluxio security stale channel purge interval dalluxio user file passive cache enabled false dalluxio user ufs block read location policy alluxio client block policy localfirstavoidevictionpolicy dalluxio home opt alluxio snapshot dalluxio conf dir opt alluxio snapshot conf dalluxio logs dir opt alluxio snapshot logs dalluxio user logs dir opt alluxio snapshot logs user configuration file opt alluxio snapshot conf properties dorg apache jasper compiler true djava net true dalluxio logger type fuse logger console alluxio fuse alluxiofuse o big writes allow other kernel cache auto cache entry timeout attr timeout m alluxio fuse r you can see dalluxio user ufs block read location policy alluxio client block policy localfirstavoideviction go to the node and copy the data from alluxio fuse to local directory du sh alluxio fuse imagenet nas alluxio fuse imagenet nas time cp r alluxio fuse imagenet nas training data images var lib docker images real user sys check the capacity i can see only the node has cache data other nodes have no cache data opt alluxio bin alluxio fsadmin report capacity capacity information for all workers total capacity tier mem size used capacity tier mem size used percentage free percentage worker name last heartbeat storage mem capacity used capacity used capacity used capacity used check the master logs i can see the node has been evicted info defaultblockmaster getworkerid workernetaddress workernetaddress host rpcport dataport webport domainsocketpath tieredidentity tieredidentity node rack null id info defaultblockmaster registerworker masterworkerinfo id workeraddress workernetaddress host rpcport dataport webport domainsocketpath tieredidentity tieredidentity node rack null capacitybytes usedbytes lastupdatedtimems blocks loststorage info defaultblockmaster block is removed on worker info defaultblockmaster block is removed on worker info defaultblockmaster block is removed on worker info defaultblockmaster block is removed on worker info defaultblockmaster block is removed on worker please let me know if there is any thing wrong in my configuration to reproduce steps to reproduce the behavior as minimally and precisely as possible expected behavior a clear and concise description of what you expected to happen urgency describe the impact and urgency of the bug additional context add any other context about the problem here | 1 |
311,533 | 9,534,804,966 | IssuesEvent | 2019-04-30 03:34:34 | vectorlit/UnofficialGenconMobile | https://api.github.com/repos/vectorlit/UnofficialGenconMobile | opened | Feature: Additional Sort/Filter Features | Medium Priority enhancement | This is the 2nd most common request from 2018: Add additional sort and filter options to the application, and fix broken ones:
- [ ] Add option to hide sold-out events (they are already grayed out, but extra option to hide)
- [ ] Fix the price sort to have a secondary sort-by time default (currently it's just a primary sort with random secondary)
- [ ] Add option to filter by category (ANI, TRD, CGM, etc)
- [ ] Anything else you can think of | 1.0 | Feature: Additional Sort/Filter Features - This is the 2nd most common request from 2018: Add additional sort and filter options to the application, and fix broken ones:
- [ ] Add option to hide sold-out events (they are already grayed out, but extra option to hide)
- [ ] Fix the price sort to have a secondary sort-by time default (currently it's just a primary sort with random secondary)
- [ ] Add option to filter by category (ANI, TRD, CGM, etc)
- [ ] Anything else you can think of | priority | feature additional sort filter features this is the most common request from add additional sort and filter options to the application and fix broken ones add option to hide sold out events they are already grayed out but extra option to hide fix the price sort to have a secondary sort by time default currently it s just a primary sort with random secondary add option to filter by category ani trd cgm etc anything else you can think of | 1 |
439,345 | 12,681,568,986 | IssuesEvent | 2020-06-19 15:38:10 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | Application Add | component:ui_next priority:medium | ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
Add support for adding an Application | 1.0 | Application Add - ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
Add support for adding an Application | priority | application add issue type feature idea summary add support for adding an application | 1 |
333,021 | 10,114,171,624 | IssuesEvent | 2019-07-30 18:31:39 | mit-cml/appinventor-sources | https://api.github.com/repos/mit-cml/appinventor-sources | opened | Support fill-opacity as an option for Map features | affects: ucr feature request help wanted issue: noted for future Work priority: medium status: new | [From the forum](https://groups.google.com/d/msg/mitappinventortest/EaqbByAhLkM/1oVN2_P1DwAJ): App Inventor supports SVG properties on features such as stroke, stroke-color, and fill-color. However, fill-opacity is not supported (and probably stroke-opacity). We should add support for these capabilities. It's important to note that we do support using ARGB hex values, so it can also be solved by updating the fill property to include the alpha channel in the color. | 1.0 | Support fill-opacity as an option for Map features - [From the forum](https://groups.google.com/d/msg/mitappinventortest/EaqbByAhLkM/1oVN2_P1DwAJ): App Inventor supports SVG properties on features such as stroke, stroke-color, and fill-color. However, fill-opacity is not supported (and probably stroke-opacity). We should add support for these capabilities. It's important to note that we do support using ARGB hex values, so it can also be solved by updating the fill property to include the alpha channel in the color. | priority | support fill opacity as an option for map features app inventor supports svg properties on features such as stroke stroke color and fill color however fill opacity is not supported and probably stroke opacity we should add support for these capabilities it s important to note that we do support using argb hex values so it can also be solved by updating the fill property to include the alpha channel in the color | 1 |
632,105 | 20,172,517,351 | IssuesEvent | 2022-02-10 11:43:12 | canonical-web-and-design/jp.ubuntu.com | https://api.github.com/repos/canonical-web-and-design/jp.ubuntu.com | closed | Failed to load resource: the server responded with a status of 404 () | Priority: Medium | ## Summary

There are some error messages `Failed to load resource: the server responded with a status of 404 ()`in the console on the `/download` pages
e.g.
https://jp.ubuntu.com/download/thank-you?version=20.04.3&architecture=amd64&platform=desktop
https://jp.ubuntu.com/download/thank-you?version=20.04.3&architecture=amd64&platform=live-server
| 1.0 | Failed to load resource: the server responded with a status of 404 () - ## Summary

There are some error messages `Failed to load resource: the server responded with a status of 404 ()`in the console on the `/download` pages
e.g.
https://jp.ubuntu.com/download/thank-you?version=20.04.3&architecture=amd64&platform=desktop
https://jp.ubuntu.com/download/thank-you?version=20.04.3&architecture=amd64&platform=live-server
| priority | failed to load resource the server responded with a status of summary there are some error messages failed to load resource the server responded with a status of in the console on the download pages e g | 1 |
178,140 | 6,599,383,207 | IssuesEvent | 2017-09-16 19:00:39 | qlicker/qlicker | https://api.github.com/repos/qlicker/qlicker | opened | Should prevent submission of student question if type not chosen | bug Medium priority | Currently, in the modal, if a student submits a question with no type, they get no feedback that the question was not submitted. | 1.0 | Should prevent submission of student question if type not chosen - Currently, in the modal, if a student submits a question with no type, they get no feedback that the question was not submitted. | priority | should prevent submission of student question if type not chosen currently in the modal if a student submits a question with no type they get no feedback that the question was not submitted | 1 |
399,906 | 11,763,394,560 | IssuesEvent | 2020-03-14 06:58:45 | CalPolyRobotics/IGVC-ROS | https://api.github.com/repos/CalPolyRobotics/IGVC-ROS | opened | Add ntripclient for RTK correction data | pkg: tools priority: medium | The ntripclient program provides a way to send the RTK correction data to the connected receiver.
- [ ] Add ntripclient to the repo.
- [ ] Create a ROS package to wrap ntripclient. | 1.0 | Add ntripclient for RTK correction data - The ntripclient program provides a way to send the RTK correction data to the connected receiver.
- [ ] Add ntripclient to the repo.
- [ ] Create a ROS package to wrap ntripclient. | priority | add ntripclient for rtk correction data the ntripclient program provides a way to send the rtk correction data to the connected receiver add ntripclient to the repo create a ros package to wrap ntripclient | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.