Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
118,378 | 4,738,328,886 | IssuesEvent | 2016-10-20 03:37:29 | CS2103AUG2016-T10-C2/main | https://api.github.com/repos/CS2103AUG2016-T10-C2/main | closed | Code Cleaning | priority.medium type.enhancement | Change approach for prepareEdit in Parser.
Refer to #92 for the suggestions provided by @kabir-plod | 1.0 | Code Cleaning - Change approach for prepareEdit in Parser.
Refer to #92 for the suggestions provided by @kabir-plod | priority | code cleaning change approach for prepareedit in parser refer to for the suggestions provided by kabir plod | 1 |
171,108 | 6,479,096,078 | IssuesEvent | 2017-08-18 09:41:24 | pest-parser/pest | https://api.github.com/repos/pest-parser/pest | closed | Matchers that do not make progress are difficult to use in error handling | Priority: Medium Status: Pending Type: Feature/Enhancement | Let's say I have a rule like this:
```rust
identifier = { !keyword ~ alpha ~ alphanumeric* }
keyword = { ["foo"] }
```
If I try to parse the word `foo`, it doesn't parse (as expected), but I don't get enough information to provide a useful error to the user.
`parser.expected()` returns `[identifier]` and all the other methods return `[]`. So the error I would report to the user is:
```
Expected an identifier
```
Notice that the rest of the rule is `alpha ~ alphanumeric*`. The word `foo` clearly matches that criteria, but the identifier rule fails to parse because of the `!keyword` part.
It would be more useful to print:
```
Expected an identifier, found a keyword instead.
```
Or something else letting them know what really happened.
This applies to all matchers that do not make progress like `!` and `&`. | 1.0 | Matchers that do not make progress are difficult to use in error handling - Let's say I have a rule like this:
```rust
identifier = { !keyword ~ alpha ~ alphanumeric* }
keyword = { ["foo"] }
```
If I try to parse the word `foo`, it doesn't parse (as expected), but I don't get enough information to provide a useful error to the user.
`parser.expected()` returns `[identifier]` and all the other methods return `[]`. So the error I would report to the user is:
```
Expected an identifier
```
Notice that the rest of the rule is `alpha ~ alphanumeric*`. The word `foo` clearly matches that criteria, but the identifier rule fails to parse because of the `!keyword` part.
It would be more useful to print:
```
Expected an identifier, found a keyword instead.
```
Or something else letting them know what really happened.
This applies to all matchers that do not make progress like `!` and `&`. | priority | matchers that do not make progress are difficult to use in error handling let s say i have a rule like this rust identifier keyword alpha alphanumeric keyword if i try to parse the word foo it doesn t parse as expected but i don t get enough information to provide a useful error to the user parser expected returns and all the other methods return so the error i would report to the user is expected an identifier notice that the rest of the rule is alpha alphanumeric the word foo clearly matches that criteria but the identifier rule fails to parse because of the keyword part it would be more useful to print expected an identifier found a keyword instead or something else letting them know what really happened this applies to all matchers that do not make progress like and | 1 |
519,419 | 15,050,932,295 | IssuesEvent | 2021-02-03 13:31:15 | rsherrera/project-caja | https://api.github.com/repos/rsherrera/project-caja | opened | Judicial. Pago Rendición. Cancela Aporte. Orden de las cuotas listadas. | Priority-Medium bug | El sistema muestra desordenado el listado de cuotas en pantalla Cancela Aporte.
Debe ordenarse desde la cuota mas antigua a la mas nueva. | 1.0 | Judicial. Pago Rendición. Cancela Aporte. Orden de las cuotas listadas. - El sistema muestra desordenado el listado de cuotas en pantalla Cancela Aporte.
Debe ordenarse desde la cuota mas antigua a la mas nueva. | priority | judicial pago rendición cancela aporte orden de las cuotas listadas el sistema muestra desordenado el listado de cuotas en pantalla cancela aporte debe ordenarse desde la cuota mas antigua a la mas nueva | 1 |
558,152 | 16,527,493,140 | IssuesEvent | 2021-05-26 22:27:09 | actually-colab/editor | https://api.github.com/repos/actually-colab/editor | opened | Emit websocket errors | AWS bug client difficulty: hard priority: medium server socket | API Gateway Route Responses do not properly support emitting errors. Come up with an alternative such as emitting errors through messages. | 1.0 | Emit websocket errors - API Gateway Route Responses do not properly support emitting errors. Come up with an alternative such as emitting errors through messages. | priority | emit websocket errors api gateway route responses do not properly support emitting errors come up with an alternative such as emitting errors through messages | 1 |
87,705 | 3,757,176,058 | IssuesEvent | 2016-03-13 20:40:18 | svof/svof | https://api.github.com/repos/svof/svof | opened | Consider new tree scenario against 2 handed | enhancement low priority medium difficulty up for grabs | reported by Aerek
> A Touch Tree scenario for curing when you ONLY have 2H fractures and no other affs would be useful. I can't figure out how to script this, myself | 1.0 | Consider new tree scenario against 2 handed - reported by Aerek
> A Touch Tree scenario for curing when you ONLY have 2H fractures and no other affs would be useful. I can't figure out how to script this, myself | priority | consider new tree scenario against handed reported by aerek a touch tree scenario for curing when you only have fractures and no other affs would be useful i can t figure out how to script this myself | 1 |
275,591 | 8,577,325,183 | IssuesEvent | 2018-11-12 23:29:17 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | segfault in python multithreaded setting | bug medium priority | Zihang Dai reports (and I've reproduced) that the autograd engine is not thread-safe.
Here's a repro script: https://gist.github.com/zihangdai/fc8f76fbb8a0f6323a6b31e6d98ceb50
Run it a few times, occassionally it segfaults.
Segfault is from a much different location, when cleaning up imports:
```
Thread 1 "python" received signal SIGSEGV, Segmentation fault.
malloc_consolidate (av=av@entry=0x7ffff70a8b20 <main_arena>) at malloc.c:4167
4167 malloc.c: No such file or directory.
(gdb) where
#0 malloc_consolidate (av=av@entry=0x7ffff70a8b20 <main_arena>) at malloc.c:4167
#1 0x00007ffff6d64678 in _int_free (av=0x7ffff70a8b20 <main_arena>, p=<optimized out>, have_lock=0) at malloc.c:4075
#2 0x00007ffff6d6853c in __GI___libc_free (mem=<optimized out>) at malloc.c:2968
#3 0x00007ffff7a6a427 in dict_dealloc (mp=0x7ffff7ed0c58) at Objects/dictobject.c:1044
#4 0x00007ffff7a682c7 in insertdict_by_entry (mp=0x7ffff7ed0d70, key='build_time_vars', hash=<optimized out>, ep=<optimized out>, value=<optimized out>) at Objects/dictobject.c:519
#5 0x00007ffff7a6b79c in insertdict (value=None, hash=-295987683324531010, key='build_time_vars', mp=0x7ffff7ed0d70) at Objects/dictobject.c:556
#6 dict_set_item_by_hash_or_entry (value=None, ep=0x0, hash=-295987683324531010, key='build_time_vars',
op={'__builtins__': {'bytearray': <type at remote 0x7ffff7d7c300>, 'IndexError': <type at remote 0x7ffff7d82bc0>, 'all': <built-in function all>, 'help': <_Helper at remote 0x7ffff7ed8c50>, 'vars': <built-in function vars>, 'SyntaxError': <type at remote 0x7ffff7d82540>, 'unicode': <type at remote 0x7ffff7d99040>, 'UnicodeDecodeError': <type at remote 0x7ffff7d833e0>, 'memoryview': <type at remote 0x7ffff7d8d900>, 'isinstance': <built-in function isinstance>, 'copyright': <_Printer(_Printer__data='Copyright (c) 2001-2016 Python Software Foundation.\nAll Rights Reserved.\n\nCopyright (c) 2000 BeOpen.com.\nAll Rights Reserved.\n\nCopyright (c) 1995-2001 Corporation for National Research Initiatives.\nAll Rights Reserved.\n\nCopyright (c) 1991-1995 Stichting Mathematisch Centrum, Amsterdam.\nAll Rights Reserved.', _Printer__lines=None, _Printer__name='copyright', _Printer__dirs=(), _Printer__files=(...)) at remote 0x7ffff7ed8a10>, 'NameError': <type at remote 0x7ffff7d82060>, 'BytesWarning': <type at remote 0x7ffff...(truncated)) at Objects/dictobject.c:795
#7 PyDict_SetItem (
op={'__builtins__': {'bytearray': <type at remote 0x7ffff7d7c300>, 'IndexError': <type at remote 0x7ffff7d82bc0>, 'all': <built-in function all>, 'help': <_Helper at remote 0x7ffff7ed8c50>, 'vars': <built-in function vars>, 'SyntaxError': <type at remote 0x7ffff7d82540>, 'unicode': <type at remote 0x7ffff7d99040>, 'UnicodeDecodeError': <type at remote 0x7ffff7d833e0>, 'memoryview': <type at remote 0x7ffff7d8d900>, 'isinstance': <built-in function isinstance>, 'copyright': <_Printer(_Printer__data='Copyright (c) 2001-2016 Python Software Foundation.\nAll Rights Reserved.\n\nCopyright (c) 2000 BeOpen.com.\nAll Rights Reserved.\n\nCopyright (c) 1995-2001 Corporation for National Research Initiatives.\nAll Rights Reserved.\n\nCopyright (c) 1991-1995 Stichting Mathematisch Centrum, Amsterdam.\nAll Rights Reserved.', _Printer__lines=None, _Printer__name='copyright', _Printer__dirs=(), _Printer__files=(...)) at remote 0x7ffff7ed8a10>, 'NameError': <type at remote 0x7ffff7d82060>, 'BytesWarning': <type at remote 0x7ffff...(truncated), key='build_time_vars', value=None)
at Objects/dictobject.c:848
#8 0x00007ffff7a6ea8d in _PyModule_Clear (m=<optimized out>) at Objects/moduleobject.c:139
#9 0x00007ffff7aec4a1 in PyImport_Cleanup () at Python/import.c:512
#10 0x00007ffff7af957b in Py_Finalize () at Python/pythonrun.c:458
#11 0x00007ffff7b0f8e5 in Py_Main (argc=<optimized out>, argv=<optimized out>) at Modules/main.c:670
#12 0x00007ffff6d04830 in __libc_start_main (main=0x4007f0 <main>, argc=2, argv=0x7fffffffe008, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffdff8) at ../csu/libc-start.c:291
#13 0x0000000000400729 in _start ()
```
| 1.0 | segfault in python multithreaded setting - Zihang Dai reports (and I've reproduced) that the autograd engine is not thread-safe.
Here's a repro script: https://gist.github.com/zihangdai/fc8f76fbb8a0f6323a6b31e6d98ceb50
Run it a few times, occassionally it segfaults.
Segfault is from a much different location, when cleaning up imports:
```
Thread 1 "python" received signal SIGSEGV, Segmentation fault.
malloc_consolidate (av=av@entry=0x7ffff70a8b20 <main_arena>) at malloc.c:4167
4167 malloc.c: No such file or directory.
(gdb) where
#0 malloc_consolidate (av=av@entry=0x7ffff70a8b20 <main_arena>) at malloc.c:4167
#1 0x00007ffff6d64678 in _int_free (av=0x7ffff70a8b20 <main_arena>, p=<optimized out>, have_lock=0) at malloc.c:4075
#2 0x00007ffff6d6853c in __GI___libc_free (mem=<optimized out>) at malloc.c:2968
#3 0x00007ffff7a6a427 in dict_dealloc (mp=0x7ffff7ed0c58) at Objects/dictobject.c:1044
#4 0x00007ffff7a682c7 in insertdict_by_entry (mp=0x7ffff7ed0d70, key='build_time_vars', hash=<optimized out>, ep=<optimized out>, value=<optimized out>) at Objects/dictobject.c:519
#5 0x00007ffff7a6b79c in insertdict (value=None, hash=-295987683324531010, key='build_time_vars', mp=0x7ffff7ed0d70) at Objects/dictobject.c:556
#6 dict_set_item_by_hash_or_entry (value=None, ep=0x0, hash=-295987683324531010, key='build_time_vars',
op={'__builtins__': {'bytearray': <type at remote 0x7ffff7d7c300>, 'IndexError': <type at remote 0x7ffff7d82bc0>, 'all': <built-in function all>, 'help': <_Helper at remote 0x7ffff7ed8c50>, 'vars': <built-in function vars>, 'SyntaxError': <type at remote 0x7ffff7d82540>, 'unicode': <type at remote 0x7ffff7d99040>, 'UnicodeDecodeError': <type at remote 0x7ffff7d833e0>, 'memoryview': <type at remote 0x7ffff7d8d900>, 'isinstance': <built-in function isinstance>, 'copyright': <_Printer(_Printer__data='Copyright (c) 2001-2016 Python Software Foundation.\nAll Rights Reserved.\n\nCopyright (c) 2000 BeOpen.com.\nAll Rights Reserved.\n\nCopyright (c) 1995-2001 Corporation for National Research Initiatives.\nAll Rights Reserved.\n\nCopyright (c) 1991-1995 Stichting Mathematisch Centrum, Amsterdam.\nAll Rights Reserved.', _Printer__lines=None, _Printer__name='copyright', _Printer__dirs=(), _Printer__files=(...)) at remote 0x7ffff7ed8a10>, 'NameError': <type at remote 0x7ffff7d82060>, 'BytesWarning': <type at remote 0x7ffff...(truncated)) at Objects/dictobject.c:795
#7 PyDict_SetItem (
op={'__builtins__': {'bytearray': <type at remote 0x7ffff7d7c300>, 'IndexError': <type at remote 0x7ffff7d82bc0>, 'all': <built-in function all>, 'help': <_Helper at remote 0x7ffff7ed8c50>, 'vars': <built-in function vars>, 'SyntaxError': <type at remote 0x7ffff7d82540>, 'unicode': <type at remote 0x7ffff7d99040>, 'UnicodeDecodeError': <type at remote 0x7ffff7d833e0>, 'memoryview': <type at remote 0x7ffff7d8d900>, 'isinstance': <built-in function isinstance>, 'copyright': <_Printer(_Printer__data='Copyright (c) 2001-2016 Python Software Foundation.\nAll Rights Reserved.\n\nCopyright (c) 2000 BeOpen.com.\nAll Rights Reserved.\n\nCopyright (c) 1995-2001 Corporation for National Research Initiatives.\nAll Rights Reserved.\n\nCopyright (c) 1991-1995 Stichting Mathematisch Centrum, Amsterdam.\nAll Rights Reserved.', _Printer__lines=None, _Printer__name='copyright', _Printer__dirs=(), _Printer__files=(...)) at remote 0x7ffff7ed8a10>, 'NameError': <type at remote 0x7ffff7d82060>, 'BytesWarning': <type at remote 0x7ffff...(truncated), key='build_time_vars', value=None)
at Objects/dictobject.c:848
#8 0x00007ffff7a6ea8d in _PyModule_Clear (m=<optimized out>) at Objects/moduleobject.c:139
#9 0x00007ffff7aec4a1 in PyImport_Cleanup () at Python/import.c:512
#10 0x00007ffff7af957b in Py_Finalize () at Python/pythonrun.c:458
#11 0x00007ffff7b0f8e5 in Py_Main (argc=<optimized out>, argv=<optimized out>) at Modules/main.c:670
#12 0x00007ffff6d04830 in __libc_start_main (main=0x4007f0 <main>, argc=2, argv=0x7fffffffe008, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffdff8) at ../csu/libc-start.c:291
#13 0x0000000000400729 in _start ()
```
| priority | segfault in python multithreaded setting zihang dai reports and i ve reproduced that the autograd engine is not thread safe here s a repro script run it a few times occassionally it segfaults segfault is from a much different location when cleaning up imports thread python received signal sigsegv segmentation fault malloc consolidate av av entry at malloc c malloc c no such file or directory gdb where malloc consolidate av av entry at malloc c in int free av p have lock at malloc c in gi libc free mem at malloc c in dict dealloc mp at objects dictobject c in insertdict by entry mp key build time vars hash ep value at objects dictobject c in insertdict value none hash key build time vars mp at objects dictobject c dict set item by hash or entry value none ep hash key build time vars op builtins bytearray indexerror all help vars syntaxerror unicode unicodedecodeerror memoryview isinstance copyright nameerror byteswarning type at remote truncated at objects dictobject c pydict setitem op builtins bytearray indexerror all help vars syntaxerror unicode unicodedecodeerror memoryview isinstance copyright nameerror byteswarning type at remote truncated key build time vars value none at objects dictobject c in pymodule clear m at objects moduleobject c in pyimport cleanup at python import c in py finalize at python pythonrun c in py main argc argv at modules main c in libc start main main argc argv init fini rtld fini stack end at csu libc start c in start | 1 |
205,572 | 7,103,388,179 | IssuesEvent | 2018-01-16 04:40:43 | morpheus65535/bazarr | https://api.github.com/repos/morpheus65535/bazarr | closed | Option to not append language to subtitle filename | enhancement priority:medium | Bazarr appends the language of the subtitle file into it's filename - e.g. `sometitle.en.srt` - making it easy to organise multiple external subtitle tracks for the same video.
However, some DLNA servers and players only serve external subtitles if they have _exactly_ the same filename. Thus, subtitle files which have the language appended to them do not get served by these devices.
For users who are only interested in one language of subtitle for a video, it would nice to have an option not to append the language in the filename, in order to make them compatible with a wider number of players.
Alternatively, implementation of #21 would probably solve this use-case equally well. | 1.0 | Option to not append language to subtitle filename - Bazarr appends the language of the subtitle file into it's filename - e.g. `sometitle.en.srt` - making it easy to organise multiple external subtitle tracks for the same video.
However, some DLNA servers and players only serve external subtitles if they have _exactly_ the same filename. Thus, subtitle files which have the language appended to them do not get served by these devices.
For users who are only interested in one language of subtitle for a video, it would nice to have an option not to append the language in the filename, in order to make them compatible with a wider number of players.
Alternatively, implementation of #21 would probably solve this use-case equally well. | priority | option to not append language to subtitle filename bazarr appends the language of the subtitle file into it s filename e g sometitle en srt making it easy to organise multiple external subtitle tracks for the same video however some dlna servers and players only serve external subtitles if they have exactly the same filename thus subtitle files which have the language appended to them do not get served by these devices for users who are only interested in one language of subtitle for a video it would nice to have an option not to append the language in the filename in order to make them compatible with a wider number of players alternatively implementation of would probably solve this use case equally well | 1 |
605,368 | 18,734,308,142 | IssuesEvent | 2021-11-04 04:08:01 | AY2122S1-CS2103T-W13-2/tp | https://api.github.com/repos/AY2122S1-CS2103T-W13-2/tp | closed | [PE-D] Help Window does not show User Guide on subsequent attempts | type.Bug priority.Medium severity.Low | I was successful in opening the User Guide through the Help Window after keying in `help`.
I then navigated to other pages by clicking the links to your team's GitHub repository via the Help Window and closed this window.
Thereafter, I re-entered `help` into the CommandBox but was presented with the webpage I last visited before I closed the window (i.e.: your team's GitHub repo) instead of the User Guide.

<!--session: 1635494331267-38517f32-21de-436c-8367-a94de66a554e-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.High` `type.FeatureFlaw`
original: stanley-1/ped#3 | 1.0 | [PE-D] Help Window does not show User Guide on subsequent attempts - I was successful in opening the User Guide through the Help Window after keying in `help`.
I then navigated to other pages by clicking the links to your team's GitHub repository via the Help Window and closed this window.
Thereafter, I re-entered `help` into the CommandBox but was presented with the webpage I last visited before I closed the window (i.e.: your team's GitHub repo) instead of the User Guide.

<!--session: 1635494331267-38517f32-21de-436c-8367-a94de66a554e-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.High` `type.FeatureFlaw`
original: stanley-1/ped#3 | priority | help window does not show user guide on subsequent attempts i was successful in opening the user guide through the help window after keying in help i then navigated to other pages by clicking the links to your team s github repository via the help window and closed this window thereafter i re entered help into the commandbox but was presented with the webpage i last visited before i closed the window i e your team s github repo instead of the user guide labels severity high type featureflaw original stanley ped | 1 |
510,415 | 14,790,326,971 | IssuesEvent | 2021-01-12 11:52:27 | canonical-web-and-design/vanilla-framework | https://api.github.com/repos/canonical-web-and-design/vanilla-framework | closed | Proposing `p-separator` for spaced <hr> | Priority: Medium | In [many sites](https://github.com/search?q=org%3Acanonical-web-and-design+p-separator&type=code), we use an in row border to separate content in the same strip. The way we do that currently is by having a local pattern called `p-separator` which when attached to an `hr` creates a horizontal divider with a few rem padding top and bottom.
## Visual

## Context
This is used mainly on marketing sites. The style has been copied or worse inlined on the `hr` element which means it is not resonsive.
## State
None that I can think of.
## Progressive enhancement
The padding should scale with the viewport.
| 1.0 | Proposing `p-separator` for spaced <hr> - In [many sites](https://github.com/search?q=org%3Acanonical-web-and-design+p-separator&type=code), we use an in row border to separate content in the same strip. The way we do that currently is by having a local pattern called `p-separator` which when attached to an `hr` creates a horizontal divider with a few rem padding top and bottom.
## Visual

## Context
This is used mainly on marketing sites. The style has been copied or worse inlined on the `hr` element which means it is not resonsive.
## State
None that I can think of.
## Progressive enhancement
The padding should scale with the viewport.
| priority | proposing p separator for spaced in we use an in row border to separate content in the same strip the way we do that currently is by having a local pattern called p separator which when attached to an hr creates a horizontal divider with a few rem padding top and bottom visual context this is used mainly on marketing sites the style has been copied or worse inlined on the hr element which means it is not resonsive state none that i can think of progressive enhancement the padding should scale with the viewport | 1 |
303,680 | 9,309,504,192 | IssuesEvent | 2019-03-25 16:37:02 | conan-io/conan-package-tools | https://api.github.com/repos/conan-io/conan-package-tools | closed | Support custom docker args for docker run command | complex: low priority: medium stage: queue type: feature | For example: `CONAN_DOCKER_ARGS` and an equivalent constructor argument.
We've discovered an issue in my company where the default settings for Windows Containers is to allocate 1GB of RAM. When building particularly large visual studio projects, we get a very random and intermittant compiler error. This occurs when using multiple versions of Visual Studio Build Tools 15.6.7, 15.7.6, and 15.8.6, and the error references a random C++ template parameter each time, from a variety of different libraries we use.
```
error C1001: An internal error has occurred in the compiler.
compiler file 'msc1.cpp', line 1511
```
There are many reports about this error, and Microsoft explains that it's a very generic error that tells them almost nothing. For us, we've found that it doesn't happen unless we're in Docker, and only on one of our docker hosts, and only on this really big project. So, we discovered a workaround to be to pass the following additional argument to Docker:
`-m 2gb`
This doubles the RAM allocated to each container, and the problem seems to have gone away. However, it's very likely that simply applying this to all builds as part of CPT would have unexpected consequences for other users or even our other projects. Additionally, there are countless other docker flags which might be necessary to set under certain circumstances.
As a result, this request is for a primitive catch all string variable that would give users some control over the docker process they are launching, without complicating CPT too much. | 1.0 | Support custom docker args for docker run command - For example: `CONAN_DOCKER_ARGS` and an equivalent constructor argument.
We've discovered an issue in my company where the default settings for Windows Containers is to allocate 1GB of RAM. When building particularly large visual studio projects, we get a very random and intermittant compiler error. This occurs when using multiple versions of Visual Studio Build Tools 15.6.7, 15.7.6, and 15.8.6, and the error references a random C++ template parameter each time, from a variety of different libraries we use.
```
error C1001: An internal error has occurred in the compiler.
compiler file 'msc1.cpp', line 1511
```
There are many reports about this error, and Microsoft explains that it's a very generic error that tells them almost nothing. For us, we've found that it doesn't happen unless we're in Docker, and only on one of our docker hosts, and only on this really big project. So, we discovered a workaround to be to pass the following additional argument to Docker:
`-m 2gb`
This doubles the RAM allocated to each container, and the problem seems to have gone away. However, it's very likely that simply applying this to all builds as part of CPT would have unexpected consequences for other users or even our other projects. Additionally, there are countless other docker flags which might be necessary to set under certain circumstances.
As a result, this request is for a primitive catch all string variable that would give users some control over the docker process they are launching, without complicating CPT too much. | priority | support custom docker args for docker run command for example conan docker args and an equivalent constructor argument we ve discovered an issue in my company where the default settings for windows containers is to allocate of ram when building particularly large visual studio projects we get a very random and intermittant compiler error this occurs when using multiple versions of visual studio build tools and and the error references a random c template parameter each time from a variety of different libraries we use error an internal error has occurred in the compiler compiler file cpp line there are many reports about this error and microsoft explains that it s a very generic error that tells them almost nothing for us we ve found that it doesn t happen unless we re in docker and only on one of our docker hosts and only on this really big project so we discovered a workaround to be to pass the following additional argument to docker m this doubles the ram allocated to each container and the problem seems to have gone away however it s very likely that simply applying this to all builds as part of cpt would have unexpected consequences for other users or even our other projects additionally there are countless other docker flags which might be necessary to set under certain circumstances as a result this request is for a primitive catch all string variable that would give users some control over the docker process they are launching without complicating cpt too much | 1 |
279,996 | 8,676,780,459 | IssuesEvent | 2018-11-30 15:03:48 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | closed | If a user's attempt to reset his OSIO environment fails, no root cause, or suggested actions are provided | SEV3-medium area/user/tenant issue/intermittent priority/P3 team/platform train/heather type/bug | Steps to recreate:
* Reset a user's environment
The reset often fails, but a root cause for the failure is not communicated to the user. Generally, it is possible to reset the environment if a 2nd (or 3rd) attempt is made. We should present error information, and a suggestion for the user's next steps.

| 1.0 | If a user's attempt to reset his OSIO environment fails, no root cause, or suggested actions are provided - Steps to recreate:
* Reset a user's environment
The reset often fails, but a root cause for the failure is not communicated to the user. Generally, it is possible to reset the environment if a 2nd (or 3rd) attempt is made. We should present error information, and a suggestion for the user's next steps.

| priority | if a user s attempt to reset his osio environment fails no root cause or suggested actions are provided steps to recreate reset a user s environment the reset often fails but a root cause for the failure is not communicated to the user generally it is possible to reset the environment if a or attempt is made we should present error information and a suggestion for the user s next steps | 1 |
595,090 | 18,059,695,808 | IssuesEvent | 2021-09-20 12:43:52 | AY2122S1-CS2103-W14-4/tp | https://api.github.com/repos/AY2122S1-CS2103-W14-4/tp | opened | Sort groups & students by their tP / iP progress (by ascending/descending order) | type.Story priority.Medium | As a user I can sort groups & students by their tP / iP progress (by ascending/descending order) so that I can see who needs help. | 1.0 | Sort groups & students by their tP / iP progress (by ascending/descending order) - As a user I can sort groups & students by their tP / iP progress (by ascending/descending order) so that I can see who needs help. | priority | sort groups students by their tp ip progress by ascending descending order as a user i can sort groups students by their tp ip progress by ascending descending order so that i can see who needs help | 1 |
356,113 | 10,588,985,248 | IssuesEvent | 2019-10-09 04:16:33 | AY1920S1-CS2103T-T09-1/main | https://api.github.com/repos/AY1920S1-CS2103T-T09-1/main | opened | As a fast typist, I can be auto-corrected for my typos | priority.Medium type.Story | so that I won't break my train of thought while planning | 1.0 | As a fast typist, I can be auto-corrected for my typos - so that I won't break my train of thought while planning | priority | as a fast typist i can be auto corrected for my typos so that i won t break my train of thought while planning | 1 |
289,144 | 8,855,198,425 | IssuesEvent | 2019-01-09 05:15:37 | visit-dav/issues-test | https://api.github.com/repos/visit-dav/issues-test | closed | XRay Image query crashes if you give it bad variable names for the absorptivity and emissivity. | bug crash likelihood medium priority reviewed severity high wrong results | Paul Amala was using the XRay Image query and forgot to set the absorptivity and emmissivity variable names and VisIt crashed because they were wrong. The engine crashed and then the viewer was in such a state that even after restarting the engine, the engine would immediately crash again.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2310
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: XRay Image query crashes if you give it bad variable names for the absorptivity and emissivity.
Assigned to: Kevin Griffin
Category: -
Target version: 2.12.0
Author: Eric Brugger
Start: 06/24/2015
Due date:
% Done: 100%
Estimated time:
Created: 06/24/2015 03:41 pm
Updated: 10/25/2016 01:41 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.9.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Paul Amala was using the XRay Image query and forgot to set the absorptivity and emmissivity variable names and VisIt crashed because they were wrong. The engine crashed and then the viewer was in such a state that even after restarting the engine, the engine would immediately crash again.
Comments:
Hello:I added a check missing variables in the Xray Image Query and also made the error messages a little more verbose (Bug #2310). Will update the release notes once the relnotes2.11.1.html file is created.2.11RC:Sending avt/Filters/avtXRayFilter.CSending avt/Queries/Queries/avtXRayImageQuery.CSending avt/Queries/Queries/avtXRayImageQuery.hTransmitting file data ...doneCommitting transaction...Committed revision 29337.TRUNK:Sending avt/Filters/avtXRayFilter.CSending avt/Queries/Queries/avtXRayImageQuery.CSending avt/Queries/Queries/avtXRayImageQuery.hTransmitting file data ...doneCommitting transaction...Committed revision 29339.-Kevin
| 1.0 | XRay Image query crashes if you give it bad variable names for the absorptivity and emissivity. - Paul Amala was using the XRay Image query and forgot to set the absorptivity and emmissivity variable names and VisIt crashed because they were wrong. The engine crashed and then the viewer was in such a state that even after restarting the engine, the engine would immediately crash again.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2310
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: XRay Image query crashes if you give it bad variable names for the absorptivity and emissivity.
Assigned to: Kevin Griffin
Category: -
Target version: 2.12.0
Author: Eric Brugger
Start: 06/24/2015
Due date:
% Done: 100%
Estimated time:
Created: 06/24/2015 03:41 pm
Updated: 10/25/2016 01:41 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.9.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Paul Amala was using the XRay Image query and forgot to set the absorptivity and emmissivity variable names and VisIt crashed because they were wrong. The engine crashed and then the viewer was in such a state that even after restarting the engine, the engine would immediately crash again.
Comments:
Hello:I added a check missing variables in the Xray Image Query and also made the error messages a little more verbose (Bug #2310). Will update the release notes once the relnotes2.11.1.html file is created.2.11RC:Sending avt/Filters/avtXRayFilter.CSending avt/Queries/Queries/avtXRayImageQuery.CSending avt/Queries/Queries/avtXRayImageQuery.hTransmitting file data ...doneCommitting transaction...Committed revision 29337.TRUNK:Sending avt/Filters/avtXRayFilter.CSending avt/Queries/Queries/avtXRayImageQuery.CSending avt/Queries/Queries/avtXRayImageQuery.hTransmitting file data ...doneCommitting transaction...Committed revision 29339.-Kevin
| priority | xray image query crashes if you give it bad variable names for the absorptivity and emissivity paul amala was using the xray image query and forgot to set the absorptivity and emmissivity variable names and visit crashed because they were wrong the engine crashed and then the viewer was in such a state that even after restarting the engine the engine would immediately crash again redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject xray image query crashes if you give it bad variable names for the absorptivity and emissivity assigned to kevin griffin category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description paul amala was using the xray image query and forgot to set the absorptivity and emmissivity variable names and visit crashed because they were wrong the engine crashed and then the viewer was in such a state that even after restarting the engine the engine would immediately crash again comments hello i added a check missing variables in the xray image query and also made the error messages a little more verbose bug will update the release notes once the html file is created sending avt filters avtxrayfilter csending avt queries queries avtxrayimagequery csending avt queries queries avtxrayimagequery htransmitting file data donecommitting transaction committed revision trunk sending avt filters avtxrayfilter csending avt queries queries avtxrayimagequery csending avt queries queries avtxrayimagequery htransmitting file data donecommitting transaction committed revision kevin | 1 |
22,151 | 2,645,693,733 | IssuesEvent | 2015-03-13 01:10:52 | prikhi/evoluspencil | https://api.github.com/repos/prikhi/evoluspencil | closed | Please add the ability to create a new page size @ a specific size | 1 star bug imported Priority-Medium | _From [tony.sev...@gmail.com](https://code.google.com/u/111188139539534463642/) on August 25, 2008 13:57:31_
New docs seem to take on the document window size at the time of creation
but it would be good to to define the document size in specific px.
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=43_ | 1.0 | Please add the ability to create a new page size @ a specific size - _From [tony.sev...@gmail.com](https://code.google.com/u/111188139539534463642/) on August 25, 2008 13:57:31_
New docs seem to take on the document window size at the time of creation
but it would be good to to define the document size in specific px.
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=43_ | priority | please add the ability to create a new page size a specific size from on august new docs seem to take on the document window size at the time of creation but it would be good to to define the document size in specific px original issue | 1 |
109,243 | 4,383,163,184 | IssuesEvent | 2016-08-07 10:49:28 | madewithlove/laravel-cqrs-es | https://api.github.com/repos/madewithlove/laravel-cqrs-es | closed | Reset read models when doing event-store:replay | Priority: Medium Status: Completed Type: Enhancement | Currently when doing `event-store:replay` it will still keep old read models around. Ideally these should be cleared.
Since read models can be stored in different drivers we should write the logic ourselves but instead provide an API so that people can hook into these events.
My proposal is to trigger the following application events:
`event-store.replay.before`
`event-store.replay.after`
This would allow package consumers to listen for these events and executed the necessary logic.
| 1.0 | Reset read models when doing event-store:replay - Currently when doing `event-store:replay` it will still keep old read models around. Ideally these should be cleared.
Since read models can be stored in different drivers we should write the logic ourselves but instead provide an API so that people can hook into these events.
My proposal is to trigger the following application events:
`event-store.replay.before`
`event-store.replay.after`
This would allow package consumers to listen for these events and executed the necessary logic.
| priority | reset read models when doing event store replay currently when doing event store replay it will still keep old read models around ideally these should be cleared since read models can be stored in different drivers we should write the logic ourselves but instead provide an api so that people can hook into these events my proposal is to trigger the following application events event store replay before event store replay after this would allow package consumers to listen for these events and executed the necessary logic | 1 |
255,602 | 8,125,821,051 | IssuesEvent | 2018-08-16 22:29:23 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Provide non-mesa linux as well as with-mesa distros. | Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal | Our linux with-mesa distributions (started with 2.9.0) have caused problems for some users (Doug Reeder is one).
https://elist.ornl.gov/pipermail/visit-users/2015-June/017430.html
When I provided Doug with a non-mesa version, his problems went away.
Perhaps we should provide both mesa and non-mesa versions.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2312
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Provide non-mesa linux as well as with-mesa distros.
Assigned to: Eric Brugger
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 06/26/2015
Due date:
% Done: 100
Estimated time: 6.0
Created: 06/26/2015 02:04 pm
Updated: 10/07/2015 02:57 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
Our linux with-mesa distributions (started with 2.9.0) have caused problems for some users (Doug Reeder is one).
https://elist.ornl.gov/pipermail/visit-users/2015-June/017430.html
When I provided Doug with a non-mesa version, his problems went away.
Perhaps we should provide both mesa and non-mesa versions.
Comments:
I committed revisions 27281 and 27283 to the 2.10 RC and trunk with thefollowing change:1) I made a number of changes to the open build and install scripts. I removed sidious. I updated them to use the 2.10.0 version of cmake. I modified the kickit build to build a version with mesa and a version without.M svn_bin/visit-build-openM svn_bin/visit-install-openI committed revisions 27377 and 27379 to the 2.10 RC and trunk with thefollowing change:1) I rebuilt the third party libraries for 2.10 on ubuntu1104_64. I also enhanced the open build script to build both with and without mesa versions of the distribution.M config-site/ubuntu1104-64.cmakeM svn_bin/visit-build-openM svn_bin/visit-install-openI committed revisions 27419 and 27421 to the 2.10 RC and trunk with thefollowing change:1) I rebuilt the third party libraries for 2.10 on ubuntu1404_64. I also enhanced the open build script to build both with and without mesa versions of the distribution. This together with checkins 27281 and 27377 (2.10RC) and 27283 and 27379 (trunk) resolves #2312.M config-site/ubuntu1404-64.sqalab-vmware.box.cmakeM svn_bin/visit-build-openM svn_bin/visit-install-open
| 1.0 | Provide non-mesa linux as well as with-mesa distros. - Our linux with-mesa distributions (started with 2.9.0) have caused problems for some users (Doug Reeder is one).
https://elist.ornl.gov/pipermail/visit-users/2015-June/017430.html
When I provided Doug with a non-mesa version, his problems went away.
Perhaps we should provide both mesa and non-mesa versions.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2312
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Provide non-mesa linux as well as with-mesa distros.
Assigned to: Eric Brugger
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 06/26/2015
Due date:
% Done: 100
Estimated time: 6.0
Created: 06/26/2015 02:04 pm
Updated: 10/07/2015 02:57 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
Our linux with-mesa distributions (started with 2.9.0) have caused problems for some users (Doug Reeder is one).
https://elist.ornl.gov/pipermail/visit-users/2015-June/017430.html
When I provided Doug with a non-mesa version, his problems went away.
Perhaps we should provide both mesa and non-mesa versions.
Comments:
I committed revisions 27281 and 27283 to the 2.10 RC and trunk with thefollowing change:1) I made a number of changes to the open build and install scripts. I removed sidious. I updated them to use the 2.10.0 version of cmake. I modified the kickit build to build a version with mesa and a version without.M svn_bin/visit-build-openM svn_bin/visit-install-openI committed revisions 27377 and 27379 to the 2.10 RC and trunk with thefollowing change:1) I rebuilt the third party libraries for 2.10 on ubuntu1104_64. I also enhanced the open build script to build both with and without mesa versions of the distribution.M config-site/ubuntu1104-64.cmakeM svn_bin/visit-build-openM svn_bin/visit-install-openI committed revisions 27419 and 27421 to the 2.10 RC and trunk with thefollowing change:1) I rebuilt the third party libraries for 2.10 on ubuntu1404_64. I also enhanced the open build script to build both with and without mesa versions of the distribution. This together with checkins 27281 and 27377 (2.10RC) and 27283 and 27379 (trunk) resolves #2312.M config-site/ubuntu1404-64.sqalab-vmware.box.cmakeM svn_bin/visit-build-openM svn_bin/visit-install-open
| priority | provide non mesa linux as well as with mesa distros our linux with mesa distributions started with have caused problems for some users doug reeder is one when i provided doug with a non mesa version his problems went away perhaps we should provide both mesa and non mesa versions redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject provide non mesa linux as well as with mesa distros assigned to eric brugger category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood severity found in version impact medium expected use occasional os all support group any description our linux with mesa distributions started with have caused problems for some users doug reeder is one when i provided doug with a non mesa version his problems went away perhaps we should provide both mesa and non mesa versions comments i committed revisions and to the rc and trunk with thefollowing change i made a number of changes to the open build and install scripts i removed sidious i updated them to use the version of cmake i modified the kickit build to build a version with mesa and a version without m svn bin visit build openm svn bin visit install openi committed revisions and to the rc and trunk with thefollowing change i rebuilt the third party libraries for on i also enhanced the open build script to build both with and without mesa versions of the distribution m config site cmakem svn bin visit build openm svn bin visit install openi committed revisions and to the rc and trunk with thefollowing change i rebuilt the third party libraries for on i also enhanced the open build script to build both with and without mesa versions of the distribution this together with checkins and and and trunk resolves m config site sqalab vmware box cmakem svn bin visit build openm svn bin visit install open | 1 |
349,475 | 10,469,993,043 | IssuesEvent | 2019-09-23 01:11:33 | ssec/sift | https://api.github.com/repos/ssec/sift | closed | Fix numba warnings when opening SIFT | optimization priority: medium | This is likely a change in numba which shows warnings by default, but @katherinekolman noticed them. Numba is now telling us that our numba.jit functions are not as optimized as they should be. We at least want to get rid of the warnings, but making them optimized would be good too. | 1.0 | Fix numba warnings when opening SIFT - This is likely a change in numba which shows warnings by default, but @katherinekolman noticed them. Numba is now telling us that our numba.jit functions are not as optimized as they should be. We at least want to get rid of the warnings, but making them optimized would be good too. | priority | fix numba warnings when opening sift this is likely a change in numba which shows warnings by default but katherinekolman noticed them numba is now telling us that our numba jit functions are not as optimized as they should be we at least want to get rid of the warnings but making them optimized would be good too | 1 |
170,787 | 6,471,567,897 | IssuesEvent | 2017-08-17 11:59:59 | systers/powerup-android | https://api.github.com/repos/systers/powerup-android | closed | Developer needs UI elements for the Home Scenario -1 | Priority: MEDIUM Program: GSoC17 Type: Enhancement | ## Description
As a developer,
I want to see the UIs of home party scene,
so that I can code the stage Home Scenario- 1.
## Mocks
N/A
## Acceptance Criteria
-Background
## Definition of Done
-Passes user testing per acceptance criteria items.
-Featured in design meeting.
-Approved by UI team. (Dil & May)
## Estimation
| 1.0 | Developer needs UI elements for the Home Scenario -1 - ## Description
As a developer,
I want to see the UIs of home party scene,
so that I can code the stage Home Scenario- 1.
## Mocks
N/A
## Acceptance Criteria
-Background
## Definition of Done
-Passes user testing per acceptance criteria items.
-Featured in design meeting.
-Approved by UI team. (Dil & May)
## Estimation
| priority | developer needs ui elements for the home scenario description as a developer i want to see the uis of home party scene so that i can code the stage home scenario mocks n a acceptance criteria background definition of done passes user testing per acceptance criteria items featured in design meeting approved by ui team dil may estimation | 1 |
721,762 | 24,836,956,816 | IssuesEvent | 2022-10-26 09:35:23 | AY2223S1-CS2103T-T12-1/tp | https://api.github.com/repos/AY2223S1-CS2103T-T12-1/tp | closed | Allow model to access unfiltered list of students. | enhancement priority.MEDIUM type.Enhancement | Perhaps it would be better if we can move `deleteModuleClass` logic to `Model` to improve abstraction.
_Originally posted by @Berted in https://github.com/AY2223S1-CS2103T-T12-1/tp/pull/137#discussion_r995335084_
| 1.0 | Allow model to access unfiltered list of students. - Perhaps it would be better if we can move `deleteModuleClass` logic to `Model` to improve abstraction.
_Originally posted by @Berted in https://github.com/AY2223S1-CS2103T-T12-1/tp/pull/137#discussion_r995335084_
| priority | allow model to access unfiltered list of students perhaps it would be better if we can move deletemoduleclass logic to model to improve abstraction originally posted by berted in | 1 |
484,628 | 13,943,032,025 | IssuesEvent | 2020-10-22 22:09:16 | cloudfoundry-incubator/kubecf | https://api.github.com/repos/cloudfoundry-incubator/kubecf | closed | Unable to deploy KubeCF 2.5.6/7/8 | Priority: Medium Status: Validation Type: Bug | **Describe the bug**
Unable to deploy KubeCF 2.5.6/7/8 on AWS EKS.
The cf-apps-dns pod remains stuck in the pod initialization stage forever.
Describing the pod shows the following log
MountVolume.SetUp failed for volume "client-tls" : secret "var-cf-app-sd-client-tls" not found
**To Reproduce**
Use a simple overrides file for values.yml and deploy cf-operator and kubecf.
#####overrides.yaml######
`
system_domain: <mydomain>
kube:
storage_class: gp2
credentials:
cf_admin_password: <strong_pass>
uaa_admin_client_secret: <yet_another_strong_pass>
features:
autoscaler:
enabled: false
`
###########
helm3 install cf-operator \
--namespace cf-operator \
--set "global.operator.watchNamespace=kubecf" \
https://s3.amazonaws.com/cf-operators/release/helm-charts/cf-operator-6.1.15%2B0.g89a56300.tgz
helm3 install kubecf \
--namespace kubecf \
--values overrides.yaml \
./kubecf-v2.5.6.tgz
**Environment**
- KubeCF Version 2.5.6
- cf-operator 6.1.15
- AWS EKS (Kubernetes 1.17.9)
- OS Image -- Amazon Linux 2
- Container runtime -- docker://19.3.6
**Additional context**
**Please note:- KubeCF 2.2.2 deployment works and always succeeds on the same cluster.**
| 1.0 | Unable to deploy KubeCF 2.5.6/7/8 - **Describe the bug**
Unable to deploy KubeCF 2.5.6/7/8 on AWS EKS.
The cf-apps-dns pod remains stuck in the pod initialization stage forever.
Describing the pod shows the following log
MountVolume.SetUp failed for volume "client-tls" : secret "var-cf-app-sd-client-tls" not found
**To Reproduce**
Use a simple overrides file for values.yml and deploy cf-operator and kubecf.
#####overrides.yaml######
`
system_domain: <mydomain>
kube:
storage_class: gp2
credentials:
cf_admin_password: <strong_pass>
uaa_admin_client_secret: <yet_another_strong_pass>
features:
autoscaler:
enabled: false
`
###########
helm3 install cf-operator \
--namespace cf-operator \
--set "global.operator.watchNamespace=kubecf" \
https://s3.amazonaws.com/cf-operators/release/helm-charts/cf-operator-6.1.15%2B0.g89a56300.tgz
helm3 install kubecf \
--namespace kubecf \
--values overrides.yaml \
./kubecf-v2.5.6.tgz
**Environment**
- KubeCF Version 2.5.6
- cf-operator 6.1.15
- AWS EKS (Kubernetes 1.17.9)
- OS Image -- Amazon Linux 2
- Container runtime -- docker://19.3.6
**Additional context**
**Please note:- KubeCF 2.2.2 deployment works and always succeeds on the same cluster.**
| priority | unable to deploy kubecf describe the bug unable to deploy kubecf on aws eks the cf apps dns pod remains stuck in the pod initialization stage forever describing the pod shows the following log mountvolume setup failed for volume client tls secret var cf app sd client tls not found to reproduce use a simple overrides file for values yml and deploy cf operator and kubecf overrides yaml system domain kube storage class credentials cf admin password uaa admin client secret features autoscaler enabled false install cf operator namespace cf operator set global operator watchnamespace kubecf install kubecf namespace kubecf values overrides yaml kubecf tgz environment kubecf version cf operator aws eks kubernetes os image amazon linux container runtime docker additional context please note kubecf deployment works and always succeeds on the same cluster | 1 |
752,606 | 26,292,185,367 | IssuesEvent | 2023-01-08 14:54:12 | deadw00d/AROS | https://api.github.com/repos/deadw00d/AROS | opened | "Floating" option in Wanderer Prefs causes wrong icons rendering | type:bug priority:medium | "By integrating "20200413" on AROS One x86 on "WMWare SVGA", I had encountered the same problem of corrupted icon graphics as as with Scalos (see screenshot), as if by magic the problem is solved by reconfiguring from scratch Wanderer Prefs"
https://ae.amigalife.org/index.php?topic=1063.0.msg14016#msg14016 | 1.0 | "Floating" option in Wanderer Prefs causes wrong icons rendering - "By integrating "20200413" on AROS One x86 on "WMWare SVGA", I had encountered the same problem of corrupted icon graphics as as with Scalos (see screenshot), as if by magic the problem is solved by reconfiguring from scratch Wanderer Prefs"
https://ae.amigalife.org/index.php?topic=1063.0.msg14016#msg14016 | priority | floating option in wanderer prefs causes wrong icons rendering by integrating on aros one on wmware svga i had encountered the same problem of corrupted icon graphics as as with scalos see screenshot as if by magic the problem is solved by reconfiguring from scratch wanderer prefs | 1 |
113,325 | 4,546,781,798 | IssuesEvent | 2016-09-12 00:11:03 | Dolibarr/dolibarr | https://api.github.com/repos/Dolibarr/dolibarr | closed | 3.7 does not fill edit image alt attribute | Bug Priority Medium version 3.7.1 version 3.8.0-beta version develop | Title tag is filled instead.
Both attributes are similar, but the "alt" attribute is shown when the image cannot be loaded, while the title attribute is shown to the user as a complementary text to the image.
Example (from https://www.w3.org/wiki/HTML/Elements/img):
```
<img src="sales.gif"
title="Sales graph"
alt="From 1998 to 2005, sales increased by the following percentages
with each year: 624%, 75%, 138%, 40%, 35%, 9%, 21%">
```
As you can see in that example, the `alt` attribute describes the image while the `title` attribute does not.
 | 1.0 | 3.7 does not fill edit image alt attribute - Title tag is filled instead.
Both attributes are similar, but the "alt" attribute is shown when the image cannot be loaded, while the title attribute is shown to the user as a complementary text to the image.
Example (from https://www.w3.org/wiki/HTML/Elements/img):
```
<img src="sales.gif"
title="Sales graph"
alt="From 1998 to 2005, sales increased by the following percentages
with each year: 624%, 75%, 138%, 40%, 35%, 9%, 21%">
```
As you can see in that example, the `alt` attribute describes the image while the `title` attribute does not.
 | priority | does not fill edit image alt attribute title tag is filled instead both attributes are similar but the alt attribute is shown when the image cannot be loaded while the title attribute is shown to the user as a complementary text to the image example from img src sales gif title sales graph alt from to sales increased by the following percentages with each year as you can see in that example the alt attribute describes the image while the title attribute does not | 1 |
364,737 | 10,772,714,461 | IssuesEvent | 2019-11-02 16:20:17 | SOSML/SOSML | https://api.github.com/repos/SOSML/SOSML | closed | Polymorphic declarations in let broken | p5: medium priority s:elaboration t:squid | In some cases, val declarations in let expressions are (seemingly) treated as ambiguous declarations:
`val (a, b) = let
val x = fn y => y
in (x 5, x 9.0) end`
does not pass elaboration although it should (and the code does work in mosml and smlnj). | 1.0 | Polymorphic declarations in let broken - In some cases, val declarations in let expressions are (seemingly) treated as ambiguous declarations:
`val (a, b) = let
val x = fn y => y
in (x 5, x 9.0) end`
does not pass elaboration although it should (and the code does work in mosml and smlnj). | priority | polymorphic declarations in let broken in some cases val declarations in let expressions are seemingly treated as ambiguous declarations val a b let val x fn y y in x x end does not pass elaboration although it should and the code does work in mosml and smlnj | 1 |
73,103 | 3,407,210,493 | IssuesEvent | 2015-12-04 01:11:14 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Possible regression in waiting for DOMContentLoaded | area-dart2js area-html Priority-Medium | Opening a new bug for possible regression reported in https://github.com/dart-lang/sdk/issues/885#issuecomment-161689055
@vsmenon did the original fix.
Comment from there by @joseluis is
On version 1.13.0 this doesn't seem true anymore.
Minimal example: querySelector('#someid'); returns null when the element is placed after the script tag that loads the Dart script compiled with dart2js. In order to make it work I must use the script defer attribute, like this:
```html
<html>
<head>
<meta charset="utf-8">
</head>
<body>
<script src="dart-program.js" defer></script>
<script id="someid"></script>
</body>
</html>
``` | 1.0 | Possible regression in waiting for DOMContentLoaded - Opening a new bug for possible regression reported in https://github.com/dart-lang/sdk/issues/885#issuecomment-161689055
@vsmenon did the original fix.
Comment from there by @joseluis is
On version 1.13.0 this doesn't seem true anymore.
Minimal example: querySelector('#someid'); returns null when the element is placed after the script tag that loads the Dart script compiled with dart2js. In order to make it work I must use the script defer attribute, like this:
```html
<html>
<head>
<meta charset="utf-8">
</head>
<body>
<script src="dart-program.js" defer></script>
<script id="someid"></script>
</body>
</html>
``` | priority | possible regression in waiting for domcontentloaded opening a new bug for possible regression reported in vsmenon did the original fix comment from there by joseluis is on version this doesn t seem true anymore minimal example queryselector someid returns null when the element is placed after the script tag that loads the dart script compiled with in order to make it work i must use the script defer attribute like this html | 1 |
265,651 | 8,357,298,275 | IssuesEvent | 2018-10-02 21:06:52 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | tests/kernel/fifo/fifo_timeout fails on nrf51_pca10028 and nrf52_pca10040 | bug in progress nRF priority: medium | Looks like this bug has been failing on this platform since the test was introduced
```
***** delaying boot 1000ms (per build configuration) *****
***** Booting Zephyr OS 1.12.0-rc2 (delayed boot 1000ms) *****
Running test suite test_fifo_timeout
===================================================================
starting test - test_timeout_empty_fifo
PASS - test_timeout_empty_fifo
===================================================================
starting test - test_timeout_non_empty_fifo
PASS - test_timeout_non_empty_fifo
===================================================================
starting test - test_timeout_fifo_thread
PASS - test_timeout_fifo_thread
===================================================================
starting test - test_timeout_threads_pend_on_fifo
thread (q order: 2, t/o: 0, fifo 0x20000000)
thread (q order: 3, t/o: 10, fifo 0x20000000)
thread (q order: 0, t/o: 20, fifo 0x20000000)
thread (q order: 4, t/o: 30, fifo 0x20000000)
thread (q order: 1, t/o: 40, fifo 0x20000000)
PASS - test_timeout_threads_pend_on_fifo
===================================================================
starting test - test_timeout_threads_pend_on_dual_fifos
thread (q order: 0, t/o: 0, fifo 0x20000010)
thread (q order: 5, t/o: 10, fifo 0x20000000)
FAIL - test_multiple_threads_pending@156. *** thread 3 woke up, expected 2
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:386: test_timeout_threads_pend_on_dual_fifos: rv not equal to TC_PASS
FAIL - test_timeout_threads_pend_on_dual_fifos
===================================================================
starting test - test_timeout_threads_pend_fail_on_fifo
FAIL - test_multiple_threads_get_data@207. *** thread 7 woke up, expected 0
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:401: test_timeout_threads_pend_fail_on_fifo: rv not equal to TC_PASS
FAIL - test_timeout_threads_pend_fail_on_fifo
===================================================================
===================================================================
RunID: ci-180601-1931-1716:k8if
PROJECT EXECUTION FAILED
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
```
This test also fails with "prj_poll.conf" that is CONFIG_POLL=y in the latest commit: 5b8e4ae4df378d0da32ce36266246fb530a6309d
```
***** delaying boot 1000ms (per build configuration) *****
***** Booting Zephyr OS v1.12.0-831-g5b8e4ae (delayed boot 1000ms) *****
Running test suite test_fifo_timeout
===================================================================
starting test - test_timeout_empty_fifo
PASS - test_timeout_empty_fifo
===================================================================
starting test - test_timeout_non_empty_fifo
PASS - test_timeout_non_empty_fifo
===================================================================
starting test - test_timeout_fifo_thread
PASS - test_timeout_fifo_thread
===================================================================
starting test - test_timeout_threads_pend_on_fifo
thread (q order: 2, t/o: 0, fifo 0x20000000)
thread (q order: 3, t/o: 10, fifo 0x20000000)
thread (q order: 0, t/o: 20, fifo 0x20000000)
thread (q order: 4, t/o: 30, fifo 0x20000000)
thread (q order: 1, t/o: 40, fifo 0x20000000)
PASS - test_timeout_threads_pend_on_fifo
===================================================================
starting test - test_timeout_threads_pend_on_dual_fifos
thread (q order: 0, t/o: 0, fifo 0x20000010)
thread (q order: 5, t/o: 10, fifo 0x20000000)
FAIL - test_multiple_threads_pending@156. *** thread 3 woke up, expected 2
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:396: test_timeout_threads_pend_on_dual_fifos: rv not equal to TC_PASS
FAIL - test_timeout_threads_pend_on_dual_fifos
===================================================================
starting test - test_timeout_threads_pend_fail_on_fifo
FAIL - test_multiple_threads_get_data@207. *** thread 7 woke up, expected 0
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:411: test_timeout_threads_pend_fail_on_fifo: rv not equal to TC_PASS
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:129: test_thread_pend_and_timeout: packet == NULL is false
FAIL - test_timeout_threads_pend_fail_on_fifo
===================================================================
===================================================================
RunID: :wjv2
PROJECT EXECUTION FAILED
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
```
This failure(with CONFIG_POLL=y) was not seen in earlier commits. | 1.0 | tests/kernel/fifo/fifo_timeout fails on nrf51_pca10028 and nrf52_pca10040 - Looks like this bug has been failing on this platform since the test was introduced
```
***** delaying boot 1000ms (per build configuration) *****
***** Booting Zephyr OS 1.12.0-rc2 (delayed boot 1000ms) *****
Running test suite test_fifo_timeout
===================================================================
starting test - test_timeout_empty_fifo
PASS - test_timeout_empty_fifo
===================================================================
starting test - test_timeout_non_empty_fifo
PASS - test_timeout_non_empty_fifo
===================================================================
starting test - test_timeout_fifo_thread
PASS - test_timeout_fifo_thread
===================================================================
starting test - test_timeout_threads_pend_on_fifo
thread (q order: 2, t/o: 0, fifo 0x20000000)
thread (q order: 3, t/o: 10, fifo 0x20000000)
thread (q order: 0, t/o: 20, fifo 0x20000000)
thread (q order: 4, t/o: 30, fifo 0x20000000)
thread (q order: 1, t/o: 40, fifo 0x20000000)
PASS - test_timeout_threads_pend_on_fifo
===================================================================
starting test - test_timeout_threads_pend_on_dual_fifos
thread (q order: 0, t/o: 0, fifo 0x20000010)
thread (q order: 5, t/o: 10, fifo 0x20000000)
FAIL - test_multiple_threads_pending@156. *** thread 3 woke up, expected 2
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:386: test_timeout_threads_pend_on_dual_fifos: rv not equal to TC_PASS
FAIL - test_timeout_threads_pend_on_dual_fifos
===================================================================
starting test - test_timeout_threads_pend_fail_on_fifo
FAIL - test_multiple_threads_get_data@207. *** thread 7 woke up, expected 0
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:401: test_timeout_threads_pend_fail_on_fifo: rv not equal to TC_PASS
FAIL - test_timeout_threads_pend_fail_on_fifo
===================================================================
===================================================================
RunID: ci-180601-1931-1716:k8if
PROJECT EXECUTION FAILED
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
```
This test also fails with "prj_poll.conf" that is CONFIG_POLL=y in the latest commit: 5b8e4ae4df378d0da32ce36266246fb530a6309d
```
***** delaying boot 1000ms (per build configuration) *****
***** Booting Zephyr OS v1.12.0-831-g5b8e4ae (delayed boot 1000ms) *****
Running test suite test_fifo_timeout
===================================================================
starting test - test_timeout_empty_fifo
PASS - test_timeout_empty_fifo
===================================================================
starting test - test_timeout_non_empty_fifo
PASS - test_timeout_non_empty_fifo
===================================================================
starting test - test_timeout_fifo_thread
PASS - test_timeout_fifo_thread
===================================================================
starting test - test_timeout_threads_pend_on_fifo
thread (q order: 2, t/o: 0, fifo 0x20000000)
thread (q order: 3, t/o: 10, fifo 0x20000000)
thread (q order: 0, t/o: 20, fifo 0x20000000)
thread (q order: 4, t/o: 30, fifo 0x20000000)
thread (q order: 1, t/o: 40, fifo 0x20000000)
PASS - test_timeout_threads_pend_on_fifo
===================================================================
starting test - test_timeout_threads_pend_on_dual_fifos
thread (q order: 0, t/o: 0, fifo 0x20000010)
thread (q order: 5, t/o: 10, fifo 0x20000000)
FAIL - test_multiple_threads_pending@156. *** thread 3 woke up, expected 2
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:396: test_timeout_threads_pend_on_dual_fifos: rv not equal to TC_PASS
FAIL - test_timeout_threads_pend_on_dual_fifos
===================================================================
starting test - test_timeout_threads_pend_fail_on_fifo
FAIL - test_multiple_threads_get_data@207. *** thread 7 woke up, expected 0
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:411: test_timeout_threads_pend_fail_on_fifo: rv not equal to TC_PASS
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:129: test_thread_pend_and_timeout: packet == NULL is false
FAIL - test_timeout_threads_pend_fail_on_fifo
===================================================================
===================================================================
RunID: :wjv2
PROJECT EXECUTION FAILED
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false
```
This failure(with CONFIG_POLL=y) was not seen in earlier commits. | priority | tests kernel fifo fifo timeout fails on and looks like this bug has been failing on this platform since the test was introduced delaying boot per build configuration booting zephyr os delayed boot running test suite test fifo timeout starting test test timeout empty fifo pass test timeout empty fifo starting test test timeout non empty fifo pass test timeout non empty fifo starting test test timeout fifo thread pass test timeout fifo thread starting test test timeout threads pend on fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo pass test timeout threads pend on fifo starting test test timeout threads pend on dual fifos thread q order t o fifo thread q order t o fifo fail test multiple threads pending thread woke up expected assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test timeout threads pend on dual fifos rv not equal to tc pass fail test timeout threads pend on dual fifos starting test test timeout threads pend fail on fifo fail test multiple threads get data thread woke up expected assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test timeout threads pend fail on fifo rv not equal to tc pass fail test timeout threads pend fail on fifo runid ci project execution failed assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false this test also fails with prj poll conf that is config poll y in the latest commit delaying boot per build configuration booting zephyr os delayed boot running test suite test fifo timeout starting test test timeout empty fifo pass test timeout empty fifo starting test test timeout non empty fifo pass test timeout non empty fifo starting test test timeout fifo thread pass test timeout fifo thread starting test test timeout threads pend on fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo pass test timeout threads pend on fifo starting test test timeout threads pend on dual fifos thread q order t o fifo thread q order t o fifo fail test multiple threads pending thread woke up expected assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test timeout threads pend on dual fifos rv not equal to tc pass fail test timeout threads pend on dual fifos starting test test timeout threads pend fail on fifo fail test multiple threads get data thread woke up expected assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test timeout threads pend fail on fifo rv not equal to tc pass assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and timeout packet null is false fail test timeout threads pend fail on fifo runid project execution failed assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false this failure with config poll y was not seen in earlier commits | 1 |
597,787 | 18,171,823,838 | IssuesEvent | 2021-09-27 20:58:16 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | Clean up react warnings | type:task priority:medium component:ui state:needs_review | ##### SUMMARY
Clean up react warnings. Go to console.
See:
```
Warning: Can't perform a React state update on an unmounted component. This is a no-op, but it indicates a memory leak in your application. To fix, cancel all subscriptions and asynchronous tasks in a useEffect cleanup function.
in InventorySourceAdd (at InventorySources.js:12)
in Route (at InventorySources.js:11) index.js:1```
```
```
Warning: React does not recognize the `isValid` prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase `isvalid` instead. If you accidentally passed it from a parent component, remove it from the DOM element.
in ul (created by Context.Consumer)
in SelectMenu (created by ForwardRef)
in ForwardRef (created by Select)
in div (created by Select)
in GenerateId (created by Select)
in Select (at SCMSubForm.js:113)
in div (created by FormGroup)
in div (created by FormGroup)
in FormGroup (at SCMSubForm.js:94)
in SCMSubForm (at InventorySourceForm.js:204)
in div (created by styled.div)
in styled.div (at InventorySourceForm.js:155)
in div (created by styled.div)
in styled.div (at InventorySourceForm.js:151)
in InventorySourceFormFields (at InventorySourceForm.js:298)
in div (created by styled.div)
in styled.div (at InventorySourceForm.js:297)
```
One more - Logged out and see
```
Warning: Cannot update a component (`SessionProvider`) while rendering a different component (`ProtectedRoute`). To locate the bad setState() call inside `ProtectedRoute`, follow the stack trace as described in https://fb.me/setstate-in-render
in ProtectedRoute (at App.js:166)
in Switch (at App.js:156)
in SessionProvider (at App.js:155)
in Unknown (at App.js:154)
in I18nProvider (at App.js:153)
in App (at App.js:180)
in Router (created by HashRouter)
in HashRouter (at App.js:179)
in Unknown (at src/index.js:10)
in StrictMode (at src/index.js:9)
``` | 1.0 | Clean up react warnings - ##### SUMMARY
Clean up react warnings. Go to console.
See:
```
Warning: Can't perform a React state update on an unmounted component. This is a no-op, but it indicates a memory leak in your application. To fix, cancel all subscriptions and asynchronous tasks in a useEffect cleanup function.
in InventorySourceAdd (at InventorySources.js:12)
in Route (at InventorySources.js:11) index.js:1```
```
```
Warning: React does not recognize the `isValid` prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase `isvalid` instead. If you accidentally passed it from a parent component, remove it from the DOM element.
in ul (created by Context.Consumer)
in SelectMenu (created by ForwardRef)
in ForwardRef (created by Select)
in div (created by Select)
in GenerateId (created by Select)
in Select (at SCMSubForm.js:113)
in div (created by FormGroup)
in div (created by FormGroup)
in FormGroup (at SCMSubForm.js:94)
in SCMSubForm (at InventorySourceForm.js:204)
in div (created by styled.div)
in styled.div (at InventorySourceForm.js:155)
in div (created by styled.div)
in styled.div (at InventorySourceForm.js:151)
in InventorySourceFormFields (at InventorySourceForm.js:298)
in div (created by styled.div)
in styled.div (at InventorySourceForm.js:297)
```
One more - Logged out and see
```
Warning: Cannot update a component (`SessionProvider`) while rendering a different component (`ProtectedRoute`). To locate the bad setState() call inside `ProtectedRoute`, follow the stack trace as described in https://fb.me/setstate-in-render
in ProtectedRoute (at App.js:166)
in Switch (at App.js:156)
in SessionProvider (at App.js:155)
in Unknown (at App.js:154)
in I18nProvider (at App.js:153)
in App (at App.js:180)
in Router (created by HashRouter)
in HashRouter (at App.js:179)
in Unknown (at src/index.js:10)
in StrictMode (at src/index.js:9)
``` | priority | clean up react warnings summary clean up react warnings go to console see warning can t perform a react state update on an unmounted component this is a no op but it indicates a memory leak in your application to fix cancel all subscriptions and asynchronous tasks in a useeffect cleanup function in inventorysourceadd at inventorysources js in route at inventorysources js index js warning react does not recognize the isvalid prop on a dom element if you intentionally want it to appear in the dom as a custom attribute spell it as lowercase isvalid instead if you accidentally passed it from a parent component remove it from the dom element in ul created by context consumer in selectmenu created by forwardref in forwardref created by select in div created by select in generateid created by select in select at scmsubform js in div created by formgroup in div created by formgroup in formgroup at scmsubform js in scmsubform at inventorysourceform js in div created by styled div in styled div at inventorysourceform js in div created by styled div in styled div at inventorysourceform js in inventorysourceformfields at inventorysourceform js in div created by styled div in styled div at inventorysourceform js one more logged out and see warning cannot update a component sessionprovider while rendering a different component protectedroute to locate the bad setstate call inside protectedroute follow the stack trace as described in in protectedroute at app js in switch at app js in sessionprovider at app js in unknown at app js in at app js in app at app js in router created by hashrouter in hashrouter at app js in unknown at src index js in strictmode at src index js | 1 |
69,839 | 3,315,861,921 | IssuesEvent | 2015-11-06 14:28:24 | craftjk/issue-migration-test4 | https://api.github.com/repos/craftjk/issue-migration-test4 | opened | Visualize the Hierarchical Type IIs Assembly Tree Process | Priority: Medium Type: Feature | <a href="https://github.com/KeeganW"><img src="https://avatars.githubusercontent.com/u/7226822?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [KeeganW](https://github.com/KeeganW)**
_Tuesday Jun 10, 2014 at 23:44 GMT_
_Originally opened as https://github.com/TeselaGen/ve/issues/80_
----
Tim Thimmaiah commented May 09, 2013 10:52 (edited May 09, 2013 11:04)
Along with accommodating hierarchical design we will build the corresponding UI tool for the visualization of the assembly tree.
RavenCAD + Pidgeon are doing this fairly well right now. We'd want the same thing conceptually, but do it better and more polished.
| 1.0 | Visualize the Hierarchical Type IIs Assembly Tree Process - <a href="https://github.com/KeeganW"><img src="https://avatars.githubusercontent.com/u/7226822?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [KeeganW](https://github.com/KeeganW)**
_Tuesday Jun 10, 2014 at 23:44 GMT_
_Originally opened as https://github.com/TeselaGen/ve/issues/80_
----
Tim Thimmaiah commented May 09, 2013 10:52 (edited May 09, 2013 11:04)
Along with accommodating hierarchical design we will build the corresponding UI tool for the visualization of the assembly tree.
RavenCAD + Pidgeon are doing this fairly well right now. We'd want the same thing conceptually, but do it better and more polished.
| priority | visualize the hierarchical type iis assembly tree process issue by tuesday jun at gmt originally opened as tim thimmaiah commented may edited may along with accommodating hierarchical design we will build the corresponding ui tool for the visualization of the assembly tree ravencad pidgeon are doing this fairly well right now we d want the same thing conceptually but do it better and more polished | 1 |
780,327 | 27,390,004,694 | IssuesEvent | 2023-02-28 15:41:57 | ansible-collections/azure | https://api.github.com/repos/ansible-collections/azure | closed | azure_rm dynamic inventory error: [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: name 'azure_cloud' is not defined | inventory medium_priority work in | I've searched all over the internet and cannot find a solution to this problem.
OS: Ubuntu 22.04.2
$ ansible --version
ansible [core 2.14.2]
config file = /home/nwmadmin/.ansible.cfg
configured module search path = ['/home/nwmadmin/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python3/dist-packages/ansible
ansible collection location = /home/nwmadmin/.ansible/collections:/usr/share/ansible/collections
executable location = /usr/bin/ansible
python version = 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] (/usr/bin/python3)
jinja version = 3.0.3
libyaml = True
$ python3 --version
Python 3.10.6
I'm trying to get the dynamic inventory for my Azure subscription working.
Per the Azure documentation, I've created the file azure_rm.yaml with the following contents:
plugin: azure.azcollection.azure_rm
include_vm_resource_groups:
- '*'
auth_source: auto
Using the command:
$ ansible-inventory -i azure_rm.yaml --graph -vvv
ansible-inventory [core 2.14.2]
config file = /home/nwmadmin/.ansible.cfg
configured module search path = ['/home/nwmadmin/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python3/dist-packages/ansible
ansible collection location = /home/nwmadmin/.ansible/collections:/usr/share/ansible/collections
executable location = /usr/bin/ansible-inventory
python version = 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] (/usr/bin/python3)
jinja version = 3.0.3
libyaml = True
Using /home/nwmadmin/.ansible.cfg as config file
redirecting (type: inventory) ansible.builtin.azure_rm to azure.azcollection.azure_rm
host_list declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method
script declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method
Using inventory plugin 'ansible_collections.azure.azcollection.plugins.inventory.azure_rm' to process inventory source '/home/nwmadmin/AWX/azure_rm.yaml'
toml declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: name 'azure_cloud' is not defined
File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source
plugin.parse(self._inventory, self._loader, source, cache=cache)
File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/auto.py", line 59, in parse
plugin.parse(inventory, loader, path, cache=cache)
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 220, in parse
self._credential_setup()
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 241, in _credential_setup
self.azure_auth = AzureRMAuth(**auth_options)
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py", line 1496, in __init__
self._cloud_environment = azure_cloud.AZURE_PUBLIC_CLOUD # SDK default
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with yaml plugin: Plugin configuration YAML file, not YAML inventory
File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source
plugin.parse(self._inventory, self._loader, source, cache=cache)
File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/yaml.py", line 114, in parse
raise AnsibleParserError('Plugin configuration YAML file, not YAML inventory')
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with ini plugin: Invalid host pattern 'plugin:' supplied, ending in ':' is not allowed, this character is
reserved to provide a port.
File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source
plugin.parse(self._inventory, self._loader, source, cache=cache)
File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/ini.py", line 137, in parse
raise AnsibleParserError(e)
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with ansible_collections.azure.azcollection.plugins.inventory.azure_rm plugin: name 'azure_cloud' is not
defined
File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source
plugin.parse(self._inventory, self._loader, source, cache=cache)
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 220, in parse
self._credential_setup()
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 241, in _credential_setup
self.azure_auth = AzureRMAuth(**auth_options)
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py", line 1496, in __init__
self._cloud_environment = azure_cloud.AZURE_PUBLIC_CLOUD # SDK default
[WARNING]: Unable to parse /home/nwmadmin/AWX/azure_rm.yaml as an inventory source
[WARNING]: No inventory was parsed, only implicit localhost is available
@all:
|--@ungrouped:
I have also tried using a ~/.azure/credentials file with the <value> replaced with my correct identifiers:
[default]
subscription_id=<value>
client_id=<value>
secret=<value>
tenant=<value>
I then change the azure_rm.yaml file to use auth_source: credential_file
The output errors are identical to what is shown above.
I then try first logging into the azure subscription with:
$ az login -u myusername
$ az account show
$ az account show
{
"environmentName": "AzureCloud",
"homeTenantId": "<value>",
"id": "<value>",
"isDefault": true,
"managedByTenants": [],
"name": "<value>",
"state": "Enabled",
"tenantId": "<value>",
"user": {
"name": "<myusername>",
"type": "user"
}
}
If I replace the auth_source in my azure_rm.yaml file with: auth_source: cli
and then run the command:
$ ansible-inventory -i azure_rm.yaml --graph -vvv
The warning for the auto plugin is:
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: Failed to import the required Python library (azure-cli) on ansibleautomationvm's Python
/usr/bin/python3. This is required for `cli` auth_source. Please read the module documentation and install it in the appropriate location. If the required library is
installed, but Ansible is using the wrong Python interpreter, please consult the documentation on ansible_python_interpreter
I've read the module documentation located here:
https://docs.ansible.com/ansible/latest/collections/azure/azcollection/azure_rm_inventory.html
I see no reference in that documentation on how to or where to install the (azure-cli) python library.
| 1.0 | azure_rm dynamic inventory error: [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: name 'azure_cloud' is not defined - I've searched all over the internet and cannot find a solution to this problem.
OS: Ubuntu 22.04.2
$ ansible --version
ansible [core 2.14.2]
config file = /home/nwmadmin/.ansible.cfg
configured module search path = ['/home/nwmadmin/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python3/dist-packages/ansible
ansible collection location = /home/nwmadmin/.ansible/collections:/usr/share/ansible/collections
executable location = /usr/bin/ansible
python version = 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] (/usr/bin/python3)
jinja version = 3.0.3
libyaml = True
$ python3 --version
Python 3.10.6
I'm trying to get the dynamic inventory for my Azure subscription working.
Per the Azure documentation, I've created the file azure_rm.yaml with the following contents:
plugin: azure.azcollection.azure_rm
include_vm_resource_groups:
- '*'
auth_source: auto
Using the command:
$ ansible-inventory -i azure_rm.yaml --graph -vvv
ansible-inventory [core 2.14.2]
config file = /home/nwmadmin/.ansible.cfg
configured module search path = ['/home/nwmadmin/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python3/dist-packages/ansible
ansible collection location = /home/nwmadmin/.ansible/collections:/usr/share/ansible/collections
executable location = /usr/bin/ansible-inventory
python version = 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] (/usr/bin/python3)
jinja version = 3.0.3
libyaml = True
Using /home/nwmadmin/.ansible.cfg as config file
redirecting (type: inventory) ansible.builtin.azure_rm to azure.azcollection.azure_rm
host_list declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method
script declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method
Using inventory plugin 'ansible_collections.azure.azcollection.plugins.inventory.azure_rm' to process inventory source '/home/nwmadmin/AWX/azure_rm.yaml'
toml declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: name 'azure_cloud' is not defined
File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source
plugin.parse(self._inventory, self._loader, source, cache=cache)
File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/auto.py", line 59, in parse
plugin.parse(inventory, loader, path, cache=cache)
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 220, in parse
self._credential_setup()
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 241, in _credential_setup
self.azure_auth = AzureRMAuth(**auth_options)
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py", line 1496, in __init__
self._cloud_environment = azure_cloud.AZURE_PUBLIC_CLOUD # SDK default
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with yaml plugin: Plugin configuration YAML file, not YAML inventory
File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source
plugin.parse(self._inventory, self._loader, source, cache=cache)
File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/yaml.py", line 114, in parse
raise AnsibleParserError('Plugin configuration YAML file, not YAML inventory')
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with ini plugin: Invalid host pattern 'plugin:' supplied, ending in ':' is not allowed, this character is
reserved to provide a port.
File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source
plugin.parse(self._inventory, self._loader, source, cache=cache)
File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/ini.py", line 137, in parse
raise AnsibleParserError(e)
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with ansible_collections.azure.azcollection.plugins.inventory.azure_rm plugin: name 'azure_cloud' is not
defined
File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source
plugin.parse(self._inventory, self._loader, source, cache=cache)
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 220, in parse
self._credential_setup()
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 241, in _credential_setup
self.azure_auth = AzureRMAuth(**auth_options)
File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py", line 1496, in __init__
self._cloud_environment = azure_cloud.AZURE_PUBLIC_CLOUD # SDK default
[WARNING]: Unable to parse /home/nwmadmin/AWX/azure_rm.yaml as an inventory source
[WARNING]: No inventory was parsed, only implicit localhost is available
@all:
|--@ungrouped:
I have also tried using a ~/.azure/credentials file with the <value> replaced with my correct identifiers:
[default]
subscription_id=<value>
client_id=<value>
secret=<value>
tenant=<value>
I then change the azure_rm.yaml file to use auth_source: credential_file
The output errors are identical to what is shown above.
I then try first logging into the azure subscription with:
$ az login -u myusername
$ az account show
$ az account show
{
"environmentName": "AzureCloud",
"homeTenantId": "<value>",
"id": "<value>",
"isDefault": true,
"managedByTenants": [],
"name": "<value>",
"state": "Enabled",
"tenantId": "<value>",
"user": {
"name": "<myusername>",
"type": "user"
}
}
If I replace the auth_source in my azure_rm.yaml file with: auth_source: cli
and then run the command:
$ ansible-inventory -i azure_rm.yaml --graph -vvv
The warning for the auto plugin is:
[WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: Failed to import the required Python library (azure-cli) on ansibleautomationvm's Python
/usr/bin/python3. This is required for `cli` auth_source. Please read the module documentation and install it in the appropriate location. If the required library is
installed, but Ansible is using the wrong Python interpreter, please consult the documentation on ansible_python_interpreter
I've read the module documentation located here:
https://docs.ansible.com/ansible/latest/collections/azure/azcollection/azure_rm_inventory.html
I see no reference in that documentation on how to or where to install the (azure-cli) python library.
| priority | azure rm dynamic inventory error failed to parse home nwmadmin awx azure rm yaml with auto plugin name azure cloud is not defined i ve searched all over the internet and cannot find a solution to this problem os ubuntu ansible version ansible config file home nwmadmin ansible cfg configured module search path ansible python module location usr lib dist packages ansible ansible collection location home nwmadmin ansible collections usr share ansible collections executable location usr bin ansible python version main nov usr bin jinja version libyaml true version python i m trying to get the dynamic inventory for my azure subscription working per the azure documentation i ve created the file azure rm yaml with the following contents plugin azure azcollection azure rm include vm resource groups auth source auto using the command ansible inventory i azure rm yaml graph vvv ansible inventory config file home nwmadmin ansible cfg configured module search path ansible python module location usr lib dist packages ansible ansible collection location home nwmadmin ansible collections usr share ansible collections executable location usr bin ansible inventory python version main nov usr bin jinja version libyaml true using home nwmadmin ansible cfg as config file redirecting type inventory ansible builtin azure rm to azure azcollection azure rm host list declined parsing home nwmadmin awx azure rm yaml as it did not pass its verify file method script declined parsing home nwmadmin awx azure rm yaml as it did not pass its verify file method using inventory plugin ansible collections azure azcollection plugins inventory azure rm to process inventory source home nwmadmin awx azure rm yaml toml declined parsing home nwmadmin awx azure rm yaml as it did not pass its verify file method failed to parse home nwmadmin awx azure rm yaml with auto plugin name azure cloud is not defined file usr lib dist packages ansible inventory manager py line in parse source plugin parse self inventory self loader source cache cache file usr lib dist packages ansible plugins inventory auto py line in parse plugin parse inventory loader path cache cache file usr lib dist packages ansible collections azure azcollection plugins inventory azure rm py line in parse self credential setup file usr lib dist packages ansible collections azure azcollection plugins inventory azure rm py line in credential setup self azure auth azurermauth auth options file usr lib dist packages ansible collections azure azcollection plugins module utils azure rm common py line in init self cloud environment azure cloud azure public cloud sdk default failed to parse home nwmadmin awx azure rm yaml with yaml plugin plugin configuration yaml file not yaml inventory file usr lib dist packages ansible inventory manager py line in parse source plugin parse self inventory self loader source cache cache file usr lib dist packages ansible plugins inventory yaml py line in parse raise ansibleparsererror plugin configuration yaml file not yaml inventory failed to parse home nwmadmin awx azure rm yaml with ini plugin invalid host pattern plugin supplied ending in is not allowed this character is reserved to provide a port file usr lib dist packages ansible inventory manager py line in parse source plugin parse self inventory self loader source cache cache file usr lib dist packages ansible plugins inventory ini py line in parse raise ansibleparsererror e failed to parse home nwmadmin awx azure rm yaml with ansible collections azure azcollection plugins inventory azure rm plugin name azure cloud is not defined file usr lib dist packages ansible inventory manager py line in parse source plugin parse self inventory self loader source cache cache file usr lib dist packages ansible collections azure azcollection plugins inventory azure rm py line in parse self credential setup file usr lib dist packages ansible collections azure azcollection plugins inventory azure rm py line in credential setup self azure auth azurermauth auth options file usr lib dist packages ansible collections azure azcollection plugins module utils azure rm common py line in init self cloud environment azure cloud azure public cloud sdk default unable to parse home nwmadmin awx azure rm yaml as an inventory source no inventory was parsed only implicit localhost is available all ungrouped i have also tried using a azure credentials file with the replaced with my correct identifiers subscription id client id secret tenant i then change the azure rm yaml file to use auth source credential file the output errors are identical to what is shown above i then try first logging into the azure subscription with az login u myusername az account show az account show environmentname azurecloud hometenantid id isdefault true managedbytenants name state enabled tenantid user name type user if i replace the auth source in my azure rm yaml file with auth source cli and then run the command ansible inventory i azure rm yaml graph vvv the warning for the auto plugin is failed to parse home nwmadmin awx azure rm yaml with auto plugin failed to import the required python library azure cli on ansibleautomationvm s python usr bin this is required for cli auth source please read the module documentation and install it in the appropriate location if the required library is installed but ansible is using the wrong python interpreter please consult the documentation on ansible python interpreter i ve read the module documentation located here i see no reference in that documentation on how to or where to install the azure cli python library | 1 |
771,171 | 27,073,055,748 | IssuesEvent | 2023-02-14 08:38:56 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Setting a Poetry constraint causes lockfile update to fail | type:bug priority-3-medium manager:poetry status:in-progress | ### How are you running Renovate?
Mend Renovate hosted app on github.com
### If you're self-hosting Renovate, tell us what version of Renovate you run.
Present on self-hosted and github app
### If you're self-hosting Renovate, select which platform you are using.
gitlab.com
### If you're self-hosting Renovate, tell us what version of the platform you run.
Also observed on gitlab.com 15.8
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
When setting a python and poetry constraint, the poetry lock update command fails because renovate attempts to pip install poetry using the wrong format for pip version constraints. See the repro here: https://github.com/WarpRat/renovate-poetry-contraint-repro/pull/6
### Relevant debug logs
<details><summary>Logs</summary>
```
DEBUG: Failed to update poetry.lock file(branch="renovate/pylint-2.x")
{
"err": {
"name": "ExecError",
"cmd": "/bin/sh -c docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"",
"stderr": "ERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n",
"stdout": "installing v2 tool python v3.9.9\nlinking tool python v3.9.9\nPython 3.9.9\npip 22.3.1 from /opt/buildpack/tools/python/3.9.9/lib/python3.9/site-packages/pip (python 3.9)\nInstalled v2 /usr/local/buildpack/tools/v2/python.sh in 24 seconds\nskip cleanup, not a docker build: 8d08c5ae1170\n",
"options": {
"cwd": "/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro",
"encoding": "utf-8",
"env": {
"PIP_CACHE_DIR": "/tmp/renovate-cache/others/pip",
"HOME": "/home/ubuntu",
"PATH": "/home/ubuntu/.local/bin:/home/ubuntu/bin:/opt/buildpack/tools/python/3.9.3/bin:/home/ubuntu/.npm-global/bin:/home/ubuntu/renovateapp/node_modules/.bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",
"LC_ALL": "C.UTF-8",
"LANG": "C.UTF-8",
"BUILDPACK_CACHE_DIR": "/tmp/containerbase",
"CONTAINERBASE_CACHE_DIR": "/tmp/containerbase"
},
"maxBuffer": 10485760,
"timeout": 900000
},
"exitCode": 1,
"message": "Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"\nERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n",
"stack": "ExecError: Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"\nERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n\n at ChildProcess.<anonymous> (/home/ubuntu/renovateapp/node_modules/renovate/dist/util/exec/common.js:87:24)\n at ChildProcess.emit (node:events:525:35)\n at ChildProcess.emit (node:domain:489:12)\n at Process.ChildProcess._handle.onexit (node:internal/child_process:293:12)"
}
}
```
</details>
### Have you created a minimal reproduction repository?
I have linked to a minimal reproduction repository in the bug description | 1.0 | Setting a Poetry constraint causes lockfile update to fail - ### How are you running Renovate?
Mend Renovate hosted app on github.com
### If you're self-hosting Renovate, tell us what version of Renovate you run.
Present on self-hosted and github app
### If you're self-hosting Renovate, select which platform you are using.
gitlab.com
### If you're self-hosting Renovate, tell us what version of the platform you run.
Also observed on gitlab.com 15.8
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
When setting a python and poetry constraint, the poetry lock update command fails because renovate attempts to pip install poetry using the wrong format for pip version constraints. See the repro here: https://github.com/WarpRat/renovate-poetry-contraint-repro/pull/6
### Relevant debug logs
<details><summary>Logs</summary>
```
DEBUG: Failed to update poetry.lock file(branch="renovate/pylint-2.x")
{
"err": {
"name": "ExecError",
"cmd": "/bin/sh -c docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"",
"stderr": "ERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n",
"stdout": "installing v2 tool python v3.9.9\nlinking tool python v3.9.9\nPython 3.9.9\npip 22.3.1 from /opt/buildpack/tools/python/3.9.9/lib/python3.9/site-packages/pip (python 3.9)\nInstalled v2 /usr/local/buildpack/tools/v2/python.sh in 24 seconds\nskip cleanup, not a docker build: 8d08c5ae1170\n",
"options": {
"cwd": "/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro",
"encoding": "utf-8",
"env": {
"PIP_CACHE_DIR": "/tmp/renovate-cache/others/pip",
"HOME": "/home/ubuntu",
"PATH": "/home/ubuntu/.local/bin:/home/ubuntu/bin:/opt/buildpack/tools/python/3.9.3/bin:/home/ubuntu/.npm-global/bin:/home/ubuntu/renovateapp/node_modules/.bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",
"LC_ALL": "C.UTF-8",
"LANG": "C.UTF-8",
"BUILDPACK_CACHE_DIR": "/tmp/containerbase",
"CONTAINERBASE_CACHE_DIR": "/tmp/containerbase"
},
"maxBuffer": 10485760,
"timeout": 900000
},
"exitCode": 1,
"message": "Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"\nERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n",
"stack": "ExecError: Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"\nERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n\n at ChildProcess.<anonymous> (/home/ubuntu/renovateapp/node_modules/renovate/dist/util/exec/common.js:87:24)\n at ChildProcess.emit (node:events:525:35)\n at ChildProcess.emit (node:domain:489:12)\n at Process.ChildProcess._handle.onexit (node:internal/child_process:293:12)"
}
}
```
</details>
### Have you created a minimal reproduction repository?
I have linked to a minimal reproduction repository in the bug description | priority | setting a poetry constraint causes lockfile update to fail how are you running renovate mend renovate hosted app on github com if you re self hosting renovate tell us what version of renovate you run present on self hosted and github app if you re self hosting renovate select which platform you are using gitlab com if you re self hosting renovate tell us what version of the platform you run also observed on gitlab com was this something which used to work for you and then stopped i never saw this working describe the bug when setting a python and poetry constraint the poetry lock update command fails because renovate attempts to pip install poetry using the wrong format for pip version constraints see the repro here relevant debug logs logs debug failed to update poetry lock file branch renovate pylint x err name execerror cmd bin sh c docker run rm name renovate sidecar label renovate child v mnt renovate gh warprat renovate poetry contraint repro mnt renovate gh warprat renovate poetry contraint repro v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir e containerbase cache dir w mnt renovate gh warprat renovate poetry contraint repro docker io renovate sidecar bash l c install tool python pip install user poetry update lock no interaction pylint stderr error could not find a version that satisfies the requirement from versions none nerror no matching distribution found for n stdout installing tool python nlinking tool python npython npip from opt buildpack tools python lib site packages pip python ninstalled usr local buildpack tools python sh in seconds nskip cleanup not a docker build n options cwd mnt renovate gh warprat renovate poetry contraint repro encoding utf env pip cache dir tmp renovate cache others pip home home ubuntu path home ubuntu local bin home ubuntu bin opt buildpack tools python bin home ubuntu npm global bin home ubuntu renovateapp node modules bin usr local sbin usr local bin usr sbin usr bin sbin bin lc all c utf lang c utf buildpack cache dir tmp containerbase containerbase cache dir tmp containerbase maxbuffer timeout exitcode message command failed docker run rm name renovate sidecar label renovate child v mnt renovate gh warprat renovate poetry contraint repro mnt renovate gh warprat renovate poetry contraint repro v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir e containerbase cache dir w mnt renovate gh warprat renovate poetry contraint repro docker io renovate sidecar bash l c install tool python pip install user poetry update lock no interaction pylint nerror could not find a version that satisfies the requirement from versions none nerror no matching distribution found for n stack execerror command failed docker run rm name renovate sidecar label renovate child v mnt renovate gh warprat renovate poetry contraint repro mnt renovate gh warprat renovate poetry contraint repro v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir e containerbase cache dir w mnt renovate gh warprat renovate poetry contraint repro docker io renovate sidecar bash l c install tool python pip install user poetry update lock no interaction pylint nerror could not find a version that satisfies the requirement from versions none nerror no matching distribution found for n n at childprocess home ubuntu renovateapp node modules renovate dist util exec common js n at childprocess emit node events n at childprocess emit node domain n at process childprocess handle onexit node internal child process have you created a minimal reproduction repository i have linked to a minimal reproduction repository in the bug description | 1 |
88,251 | 3,775,408,747 | IssuesEvent | 2016-03-17 13:28:19 | marklogic/marklogic-data-hub | https://api.github.com/repos/marklogic/marklogic-data-hub | closed | Input Flow doesn't report back failures | Component:QuickStart priority:medium | When I run an input flow that fails it doesn't report the errors or the failure back to the quick-start UI | 1.0 | Input Flow doesn't report back failures - When I run an input flow that fails it doesn't report the errors or the failure back to the quick-start UI | priority | input flow doesn t report back failures when i run an input flow that fails it doesn t report the errors or the failure back to the quick start ui | 1 |
255,607 | 8,125,822,992 | IssuesEvent | 2018-08-16 22:29:58 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Add tests for 2D and 3D Line annotations | Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal | Subject says it all.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2337
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Add tests for 2D and 3D Line annotations
Assigned to: Kathleen Biagas
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 07/14/2015
Due date:
% Done: 100
Estimated time:
Created: 07/14/2015 12:53 am
Updated: 08/27/2015 08:22 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
Subject says it all.
Comments:
SVN Revision 26740A tests/rendering/line2d.pyA baseline/rendering/line2d/line2d00.pngA baseline/rendering/line2d/line2d10.pngA baseline/rendering/line2d/line2d01.pngA baseline/rendering/line2d/line2d02.pngA baseline/rendering/line2d/line2d03.pngA baseline/rendering/line2d/line2d04.pngA baseline/rendering/line2d/line2d05.pngA baseline/rendering/line2d/line2d06.pngA baseline/rendering/line2d/line2d07.pngA baseline/rendering/line2d/line2d08.pngA baseline/rendering/line2d/line2d09.pngSVN Revision 27056A baseline/rendering/line3d/line3d00.pngA baseline/rendering/line3d/Line3d01.pngA baseline/rendering/line3d/line3d10.pngA baseline/rendering/line3d/Line3d02.pngA baseline/rendering/line3d/line3d11.pngA baseline/rendering/line3d/Line3d03.pngA baseline/rendering/line3d/line3d12.pngA baseline/rendering/line3d/Line3d04.pngA baseline/rendering/line3d/line3d13.pngA baseline/rendering/line3d/line3d05.pngA baseline/rendering/line3d/line3d06.pngA baseline/rendering/line3d/line3d07.pngA baseline/rendering/line3d/line3d08.pngA baseline/rendering/line3d/line3d09.png
| 1.0 | Add tests for 2D and 3D Line annotations - Subject says it all.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2337
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Add tests for 2D and 3D Line annotations
Assigned to: Kathleen Biagas
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 07/14/2015
Due date:
% Done: 100
Estimated time:
Created: 07/14/2015 12:53 am
Updated: 08/27/2015 08:22 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
Subject says it all.
Comments:
SVN Revision 26740A tests/rendering/line2d.pyA baseline/rendering/line2d/line2d00.pngA baseline/rendering/line2d/line2d10.pngA baseline/rendering/line2d/line2d01.pngA baseline/rendering/line2d/line2d02.pngA baseline/rendering/line2d/line2d03.pngA baseline/rendering/line2d/line2d04.pngA baseline/rendering/line2d/line2d05.pngA baseline/rendering/line2d/line2d06.pngA baseline/rendering/line2d/line2d07.pngA baseline/rendering/line2d/line2d08.pngA baseline/rendering/line2d/line2d09.pngSVN Revision 27056A baseline/rendering/line3d/line3d00.pngA baseline/rendering/line3d/Line3d01.pngA baseline/rendering/line3d/line3d10.pngA baseline/rendering/line3d/Line3d02.pngA baseline/rendering/line3d/line3d11.pngA baseline/rendering/line3d/Line3d03.pngA baseline/rendering/line3d/line3d12.pngA baseline/rendering/line3d/Line3d04.pngA baseline/rendering/line3d/line3d13.pngA baseline/rendering/line3d/line3d05.pngA baseline/rendering/line3d/line3d06.pngA baseline/rendering/line3d/line3d07.pngA baseline/rendering/line3d/line3d08.pngA baseline/rendering/line3d/line3d09.png
| priority | add tests for and line annotations subject says it all redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject add tests for and line annotations assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created am updated pm likelihood severity found in version impact medium expected use occasional os all support group any description subject says it all comments svn revision tests rendering pya baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pngsvn revision baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering png | 1 |
417,799 | 12,179,396,154 | IssuesEvent | 2020-04-28 10:36:13 | Arquisoft/viade_en1a | https://api.github.com/repos/Arquisoft/viade_en1a | closed | You can not delete shared routes | bug medium priority | The button of deleting routes someone shared with you does not work. You should be able to dismiss shared routes you do not want.
| 1.0 | You can not delete shared routes - The button of deleting routes someone shared with you does not work. You should be able to dismiss shared routes you do not want.
| priority | you can not delete shared routes the button of deleting routes someone shared with you does not work you should be able to dismiss shared routes you do not want | 1 |
512,587 | 14,901,413,177 | IssuesEvent | 2021-01-21 16:26:21 | staxrip/staxrip | https://api.github.com/repos/staxrip/staxrip | closed | Please add NvEnc --vpp-colorspace options. | added/fixed/done feature request priority medium | HDR looks terrible on my TVs with washed out blacks and faded color. Both TVs support HDR and show it as enabled when HDR content is present. However, the colors and blacks still look terrible compared to normal SDR which looks great. This HDR issue is well confirmed on the internet and Tone mapping HDR to SDR has become the current best solution.
I use Staxrip with the Nvidia H265 encoder as this is much faster than x265 and retains excellent quality. Nvidia encoder supports fast hardware tone mapping using "--vpp-colorspace hdr2sdr=reinhard" command line option. Currently we have to add this option manually in Encoder Options/Other/Custom. Works well with excellent results.
Can we have a "Colorspace Conversion" option added to the VPP GUI section of the NvEnc Encoder Options? Should include Reinhard, Hable and Mobius conversion options as a selection box. NvEnc supports a heap of manual options for Colorspace Conversions as well. These could be added too, perhaps at a later stage as I think most people will be happy with the 3 Auto selections listed above.
Will be great to have the Colorspace options available in their proper easy place in the GUI so they can be used by the many and not have to be manually added by the very determined few. | 1.0 | Please add NvEnc --vpp-colorspace options. - HDR looks terrible on my TVs with washed out blacks and faded color. Both TVs support HDR and show it as enabled when HDR content is present. However, the colors and blacks still look terrible compared to normal SDR which looks great. This HDR issue is well confirmed on the internet and Tone mapping HDR to SDR has become the current best solution.
I use Staxrip with the Nvidia H265 encoder as this is much faster than x265 and retains excellent quality. Nvidia encoder supports fast hardware tone mapping using "--vpp-colorspace hdr2sdr=reinhard" command line option. Currently we have to add this option manually in Encoder Options/Other/Custom. Works well with excellent results.
Can we have a "Colorspace Conversion" option added to the VPP GUI section of the NvEnc Encoder Options? Should include Reinhard, Hable and Mobius conversion options as a selection box. NvEnc supports a heap of manual options for Colorspace Conversions as well. These could be added too, perhaps at a later stage as I think most people will be happy with the 3 Auto selections listed above.
Will be great to have the Colorspace options available in their proper easy place in the GUI so they can be used by the many and not have to be manually added by the very determined few. | priority | please add nvenc vpp colorspace options hdr looks terrible on my tvs with washed out blacks and faded color both tvs support hdr and show it as enabled when hdr content is present however the colors and blacks still look terrible compared to normal sdr which looks great this hdr issue is well confirmed on the internet and tone mapping hdr to sdr has become the current best solution i use staxrip with the nvidia encoder as this is much faster than and retains excellent quality nvidia encoder supports fast hardware tone mapping using vpp colorspace reinhard command line option currently we have to add this option manually in encoder options other custom works well with excellent results can we have a colorspace conversion option added to the vpp gui section of the nvenc encoder options should include reinhard hable and mobius conversion options as a selection box nvenc supports a heap of manual options for colorspace conversions as well these could be added too perhaps at a later stage as i think most people will be happy with the auto selections listed above will be great to have the colorspace options available in their proper easy place in the gui so they can be used by the many and not have to be manually added by the very determined few | 1 |
394,043 | 11,628,622,961 | IssuesEvent | 2020-02-27 18:42:21 | wynn-rj/sword-and-bored-game | https://api.github.com/repos/wynn-rj/sword-and-bored-game | opened | Functioning Unit and Squad System | enhancement medium-priority strategy view | - Add "Available Unit" pool into base
- Upon creating a unit you can place it in any City under your control. Creation is instant, sending it to another city takes 1 turn. If you leave it in the base you can use it on the same turn it is created.
Add Squads that you can put units together in:
- Between 1-6 units per squad
- Squads have single turn on World Map, act together
| 1.0 | Functioning Unit and Squad System - - Add "Available Unit" pool into base
- Upon creating a unit you can place it in any City under your control. Creation is instant, sending it to another city takes 1 turn. If you leave it in the base you can use it on the same turn it is created.
Add Squads that you can put units together in:
- Between 1-6 units per squad
- Squads have single turn on World Map, act together
| priority | functioning unit and squad system add available unit pool into base upon creating a unit you can place it in any city under your control creation is instant sending it to another city takes turn if you leave it in the base you can use it on the same turn it is created add squads that you can put units together in between units per squad squads have single turn on world map act together | 1 |
769,678 | 27,016,296,170 | IssuesEvent | 2023-02-10 19:43:57 | CDCgov/prime-reportstream | https://api.github.com/repos/CDCgov/prime-reportstream | closed | CA - LA County Test Date Error | onboarding-ops Medium Priority California | LA County California is reporting issues with test dates being wrong.
##Background Information
At LA county, we have received some reports with date issue. See summaries below and examples attached. Please let us know if you’re able to help resolve these issues.
 | 1.0 | CA - LA County Test Date Error - LA County California is reporting issues with test dates being wrong.
##Background Information
At LA county, we have received some reports with date issue. See summaries below and examples attached. Please let us know if you’re able to help resolve these issues.
 | priority | ca la county test date error la county california is reporting issues with test dates being wrong background information at la county we have received some reports with date issue see summaries below and examples attached please let us know if you’re able to help resolve these issues | 1 |
454,460 | 13,101,476,922 | IssuesEvent | 2020-08-04 03:52:52 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | Indicator for unsynced local changes | Accepted Anki Ecosystem Compatibility Enhancement Keep Open Priority-Medium Sync | Originally reported on Google Code with ID 2638
```
I would like an indicator that shows that there are unsynced local changes (learning
progress).
I learn on many devices (staniory (PC) and mobile) in parallel and commonly sync to
many times because i am unsure about the state of the devices. “Did i learn anything
on the tablet today? Don’t know, better sync once more to be safe.”
It could be something unobtrusive. Maybe an * added to the “You learned X cards in
Y” line at the bottom of the deck picker.
Is this feature available on similar apps (…)?
Not really in learning apps afaik. The syncing is one of the great pluses of AnkiDroid
compared to other apps.
But Emacs uses a * to indicate unsaved (rather than unsynced) changes, which i find
quite useful.
```
Reported by `ospalh` on 2015-06-10 09:46:18
| 1.0 | Indicator for unsynced local changes - Originally reported on Google Code with ID 2638
```
I would like an indicator that shows that there are unsynced local changes (learning
progress).
I learn on many devices (staniory (PC) and mobile) in parallel and commonly sync to
many times because i am unsure about the state of the devices. “Did i learn anything
on the tablet today? Don’t know, better sync once more to be safe.”
It could be something unobtrusive. Maybe an * added to the “You learned X cards in
Y” line at the bottom of the deck picker.
Is this feature available on similar apps (…)?
Not really in learning apps afaik. The syncing is one of the great pluses of AnkiDroid
compared to other apps.
But Emacs uses a * to indicate unsaved (rather than unsynced) changes, which i find
quite useful.
```
Reported by `ospalh` on 2015-06-10 09:46:18
| priority | indicator for unsynced local changes originally reported on google code with id i would like an indicator that shows that there are unsynced local changes learning progress i learn on many devices staniory pc and mobile in parallel and commonly sync to many times because i am unsure about the state of the devices “did i learn anything on the tablet today don’t know better sync once more to be safe ” it could be something unobtrusive maybe an added to the “you learned x cards in y” line at the bottom of the deck picker is this feature available on similar apps … not really in learning apps afaik the syncing is one of the great pluses of ankidroid compared to other apps but emacs uses a to indicate unsaved rather than unsynced changes which i find quite useful reported by ospalh on | 1 |
778,471 | 27,318,042,161 | IssuesEvent | 2023-02-24 17:14:32 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Enable single-row optimizations for valid RETURNING expressions | kind/enhancement area/ysql priority/medium | Jira Link: [DB-2310](https://yugabyte.atlassian.net/browse/DB-2310)
In the fix for #1028 we are adding support for single-row optimization on UPDATE/DELETEs, however currently it only allows for simple RETURNING clauses that just specify columns.
However this is too conservative, as there are valid single-row returning expressions (i.e. those that only use column values included in the query itself) that are more complex than just a standalone column.
Example of query that is currently (once diff lands) optimized:
`DELETE FROM t WHERE a = 1 RETURNING a;`
Example of query that is valid but is not optimized:
`DELETE FROM t WHERE a = 1 RETURNING a + 1`
This issue tracks adding optimization support for the latter case and all general expressions that only involve column values that are included in the query. | 1.0 | [YSQL] Enable single-row optimizations for valid RETURNING expressions - Jira Link: [DB-2310](https://yugabyte.atlassian.net/browse/DB-2310)
In the fix for #1028 we are adding support for single-row optimization on UPDATE/DELETEs, however currently it only allows for simple RETURNING clauses that just specify columns.
However this is too conservative, as there are valid single-row returning expressions (i.e. those that only use column values included in the query itself) that are more complex than just a standalone column.
Example of query that is currently (once diff lands) optimized:
`DELETE FROM t WHERE a = 1 RETURNING a;`
Example of query that is valid but is not optimized:
`DELETE FROM t WHERE a = 1 RETURNING a + 1`
This issue tracks adding optimization support for the latter case and all general expressions that only involve column values that are included in the query. | priority | enable single row optimizations for valid returning expressions jira link in the fix for we are adding support for single row optimization on update deletes however currently it only allows for simple returning clauses that just specify columns however this is too conservative as there are valid single row returning expressions i e those that only use column values included in the query itself that are more complex than just a standalone column example of query that is currently once diff lands optimized delete from t where a returning a example of query that is valid but is not optimized delete from t where a returning a this issue tracks adding optimization support for the latter case and all general expressions that only involve column values that are included in the query | 1 |
829,504 | 31,881,440,433 | IssuesEvent | 2023-09-16 12:28:24 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | closed | Notification Delay not working | Type: Bug Priority: Medium Status: Available | ### Describe the bug
WATCHTOWER_NOTIFICATION_DELAY / WATCHTOWER_NOTIFICATIONS_DELAY environment variables seems to have no effect.
Notifications are sent directly
Please note
https://containrrr.dev/watchtower/notifications/ says it's WATCHTOWER_NOTIFICATION_DELAY
but source code seems to indicate WATCHTOWER_NOTIFICATIONS_DELAY (notice the "S")
https://github.com/containrrr/watchtower/blob/a56b9bdb7c7c8de96fc4e5f5fe5b5f4a952447be/internal/flags/flags.go#L208
To be sure I set both in my container
### Steps to reproduce
1. Set WATCHTOWER_NOTIFICATION_DELAY=120 and WATCHTOWER_NOTIFICATION_URL
2. Trigger a notification (by starting up an old image or setting WATCHTOWER_NO_STARTUP_MESSAGE to false and restart container)
### Expected behavior
Notification should be triggered later
### Screenshots
_No response_
### Environment
- Debian Bookworm
- x86_64
- Docker Version : 24.0.5
"Env": [
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",
"TZ=Europe/Paris",
"WATCHTOWER_LABEL_ENABLE=true",
"WATCHTOWER_MONITOR_ONLY=False",
"WATCHTOWER_NOTIFICATION_URL=generic://xxxxxxxxxxxx/webhook/xxxxxxxxxx?template=json",
"WATCHTOWER_NOTIFICATIONS=shoutrrr",
"WATCHTOWER_SCHEDULE=0 0 21 * * *",
"WATCHTOWER_NO_STARTUP_MESSAGE=False",
"WATCHTOWER_NOTIFICATIONS_DELAY=120",
"WATCHTOWER_NOTIFICATION_DELAY=120",
"WATCHTOWER_DEBUG=True"
],
### Your logs
```text
2023-09-13T18:46:34.665775703Z time="2023-09-13T20:46:34+02:00" level=debug msg="Sleeping for a second to ensure the docker api client has been properly initialized."
2023-09-13T18:46:35.665954195Z time="2023-09-13T20:46:35+02:00" level=debug msg="Making sure everything is sane before starting"
2023-09-13T18:46:35.665993015Z time="2023-09-13T20:46:35+02:00" level=debug msg="Retrieving running containers"
2023-09-13T18:46:35.752821755Z time="2023-09-13T20:46:35+02:00" level=debug msg="There are no additional watchtower containers"
2023-09-13T18:46:35.752845425Z time="2023-09-13T20:46:35+02:00" level=debug msg="Watchtower HTTP API skipped."
2023-09-13T18:46:35.752851085Z time="2023-09-13T20:46:35+02:00" level=info msg="Watchtower 1.5.3"
2023-09-13T18:46:35.752855385Z time="2023-09-13T20:46:35+02:00" level=info msg="Using notifications: generic"
2023-09-13T18:46:35.752859425Z time="2023-09-13T20:46:35+02:00" level=info msg="Only checking containers using enable label"
2023-09-13T18:46:35.752867245Z time="2023-09-13T20:46:35+02:00" level=info msg="Scheduling first run: 2023-09-13 21:00:00 +0200 CEST"
2023-09-13T18:46:35.752871475Z time="2023-09-13T20:46:35+02:00" level=info msg="Note that the first check will be performed in 13 minutes, 24 seconds"
2023-09-13T19:00:00.000458126Z time="2023-09-13T21:00:00+02:00" level=debug msg="Checking containers for updated images"
2023-09-13T19:00:00.000508366Z time="2023-09-13T21:00:00+02:00" level=debug msg="Retrieving running containers"
2023-09-13T19:00:01.164754872Z time="2023-09-13T21:00:01+02:00" level=debug msg="Trying to load authentication credentials." container=/homeassistant image="docker.io/homeassistant/home-assistant:stable"
2023-09-13T19:00:01.164842423Z time="2023-09-13T21:00:01+02:00" level=debug msg="No credentials for docker.io found" config_file=/config.json
2023-09-13T19:00:01.164865053Z time="2023-09-13T21:00:01+02:00" level=debug msg="Got image name: docker.io/homeassistant/home-assistant:stable"
2023-09-13T19:00:01.164881323Z time="2023-09-13T21:00:01+02:00" level=debug msg="Checking if pull is needed" container=/homeassistant image="docker.io/homeassistant/home-assistant:stable"
2023-09-13T19:00:01.164909713Z time="2023-09-13T21:00:01+02:00" level=debug msg="Building challenge URL" URL="https://index.docker.io/v2/"
2023-09-13T19:00:01.488038988Z time="2023-09-13T21:00:01+02:00" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://auth.docker.io/token\",service=\"registry.docker.io\"" status="401 Unauthorized"
2023-09-13T19:00:01.488077238Z time="2023-09-13T21:00:01+02:00" level=debug msg="Checking challenge header content" realm="https://auth.docker.io/token" service=registry.docker.io
2023-09-13T19:00:01.488091068Z time="2023-09-13T21:00:01+02:00" level=debug msg="Setting scope for auth token" image=docker.io/homeassistant/home-assistant scope="repository:homeassistant/home-assistant:pull"
2023-09-13T19:00:01.488106329Z time="2023-09-13T21:00:01+02:00" level=debug msg="No credentials found."
2023-09-13T19:00:01.839030275Z time="2023-09-13T21:00:01+02:00" level=debug msg="Parsing image ref" host=index.docker.io image=docker.io/homeassistant/home-assistant normalized="docker.io/homeassistant/home-assistant:stable" tag=stable
2023-09-13T19:00:01.839075615Z time="2023-09-13T21:00:01+02:00" level=debug msg="Doing a HEAD request to fetch a digest" url="https://index.docker.io/v2/homeassistant/home-assistant/manifests/stable"
2023-09-13T19:00:02.228619850Z time="2023-09-13T21:00:02+02:00" level=debug msg="Found a remote digest to compare with" remote="sha256:223c9986ff318eaf35fac70351900cf766c15faf82fbe54150b73b7cfc3c933a"
2023-09-13T19:00:02.228666641Z time="2023-09-13T21:00:02+02:00" level=debug msg=Comparing local="sha256:183513a13fbcb758eb525f857cf3ec69a619a5ec27c117a2077ab2f4e7a3c052" remote="sha256:223c9986ff318eaf35fac70351900cf766c15faf82fbe54150b73b7cfc3c933a"
2023-09-13T19:00:02.228699001Z time="2023-09-13T21:00:02+02:00" level=debug msg="Digests did not match, doing a pull."
2023-09-13T19:00:02.228707471Z time="2023-09-13T21:00:02+02:00" level=debug msg="Pulling image" container=/homeassistant image="docker.io/homeassistant/home-assistant:stable"
2023-09-13T19:00:54.987751992Z time="2023-09-13T21:00:54+02:00" level=info msg="Found new docker.io/homeassistant/home-assistant:stable image (1bcf95d7b112)"
2023-09-13T19:01:14.350469904Z time="2023-09-13T21:01:14+02:00" level=debug msg="Removing container 10736233407b"
2023-09-13T19:01:14.393279697Z time="2023-09-13T21:01:14+02:00" level=info msg="Stopping /homeassistant (6da3dfd8d3a6) with SIGTERM"
2023-09-13T19:01:21.908834811Z time="2023-09-13T21:01:21+02:00" level=info msg="Session done" Failed=0 Scanned=11 Updated=2 notify=no
2023-09-13T19:01:21.910918050Z time="2023-09-13T21:01:21+02:00" level=debug msg="Scheduled next run: 2023-09-14 21:00:00 +0200 CEST"
2023-09-13T19:01:21.962319729Z time="2023-09-13T21:01:21+02:00" level=error msg="Failed to send shoutrrr notification" error="an error occurred while sending notification to generic webhook: server returned response status code 502 Bad Gateway" index=0 notify=no service=generic
```
### Additional context
_No response_ | 1.0 | Notification Delay not working - ### Describe the bug
WATCHTOWER_NOTIFICATION_DELAY / WATCHTOWER_NOTIFICATIONS_DELAY environment variables seems to have no effect.
Notifications are sent directly
Please note
https://containrrr.dev/watchtower/notifications/ says it's WATCHTOWER_NOTIFICATION_DELAY
but source code seems to indicate WATCHTOWER_NOTIFICATIONS_DELAY (notice the "S")
https://github.com/containrrr/watchtower/blob/a56b9bdb7c7c8de96fc4e5f5fe5b5f4a952447be/internal/flags/flags.go#L208
To be sure I set both in my container
### Steps to reproduce
1. Set WATCHTOWER_NOTIFICATION_DELAY=120 and WATCHTOWER_NOTIFICATION_URL
2. Trigger a notification (by starting up an old image or setting WATCHTOWER_NO_STARTUP_MESSAGE to false and restart container)
### Expected behavior
Notification should be triggered later
### Screenshots
_No response_
### Environment
- Debian Bookworm
- x86_64
- Docker Version : 24.0.5
"Env": [
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",
"TZ=Europe/Paris",
"WATCHTOWER_LABEL_ENABLE=true",
"WATCHTOWER_MONITOR_ONLY=False",
"WATCHTOWER_NOTIFICATION_URL=generic://xxxxxxxxxxxx/webhook/xxxxxxxxxx?template=json",
"WATCHTOWER_NOTIFICATIONS=shoutrrr",
"WATCHTOWER_SCHEDULE=0 0 21 * * *",
"WATCHTOWER_NO_STARTUP_MESSAGE=False",
"WATCHTOWER_NOTIFICATIONS_DELAY=120",
"WATCHTOWER_NOTIFICATION_DELAY=120",
"WATCHTOWER_DEBUG=True"
],
### Your logs
```text
2023-09-13T18:46:34.665775703Z time="2023-09-13T20:46:34+02:00" level=debug msg="Sleeping for a second to ensure the docker api client has been properly initialized."
2023-09-13T18:46:35.665954195Z time="2023-09-13T20:46:35+02:00" level=debug msg="Making sure everything is sane before starting"
2023-09-13T18:46:35.665993015Z time="2023-09-13T20:46:35+02:00" level=debug msg="Retrieving running containers"
2023-09-13T18:46:35.752821755Z time="2023-09-13T20:46:35+02:00" level=debug msg="There are no additional watchtower containers"
2023-09-13T18:46:35.752845425Z time="2023-09-13T20:46:35+02:00" level=debug msg="Watchtower HTTP API skipped."
2023-09-13T18:46:35.752851085Z time="2023-09-13T20:46:35+02:00" level=info msg="Watchtower 1.5.3"
2023-09-13T18:46:35.752855385Z time="2023-09-13T20:46:35+02:00" level=info msg="Using notifications: generic"
2023-09-13T18:46:35.752859425Z time="2023-09-13T20:46:35+02:00" level=info msg="Only checking containers using enable label"
2023-09-13T18:46:35.752867245Z time="2023-09-13T20:46:35+02:00" level=info msg="Scheduling first run: 2023-09-13 21:00:00 +0200 CEST"
2023-09-13T18:46:35.752871475Z time="2023-09-13T20:46:35+02:00" level=info msg="Note that the first check will be performed in 13 minutes, 24 seconds"
2023-09-13T19:00:00.000458126Z time="2023-09-13T21:00:00+02:00" level=debug msg="Checking containers for updated images"
2023-09-13T19:00:00.000508366Z time="2023-09-13T21:00:00+02:00" level=debug msg="Retrieving running containers"
2023-09-13T19:00:01.164754872Z time="2023-09-13T21:00:01+02:00" level=debug msg="Trying to load authentication credentials." container=/homeassistant image="docker.io/homeassistant/home-assistant:stable"
2023-09-13T19:00:01.164842423Z time="2023-09-13T21:00:01+02:00" level=debug msg="No credentials for docker.io found" config_file=/config.json
2023-09-13T19:00:01.164865053Z time="2023-09-13T21:00:01+02:00" level=debug msg="Got image name: docker.io/homeassistant/home-assistant:stable"
2023-09-13T19:00:01.164881323Z time="2023-09-13T21:00:01+02:00" level=debug msg="Checking if pull is needed" container=/homeassistant image="docker.io/homeassistant/home-assistant:stable"
2023-09-13T19:00:01.164909713Z time="2023-09-13T21:00:01+02:00" level=debug msg="Building challenge URL" URL="https://index.docker.io/v2/"
2023-09-13T19:00:01.488038988Z time="2023-09-13T21:00:01+02:00" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://auth.docker.io/token\",service=\"registry.docker.io\"" status="401 Unauthorized"
2023-09-13T19:00:01.488077238Z time="2023-09-13T21:00:01+02:00" level=debug msg="Checking challenge header content" realm="https://auth.docker.io/token" service=registry.docker.io
2023-09-13T19:00:01.488091068Z time="2023-09-13T21:00:01+02:00" level=debug msg="Setting scope for auth token" image=docker.io/homeassistant/home-assistant scope="repository:homeassistant/home-assistant:pull"
2023-09-13T19:00:01.488106329Z time="2023-09-13T21:00:01+02:00" level=debug msg="No credentials found."
2023-09-13T19:00:01.839030275Z time="2023-09-13T21:00:01+02:00" level=debug msg="Parsing image ref" host=index.docker.io image=docker.io/homeassistant/home-assistant normalized="docker.io/homeassistant/home-assistant:stable" tag=stable
2023-09-13T19:00:01.839075615Z time="2023-09-13T21:00:01+02:00" level=debug msg="Doing a HEAD request to fetch a digest" url="https://index.docker.io/v2/homeassistant/home-assistant/manifests/stable"
2023-09-13T19:00:02.228619850Z time="2023-09-13T21:00:02+02:00" level=debug msg="Found a remote digest to compare with" remote="sha256:223c9986ff318eaf35fac70351900cf766c15faf82fbe54150b73b7cfc3c933a"
2023-09-13T19:00:02.228666641Z time="2023-09-13T21:00:02+02:00" level=debug msg=Comparing local="sha256:183513a13fbcb758eb525f857cf3ec69a619a5ec27c117a2077ab2f4e7a3c052" remote="sha256:223c9986ff318eaf35fac70351900cf766c15faf82fbe54150b73b7cfc3c933a"
2023-09-13T19:00:02.228699001Z time="2023-09-13T21:00:02+02:00" level=debug msg="Digests did not match, doing a pull."
2023-09-13T19:00:02.228707471Z time="2023-09-13T21:00:02+02:00" level=debug msg="Pulling image" container=/homeassistant image="docker.io/homeassistant/home-assistant:stable"
2023-09-13T19:00:54.987751992Z time="2023-09-13T21:00:54+02:00" level=info msg="Found new docker.io/homeassistant/home-assistant:stable image (1bcf95d7b112)"
2023-09-13T19:01:14.350469904Z time="2023-09-13T21:01:14+02:00" level=debug msg="Removing container 10736233407b"
2023-09-13T19:01:14.393279697Z time="2023-09-13T21:01:14+02:00" level=info msg="Stopping /homeassistant (6da3dfd8d3a6) with SIGTERM"
2023-09-13T19:01:21.908834811Z time="2023-09-13T21:01:21+02:00" level=info msg="Session done" Failed=0 Scanned=11 Updated=2 notify=no
2023-09-13T19:01:21.910918050Z time="2023-09-13T21:01:21+02:00" level=debug msg="Scheduled next run: 2023-09-14 21:00:00 +0200 CEST"
2023-09-13T19:01:21.962319729Z time="2023-09-13T21:01:21+02:00" level=error msg="Failed to send shoutrrr notification" error="an error occurred while sending notification to generic webhook: server returned response status code 502 Bad Gateway" index=0 notify=no service=generic
```
### Additional context
_No response_ | priority | notification delay not working describe the bug watchtower notification delay watchtower notifications delay environment variables seems to have no effect notifications are sent directly please note says it s watchtower notification delay but source code seems to indicate watchtower notifications delay notice the s to be sure i set both in my container steps to reproduce set watchtower notification delay and watchtower notification url trigger a notification by starting up an old image or setting watchtower no startup message to false and restart container expected behavior notification should be triggered later screenshots no response environment debian bookworm docker version env path usr local sbin usr local bin usr sbin usr bin sbin bin tz europe paris watchtower label enable true watchtower monitor only false watchtower notification url generic xxxxxxxxxxxx webhook xxxxxxxxxx template json watchtower notifications shoutrrr watchtower schedule watchtower no startup message false watchtower notifications delay watchtower notification delay watchtower debug true your logs text time level debug msg sleeping for a second to ensure the docker api client has been properly initialized time level debug msg making sure everything is sane before starting time level debug msg retrieving running containers time level debug msg there are no additional watchtower containers time level debug msg watchtower http api skipped time level info msg watchtower time level info msg using notifications generic time level info msg only checking containers using enable label time level info msg scheduling first run cest time level info msg note that the first check will be performed in minutes seconds time level debug msg checking containers for updated images time level debug msg retrieving running containers time level debug msg trying to load authentication credentials container homeassistant image docker io homeassistant home assistant stable time level debug msg no credentials for docker io found config file config json time level debug msg got image name docker io homeassistant home assistant stable time level debug msg checking if pull is needed container homeassistant image docker io homeassistant home assistant stable time level debug msg building challenge url url time level debug msg got response to challenge request header bearer realm status unauthorized time level debug msg checking challenge header content realm service registry docker io time level debug msg setting scope for auth token image docker io homeassistant home assistant scope repository homeassistant home assistant pull time level debug msg no credentials found time level debug msg parsing image ref host index docker io image docker io homeassistant home assistant normalized docker io homeassistant home assistant stable tag stable time level debug msg doing a head request to fetch a digest url time level debug msg found a remote digest to compare with remote time level debug msg comparing local remote time level debug msg digests did not match doing a pull time level debug msg pulling image container homeassistant image docker io homeassistant home assistant stable time level info msg found new docker io homeassistant home assistant stable image time level debug msg removing container time level info msg stopping homeassistant with sigterm time level info msg session done failed scanned updated notify no time level debug msg scheduled next run cest time level error msg failed to send shoutrrr notification error an error occurred while sending notification to generic webhook server returned response status code bad gateway index notify no service generic additional context no response | 1 |
508,519 | 14,701,734,535 | IssuesEvent | 2021-01-04 12:24:47 | seung-lab/neuroglancer | https://api.github.com/repos/seung-lab/neuroglancer | closed | meshes do not show up in Chrome | Priority: Medium Type: Bug | it works in Firefox.
My chrome:
Version 83.0.4103.116 (Official Build) (64-bit)
OS: Ubuntu 18.04 | 1.0 | meshes do not show up in Chrome - it works in Firefox.
My chrome:
Version 83.0.4103.116 (Official Build) (64-bit)
OS: Ubuntu 18.04 | priority | meshes do not show up in chrome it works in firefox my chrome version official build bit os ubuntu | 1 |
41,972 | 2,869,091,092 | IssuesEvent | 2015-06-05 23:15:45 | dart-lang/polymer-dart | https://api.github.com/repos/dart-lang/polymer-dart | opened | Cannot use iterables in bindings | bug PolymerMilestone-Next Priority-Medium | <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)**
_Originally opened as dart-lang/sdk#19945_
----
They could generate infinite loops. Right now we are OK with them because we are not discarding changes in 'get value' in polymer-expressions, but you'd hit the issue if we fix that. We also hit this if you use path_observers.
Here is a change that illustrates the problem:
https://codereview.chromium.org/382683005
ObservableMap.keys would always be different, so we would think there was a change every time we do a read. The CL patches ObservableMap.keys to cache it's previous result to illustrate that the issue goes away if we know somehow that the value didn't change.
I think we need to guarantee that with any Bindable, calling `get value` twice in a row should return either the same thing, or equal values.
Haven't we solved this issue before already?
| 1.0 | Cannot use iterables in bindings - <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)**
_Originally opened as dart-lang/sdk#19945_
----
They could generate infinite loops. Right now we are OK with them because we are not discarding changes in 'get value' in polymer-expressions, but you'd hit the issue if we fix that. We also hit this if you use path_observers.
Here is a change that illustrates the problem:
https://codereview.chromium.org/382683005
ObservableMap.keys would always be different, so we would think there was a change every time we do a read. The CL patches ObservableMap.keys to cache it's previous result to illustrate that the issue goes away if we know somehow that the value didn't change.
I think we need to guarantee that with any Bindable, calling `get value` twice in a row should return either the same thing, or equal values.
Haven't we solved this issue before already?
| priority | cannot use iterables in bindings issue by originally opened as dart lang sdk they could generate infinite loops right now we are ok with them because we are not discarding changes in get value in polymer expressions but you d hit the issue if we fix that we also hit this if you use path observers here is a change that illustrates the problem observablemap keys would always be different so we would think there was a change every time we do a read the cl patches observablemap keys to cache it s previous result to illustrate that the issue goes away if we know somehow that the value didn t change i think we need to guarantee that with any bindable calling get value twice in a row should return either the same thing or equal values haven t we solved this issue before already | 1 |
372,486 | 11,015,550,320 | IssuesEvent | 2019-12-05 02:01:53 | NIFCLOUD-mbaas/ncmb_ios | https://api.github.com/repos/NIFCLOUD-mbaas/ncmb_ios | closed | UIWebview - Deprecated | Priority: Medium | つい最近からだと思いますが。アプリをApp Storeにアップロードすると、UIWebViewを使用しているアプリだと以下のようにワーニング出るようになってしまいました。
NCMBRichPushViewクラスにて使用しているようなので、UIWebViewからWKWebViewへの変更をお願いします。
We identified one or more issues with a recent delivery for your app, "xxxx". Your delivery was successful, but you may wish to correct the following issues in your next delivery:
ITMS-90809: Deprecated API Usage - Apple will stop accepting submissions of apps that use UIWebView APIs .
| 1.0 | UIWebview - Deprecated - つい最近からだと思いますが。アプリをApp Storeにアップロードすると、UIWebViewを使用しているアプリだと以下のようにワーニング出るようになってしまいました。
NCMBRichPushViewクラスにて使用しているようなので、UIWebViewからWKWebViewへの変更をお願いします。
We identified one or more issues with a recent delivery for your app, "xxxx". Your delivery was successful, but you may wish to correct the following issues in your next delivery:
ITMS-90809: Deprecated API Usage - Apple will stop accepting submissions of apps that use UIWebView APIs .
| priority | uiwebview deprecated つい最近からだと思いますが。アプリをapp storeにアップロードすると、uiwebviewを使用しているアプリだと以下のようにワーニング出るようになってしまいました。 ncmbrichpushviewクラスにて使用しているようなので、uiwebviewからwkwebviewへの変更をお願いします。 we identified one or more issues with a recent delivery for your app xxxx your delivery was successful but you may wish to correct the following issues in your next delivery itms deprecated api usage apple will stop accepting submissions of apps that use uiwebview apis | 1 |
682,441 | 23,345,322,792 | IssuesEvent | 2022-08-09 17:23:09 | Javacord/Javacord | https://api.github.com/repos/Javacord/Javacord | closed | MessageUpdateHandler is outdated | medium priority | The MessageUpdateHandler is outdated and for example does not update the message if components have been changed and also does not fire an event in this case | 1.0 | MessageUpdateHandler is outdated - The MessageUpdateHandler is outdated and for example does not update the message if components have been changed and also does not fire an event in this case | priority | messageupdatehandler is outdated the messageupdatehandler is outdated and for example does not update the message if components have been changed and also does not fire an event in this case | 1 |
111,104 | 4,461,558,711 | IssuesEvent | 2016-08-24 06:20:27 | bespokeinteractive/ipdapp | https://api.github.com/repos/bespokeinteractive/ipdapp | opened | IPD module: Search functionality within the IPD module not working as it should | bug Medium priority | PD module: Search functionality within the IPD module not working as it should
Suggestion
-Work on clearing on searched item to reveal all items in the list | 1.0 | IPD module: Search functionality within the IPD module not working as it should - PD module: Search functionality within the IPD module not working as it should
Suggestion
-Work on clearing on searched item to reveal all items in the list | priority | ipd module search functionality within the ipd module not working as it should pd module search functionality within the ipd module not working as it should suggestion work on clearing on searched item to reveal all items in the list | 1 |
667,648 | 22,495,497,778 | IssuesEvent | 2022-06-23 07:11:03 | OffprintStudios/dragonfish | https://api.github.com/repos/OffprintStudios/dragonfish | opened | Refreshing approval queue resulted in authentication issue where said work approver doesn't have permission to view | Bug Medium Priority | Reported by work approver:
I just refreshed the queue page, and went from seeing the story that had been claimed by Techie, to getting the 'you do not have permission to view this page' message.
Clicking on 'home' and then back to the queue via the dash worked as expected, though, now showing 'The queue is clear!'
This likely ties into other general issues with authentication | 1.0 | Refreshing approval queue resulted in authentication issue where said work approver doesn't have permission to view - Reported by work approver:
I just refreshed the queue page, and went from seeing the story that had been claimed by Techie, to getting the 'you do not have permission to view this page' message.
Clicking on 'home' and then back to the queue via the dash worked as expected, though, now showing 'The queue is clear!'
This likely ties into other general issues with authentication | priority | refreshing approval queue resulted in authentication issue where said work approver doesn t have permission to view reported by work approver i just refreshed the queue page and went from seeing the story that had been claimed by techie to getting the you do not have permission to view this page message clicking on home and then back to the queue via the dash worked as expected though now showing the queue is clear this likely ties into other general issues with authentication | 1 |
13,785 | 2,610,300,572 | IssuesEvent | 2015-02-26 19:36:36 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Icons for hedgewars demo *.hwd and hedgewars save *.hws files | auto-migrated Priority-Medium Type-Enhancement | ```
Hi. This isnt a real problem. its just cosmetic. On KDE Plasma, there are
really talented icon designers. (oxygen icons) so maybe you could ask them if
they could make this two icons für you or ill ask them if this is ok. Because
this [?] symbol isnt so good :(
```
-----
Original issue reported on code.google.com by `apfelmau...@googlemail.com` on 7 Apr 2011 at 9:51 | 1.0 | Icons for hedgewars demo *.hwd and hedgewars save *.hws files - ```
Hi. This isnt a real problem. its just cosmetic. On KDE Plasma, there are
really talented icon designers. (oxygen icons) so maybe you could ask them if
they could make this two icons für you or ill ask them if this is ok. Because
this [?] symbol isnt so good :(
```
-----
Original issue reported on code.google.com by `apfelmau...@googlemail.com` on 7 Apr 2011 at 9:51 | priority | icons for hedgewars demo hwd and hedgewars save hws files hi this isnt a real problem its just cosmetic on kde plasma there are really talented icon designers oxygen icons so maybe you could ask them if they could make this two icons für you or ill ask them if this is ok because this symbol isnt so good original issue reported on code google com by apfelmau googlemail com on apr at | 1 |
798,399 | 28,262,986,501 | IssuesEvent | 2023-04-07 02:15:26 | AY2223S2-CS2103-W17-3/tp | https://api.github.com/repos/AY2223S2-CS2103-W17-3/tp | closed | [PE-D][Tester D] Find functionality inconsistent with UserGuide (partial vs full word search) | bug docs priority.Medium | ## Bug Report - ezSchedule App
### Issue Description
There is an issue with find, where finding strings that are partial in a word will return the event.
For example, `find e` will return any events which names' contain the letter e. This contradicts the userguide, which says that find finds based on **keyword**.

### Expected Behavior
I expected the app to find events based on word alone(non-case sensitive). As some words may be part of another word for example, event `eating fish` and event `fishing`, I may only want to find events that contain fish, and not care about events like fishing.
### Actual Behavior
Currently, the app's find returns any string as long as they contain any letter passed as a parameter to find.
### Steps to Reproduce
1. Open the ezSchedule app.
2. Add a tennis event
3. Use the find command `find e`
### Environment
- OS: Mac OS 13.3 Ventura
- App Version: 1.3.1
### Priority
We consider this to be a very low priority bug, as it can cause confusion for users and make it slightly unintuitive to use the app, but not a huge problem.
<!--session: 1680242278740-fb3aa958-ae37-4e91-b3e1-1f77cd8ac339--><!--Version: Web v3.4.7-->
-------------
Labels: `type.FunctionalityBug` `severity.VeryLow`
original: wengkit1/CS2103-T-PE-Dry-run#12 | 1.0 | [PE-D][Tester D] Find functionality inconsistent with UserGuide (partial vs full word search) - ## Bug Report - ezSchedule App
### Issue Description
There is an issue with find, where finding strings that are partial in a word will return the event.
For example, `find e` will return any events which names' contain the letter e. This contradicts the userguide, which says that find finds based on **keyword**.

### Expected Behavior
I expected the app to find events based on word alone(non-case sensitive). As some words may be part of another word for example, event `eating fish` and event `fishing`, I may only want to find events that contain fish, and not care about events like fishing.
### Actual Behavior
Currently, the app's find returns any string as long as they contain any letter passed as a parameter to find.
### Steps to Reproduce
1. Open the ezSchedule app.
2. Add a tennis event
3. Use the find command `find e`
### Environment
- OS: Mac OS 13.3 Ventura
- App Version: 1.3.1
### Priority
We consider this to be a very low priority bug, as it can cause confusion for users and make it slightly unintuitive to use the app, but not a huge problem.
<!--session: 1680242278740-fb3aa958-ae37-4e91-b3e1-1f77cd8ac339--><!--Version: Web v3.4.7-->
-------------
Labels: `type.FunctionalityBug` `severity.VeryLow`
original: wengkit1/CS2103-T-PE-Dry-run#12 | priority | find functionality inconsistent with userguide partial vs full word search bug report ezschedule app issue description there is an issue with find where finding strings that are partial in a word will return the event for example find e will return any events which names contain the letter e this contradicts the userguide which says that find finds based on keyword expected behavior i expected the app to find events based on word alone non case sensitive as some words may be part of another word for example event eating fish and event fishing i may only want to find events that contain fish and not care about events like fishing actual behavior currently the app s find returns any string as long as they contain any letter passed as a parameter to find steps to reproduce open the ezschedule app add a tennis event use the find command find e environment os mac os ventura app version priority we consider this to be a very low priority bug as it can cause confusion for users and make it slightly unintuitive to use the app but not a huge problem labels type functionalitybug severity verylow original t pe dry run | 1 |
47,211 | 2,974,600,406 | IssuesEvent | 2015-07-15 02:15:09 | Reimashi/jotai | https://api.github.com/repos/Reimashi/jotai | closed | No motherboard readings for MSI P67a-g43 | auto-migrated Priority-Medium Type-Enhancement | ```
What is the expected output? What do you see instead?
details for the motherboard sesnsors
What version of the product are you using? On what operating system?
0.3.2, windows 7 x64
Please provide any additional information below.
MSI p67a-g43 motherboard
Please attach a Report created with "File / Save Report...".
```
Original issue reported on code.google.com by `spudsd...@gmail.com` on 23 May 2011 at 2:41
Attachments:
* [OpenHardwareMonitor.Report.txt](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-226/comment-0/OpenHardwareMonitor.Report.txt)
| 1.0 | No motherboard readings for MSI P67a-g43 - ```
What is the expected output? What do you see instead?
details for the motherboard sesnsors
What version of the product are you using? On what operating system?
0.3.2, windows 7 x64
Please provide any additional information below.
MSI p67a-g43 motherboard
Please attach a Report created with "File / Save Report...".
```
Original issue reported on code.google.com by `spudsd...@gmail.com` on 23 May 2011 at 2:41
Attachments:
* [OpenHardwareMonitor.Report.txt](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-226/comment-0/OpenHardwareMonitor.Report.txt)
| priority | no motherboard readings for msi what is the expected output what do you see instead details for the motherboard sesnsors what version of the product are you using on what operating system windows please provide any additional information below msi motherboard please attach a report created with file save report original issue reported on code google com by spudsd gmail com on may at attachments | 1 |
54,961 | 3,071,723,437 | IssuesEvent | 2015-08-19 13:44:35 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | opened | "Описание" в юникоде отабражается знаками вопроса в списке юзеров. | bug Component-Logic imported Priority-Medium Usability | _From [tret2...@gmail.com](https://code.google.com/u/116508191076211387118/) on November 17, 2010 12:32:47_
"Описание" в юникоде так и отабражается знаками вопроса в списке юзеров. В другом клиенте, на том же хабе все в норме.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=222_ | 1.0 | "Описание" в юникоде отабражается знаками вопроса в списке юзеров. - _From [tret2...@gmail.com](https://code.google.com/u/116508191076211387118/) on November 17, 2010 12:32:47_
"Описание" в юникоде так и отабражается знаками вопроса в списке юзеров. В другом клиенте, на том же хабе все в норме.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=222_ | priority | описание в юникоде отабражается знаками вопроса в списке юзеров from on november описание в юникоде так и отабражается знаками вопроса в списке юзеров в другом клиенте на том же хабе все в норме original issue | 1 |
576,799 | 17,095,123,504 | IssuesEvent | 2021-07-09 00:28:38 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | opened | Need tangent matrix with respect to velocities for deformable dofs | priority: medium team: dynamics type: idea | ### Background:
In the current FEM formulation for deformable bodies, we are solving for the equation
Ma − f_d(x,v) − fₑ(x) = fₑₓₜ, (1)
which then gets discretized as
F(a) = Ma − f_d(x₀,v(a, v₀)) − fₑ(x(a, v₀, x₀)) - fₑₓₜ = 0, (2)
with the Newmark scheme. We then use Newton's method to solve the discretized equations with `a` as the unknown variables. In that process, we produce the tangent matrix ∇ₐF = dF/da, which is related to, but not the same as, the tangent matrix ∇ᵥF = dF/dv consumed by the contact solver. It would be nice to be able to reuse the tangent matrix from FEM solve when Newton's method converges instead of recalculating another tangent matrix for contact.
### Proposed solution:
Change variable in the discretization of (1) and solve
F(v) = Ma(v, a₀) − f_d(x₀, v) − fₑ(x(a₀, v, x₀)) - fₑₓₜ = 0, (3)
instead with `v` as the unknown variables. The tangent matrix needed for solving (3) with Newton's method is ∇ᵥF = dF/dv, so we can reuse the tangent matrix in the contact solve when the Newton's iterations converge.
To enable this alternative formulation, we need a new class similar to `NewmarkScheme` that captures the same time discretization scheme, but uses `v` instead of `a` as the "unknown variable".
@amcastro-tri hinted at this idea in a PR review a while back. | 1.0 | Need tangent matrix with respect to velocities for deformable dofs - ### Background:
In the current FEM formulation for deformable bodies, we are solving for the equation
Ma − f_d(x,v) − fₑ(x) = fₑₓₜ, (1)
which then gets discretized as
F(a) = Ma − f_d(x₀,v(a, v₀)) − fₑ(x(a, v₀, x₀)) - fₑₓₜ = 0, (2)
with the Newmark scheme. We then use Newton's method to solve the discretized equations with `a` as the unknown variables. In that process, we produce the tangent matrix ∇ₐF = dF/da, which is related to, but not the same as, the tangent matrix ∇ᵥF = dF/dv consumed by the contact solver. It would be nice to be able to reuse the tangent matrix from FEM solve when Newton's method converges instead of recalculating another tangent matrix for contact.
### Proposed solution:
Change variable in the discretization of (1) and solve
F(v) = Ma(v, a₀) − f_d(x₀, v) − fₑ(x(a₀, v, x₀)) - fₑₓₜ = 0, (3)
instead with `v` as the unknown variables. The tangent matrix needed for solving (3) with Newton's method is ∇ᵥF = dF/dv, so we can reuse the tangent matrix in the contact solve when the Newton's iterations converge.
To enable this alternative formulation, we need a new class similar to `NewmarkScheme` that captures the same time discretization scheme, but uses `v` instead of `a` as the "unknown variable".
@amcastro-tri hinted at this idea in a PR review a while back. | priority | need tangent matrix with respect to velocities for deformable dofs background in the current fem formulation for deformable bodies we are solving for the equation ma − f d x v − fₑ x fₑₓₜ which then gets discretized as f a ma − f d x₀ v a v₀ − fₑ x a v₀ x₀ fₑₓₜ with the newmark scheme we then use newton s method to solve the discretized equations with a as the unknown variables in that process we produce the tangent matrix ∇ₐf df da which is related to but not the same as the tangent matrix ∇ᵥf df dv consumed by the contact solver it would be nice to be able to reuse the tangent matrix from fem solve when newton s method converges instead of recalculating another tangent matrix for contact proposed solution change variable in the discretization of and solve f v ma v a₀ − f d x₀ v − fₑ x a₀ v x₀ fₑₓₜ instead with v as the unknown variables the tangent matrix needed for solving with newton s method is ∇ᵥf df dv so we can reuse the tangent matrix in the contact solve when the newton s iterations converge to enable this alternative formulation we need a new class similar to newmarkscheme that captures the same time discretization scheme but uses v instead of a as the unknown variable amcastro tri hinted at this idea in a pr review a while back | 1 |
498,450 | 14,407,243,543 | IssuesEvent | 2020-12-03 21:31:15 | minio/docs | https://api.github.com/repos/minio/docs | closed | Initial Skeleton on MinIO + Private/Public Cloud | WIP priority: medium | # Summary
Create the initial skeleton structure for MinIO + Kubernetes for Private/Public Cloud
# Scope
This issue covers creating skeleton structure for the following pages:
- Getting Started (Quickstart w/ MinIO 4x4 + console, Self SIgned TLS, no KES)
- Core Concepts
- Architecture
- Erasure Coding
- Bitrot Protection
- Deploy a MinIO Tenant
- Manage a MinIO Tenant
- Connect your Application
- Security
- User Management
- Server Side Encryption
- Transport Layer Security (TLS)
- Production Recommendations
| 1.0 | Initial Skeleton on MinIO + Private/Public Cloud - # Summary
Create the initial skeleton structure for MinIO + Kubernetes for Private/Public Cloud
# Scope
This issue covers creating skeleton structure for the following pages:
- Getting Started (Quickstart w/ MinIO 4x4 + console, Self SIgned TLS, no KES)
- Core Concepts
- Architecture
- Erasure Coding
- Bitrot Protection
- Deploy a MinIO Tenant
- Manage a MinIO Tenant
- Connect your Application
- Security
- User Management
- Server Side Encryption
- Transport Layer Security (TLS)
- Production Recommendations
| priority | initial skeleton on minio private public cloud summary create the initial skeleton structure for minio kubernetes for private public cloud scope this issue covers creating skeleton structure for the following pages getting started quickstart w minio console self signed tls no kes core concepts architecture erasure coding bitrot protection deploy a minio tenant manage a minio tenant connect your application security user management server side encryption transport layer security tls production recommendations | 1 |
78,028 | 3,508,788,391 | IssuesEvent | 2016-01-08 19:31:06 | vigetlabs/trackomatic | https://api.github.com/repos/vigetlabs/trackomatic | closed | Set up docs and releases | medium priority | We should talk about how GitHub releases work, and how they interact with the Changelog. We should also probably tweak CONTRIBUTING.md so it reflects the project more specifically. | 1.0 | Set up docs and releases - We should talk about how GitHub releases work, and how they interact with the Changelog. We should also probably tweak CONTRIBUTING.md so it reflects the project more specifically. | priority | set up docs and releases we should talk about how github releases work and how they interact with the changelog we should also probably tweak contributing md so it reflects the project more specifically | 1 |
30,054 | 2,722,146,282 | IssuesEvent | 2015-04-14 00:23:46 | CruxFramework/crux-smart-faces | https://api.github.com/repos/CruxFramework/crux-smart-faces | closed | Centralize the Faces button | enhancement imported Milestone-M14-C4 Module-CruxSmartFaces Priority-Medium TargetVersion-5.3.0 | _From [flavia.jesus@triggolabs.com](https://code.google.com/u/flavia.jesus@triggolabs.com/) on March 17, 2015 10:13:06_
The Faces button is decentralized.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=637_ | 1.0 | Centralize the Faces button - _From [flavia.jesus@triggolabs.com](https://code.google.com/u/flavia.jesus@triggolabs.com/) on March 17, 2015 10:13:06_
The Faces button is decentralized.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=637_ | priority | centralize the faces button from on march the faces button is decentralized original issue | 1 |
799,426 | 28,306,669,892 | IssuesEvent | 2023-04-10 11:44:51 | PrefectHQ/prefect | https://api.github.com/repos/PrefectHQ/prefect | closed | Subflow names not shown in timeline view | bug ui great writeup priority:medium status:in-progress from:sales team:observability | ### First check
- [X] I added a descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I refreshed the page and this issue still occurred.
- [X] I checked if this issue was specific to the browser I was using by testing with a different browser.
### Bug summary
Subflow names are not shown in the timeline view, only the randomly generated flow run name is shown.
The radar chart was able to show sub flow names.

### Reproduction
[Flow Code](https://github.com/taylor-curran/flow-patterns/blob/main/flows/subflows/task_wrapped_deployments.py)
```
from prefect import flow, task
from prefect.deployments import run_deployment
from prefect.task_runners import ConcurrentTaskRunner
from pydantic import BaseModel
@task()
def upstream_task_h():
print("upstream task")
return {"h": "upstream task"}
@task()
def upstream_task_i():
print("upstream task")
return {"i": "upstream task"}
@task()
def wrapper_task_a(i, sim_failure_child_flow_a):
print("wrapper task")
a = run_deployment(
"child-flow-a/dep-child-a",
parameters={"i": i, "sim_failure_child_flow_a": sim_failure_child_flow_a},
)
return {"a": a.state.result()}
@task()
def wrapper_task_b(sim_failure_child_flow_b):
print("wrapper task")
b = run_deployment(
name="child-flow-b/dep-child-b",
parameters={"sim_failure_child_flow_b": sim_failure_child_flow_b},
)
# WARNING: We do not evaluate the result or state in this
# wrapper task decoupling this wrapper task from its
# subflow's state.
return {"b": "not flow result"}
@task()
def wrapper_task_c():
print("wrapper task")
c = run_deployment("child-flow-c/dep-child-c")
return {"c": c.state.result()}
@task()
def downstream_task_j(a):
print("downstream task")
return {"j": "downstream task"}
@task()
async def downstream_task_j(a, c, sim_failure_downstream_task_j):
if sim_failure_downstream_task_j:
raise Exception("This is a test exception")
else:
print("downstream task")
return {"j": "downstream task"}
@task()
def downstream_task_k(b="b"):
print(b)
print("downstream task")
return {"k": "downstream task"}
# ---
class SimulatedFailure(BaseModel):
child_flow_a: bool = False
child_flow_b: bool = False
downstream_task_j: bool = False
default_simulated_failure = SimulatedFailure(
child_flow_a=False, child_flow_b=False, downstream_task_j=False
)
# prefect deployment build task_wrapped_deployments.py:task_wrapped_deployments -n dep_task_wrapped -t sub-flows -t task-wrapped -t parent -a
@flow(task_runner=ConcurrentTaskRunner(), persist_result=True)
def task_wrapped_deployments(sim_failure: SimulatedFailure = default_simulated_failure):
h = upstream_task_h.submit()
i = upstream_task_i.submit()
a = wrapper_task_a.submit(i, sim_failure.child_flow_a)
b = wrapper_task_b.submit(
sim_failure_child_flow_b=sim_failure.child_flow_b, wait_for=[i]
)
c = wrapper_task_c.submit()
j = downstream_task_j.submit(a, c, sim_failure.downstream_task_j)
k = downstream_task_k.submit(wait_for=[b])
return {"j": j, "k": k}
# ---
if __name__ == "__main__":
task_wrapped_deployments(
sim_failure=SimulatedFailure(
child_flow_a=False, child_flow_b=True, downstream_task_j=False
)
)
```
### Error

### Browers
- [X] Chrome
- [ ] Firefox
- [ ] Safari
- [ ] Edge
### Prefect version
```Text
Version: 2.8.6
API version: 0.8.4
Python version: 3.10.10
Git commit: 061d877b
Built: Thu, Mar 16, 2023 2:58 PM
OS/Arch: darwin/arm64
Profile: default
Server type: cloud
```
### Additional context
_No response_ | 1.0 | Subflow names not shown in timeline view - ### First check
- [X] I added a descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I refreshed the page and this issue still occurred.
- [X] I checked if this issue was specific to the browser I was using by testing with a different browser.
### Bug summary
Subflow names are not shown in the timeline view, only the randomly generated flow run name is shown.
The radar chart was able to show sub flow names.

### Reproduction
[Flow Code](https://github.com/taylor-curran/flow-patterns/blob/main/flows/subflows/task_wrapped_deployments.py)
```
from prefect import flow, task
from prefect.deployments import run_deployment
from prefect.task_runners import ConcurrentTaskRunner
from pydantic import BaseModel
@task()
def upstream_task_h():
print("upstream task")
return {"h": "upstream task"}
@task()
def upstream_task_i():
print("upstream task")
return {"i": "upstream task"}
@task()
def wrapper_task_a(i, sim_failure_child_flow_a):
print("wrapper task")
a = run_deployment(
"child-flow-a/dep-child-a",
parameters={"i": i, "sim_failure_child_flow_a": sim_failure_child_flow_a},
)
return {"a": a.state.result()}
@task()
def wrapper_task_b(sim_failure_child_flow_b):
print("wrapper task")
b = run_deployment(
name="child-flow-b/dep-child-b",
parameters={"sim_failure_child_flow_b": sim_failure_child_flow_b},
)
# WARNING: We do not evaluate the result or state in this
# wrapper task decoupling this wrapper task from its
# subflow's state.
return {"b": "not flow result"}
@task()
def wrapper_task_c():
print("wrapper task")
c = run_deployment("child-flow-c/dep-child-c")
return {"c": c.state.result()}
@task()
def downstream_task_j(a):
print("downstream task")
return {"j": "downstream task"}
@task()
async def downstream_task_j(a, c, sim_failure_downstream_task_j):
if sim_failure_downstream_task_j:
raise Exception("This is a test exception")
else:
print("downstream task")
return {"j": "downstream task"}
@task()
def downstream_task_k(b="b"):
print(b)
print("downstream task")
return {"k": "downstream task"}
# ---
class SimulatedFailure(BaseModel):
child_flow_a: bool = False
child_flow_b: bool = False
downstream_task_j: bool = False
default_simulated_failure = SimulatedFailure(
child_flow_a=False, child_flow_b=False, downstream_task_j=False
)
# prefect deployment build task_wrapped_deployments.py:task_wrapped_deployments -n dep_task_wrapped -t sub-flows -t task-wrapped -t parent -a
@flow(task_runner=ConcurrentTaskRunner(), persist_result=True)
def task_wrapped_deployments(sim_failure: SimulatedFailure = default_simulated_failure):
h = upstream_task_h.submit()
i = upstream_task_i.submit()
a = wrapper_task_a.submit(i, sim_failure.child_flow_a)
b = wrapper_task_b.submit(
sim_failure_child_flow_b=sim_failure.child_flow_b, wait_for=[i]
)
c = wrapper_task_c.submit()
j = downstream_task_j.submit(a, c, sim_failure.downstream_task_j)
k = downstream_task_k.submit(wait_for=[b])
return {"j": j, "k": k}
# ---
if __name__ == "__main__":
task_wrapped_deployments(
sim_failure=SimulatedFailure(
child_flow_a=False, child_flow_b=True, downstream_task_j=False
)
)
```
### Error

### Browers
- [X] Chrome
- [ ] Firefox
- [ ] Safari
- [ ] Edge
### Prefect version
```Text
Version: 2.8.6
API version: 0.8.4
Python version: 3.10.10
Git commit: 061d877b
Built: Thu, Mar 16, 2023 2:58 PM
OS/Arch: darwin/arm64
Profile: default
Server type: cloud
```
### Additional context
_No response_ | priority | subflow names not shown in timeline view first check i added a descriptive title to this issue i used the github search to find a similar issue and didn t find it i refreshed the page and this issue still occurred i checked if this issue was specific to the browser i was using by testing with a different browser bug summary subflow names are not shown in the timeline view only the randomly generated flow run name is shown the radar chart was able to show sub flow names reproduction from prefect import flow task from prefect deployments import run deployment from prefect task runners import concurrenttaskrunner from pydantic import basemodel task def upstream task h print upstream task return h upstream task task def upstream task i print upstream task return i upstream task task def wrapper task a i sim failure child flow a print wrapper task a run deployment child flow a dep child a parameters i i sim failure child flow a sim failure child flow a return a a state result task def wrapper task b sim failure child flow b print wrapper task b run deployment name child flow b dep child b parameters sim failure child flow b sim failure child flow b warning we do not evaluate the result or state in this wrapper task decoupling this wrapper task from its subflow s state return b not flow result task def wrapper task c print wrapper task c run deployment child flow c dep child c return c c state result task def downstream task j a print downstream task return j downstream task task async def downstream task j a c sim failure downstream task j if sim failure downstream task j raise exception this is a test exception else print downstream task return j downstream task task def downstream task k b b print b print downstream task return k downstream task class simulatedfailure basemodel child flow a bool false child flow b bool false downstream task j bool false default simulated failure simulatedfailure child flow a false child flow b false downstream task j false prefect deployment build task wrapped deployments py task wrapped deployments n dep task wrapped t sub flows t task wrapped t parent a flow task runner concurrenttaskrunner persist result true def task wrapped deployments sim failure simulatedfailure default simulated failure h upstream task h submit i upstream task i submit a wrapper task a submit i sim failure child flow a b wrapper task b submit sim failure child flow b sim failure child flow b wait for c wrapper task c submit j downstream task j submit a c sim failure downstream task j k downstream task k submit wait for return j j k k if name main task wrapped deployments sim failure simulatedfailure child flow a false child flow b true downstream task j false error browers chrome firefox safari edge prefect version text version api version python version git commit built thu mar pm os arch darwin profile default server type cloud additional context no response | 1 |
468,095 | 13,461,639,873 | IssuesEvent | 2020-09-09 15:03:14 | hubmapconsortium/ccf-ui | https://api.github.com/repos/hubmapconsortium/ccf-ui | closed | As a researcher I need to in the Ingestion Portal modal that contains the app have a means to close the modal so that I can return to the Portal to add another registration or other task | Priority: Medium High SP:1 user-story | Ensure the researcher can get back to the Ingestion Portal by:
* Providing a BACK button in the Logo Bar
* Closing the modal when the user clicks the BACK button and revealing the Ingestion Portal that the modal overlays
* Providing a chevron icon that points to the left and the text "BACK"
Original ID: 3 | 1.0 | As a researcher I need to in the Ingestion Portal modal that contains the app have a means to close the modal so that I can return to the Portal to add another registration or other task - Ensure the researcher can get back to the Ingestion Portal by:
* Providing a BACK button in the Logo Bar
* Closing the modal when the user clicks the BACK button and revealing the Ingestion Portal that the modal overlays
* Providing a chevron icon that points to the left and the text "BACK"
Original ID: 3 | priority | as a researcher i need to in the ingestion portal modal that contains the app have a means to close the modal so that i can return to the portal to add another registration or other task ensure the researcher can get back to the ingestion portal by providing a back button in the logo bar closing the modal when the user clicks the back button and revealing the ingestion portal that the modal overlays providing a chevron icon that points to the left and the text back original id | 1 |
238,206 | 7,776,362,661 | IssuesEvent | 2018-06-05 07:48:32 | strapi/strapi | https://api.github.com/repos/strapi/strapi | closed | Relations with 1-N to same content-type becomes 1-1 relation | priority: medium status: confirmed 👍 type: bug 🐛 | **Informations**
- **Node.js version**: 9.9.0
- **npm version**: 5.6.0
- **Strapi version**: 3.0.0-alpha.12.1.3
- **Database**: MongoDB
- **Operating system**: Windows 10
**What is the current behavior?**
Adding a relation from a content-type to itself does not work for 1-N relations. After creation of the content-type, the server restarts and clicking on the content-type shows that the relation is indeed 1-N. But when creating a new object of that content-type, the content-type becomes a 1-1 relation for no apparant reason. N-N relations also work like they should.
**Steps to reproduce the problem**
1) create a new content-type with a name and a 1-N relation with itself
2) navigate to the view of all content-types and add a new object of the content-type with a name, and save it
3) add a second object of that content-type in similar fashion
4) add a third object of that content-type and attempt to add the other 2 objects of that content-type to the object. The dropdown list now only allows to choose 1 of the 2 other objects, not both.
5) returning to the content-type settings, the content-type magically became a 1-1 relation
**What is the expected behavior?**
1-N relations should get saved and not change to 1-1 when creating a new object
**Suggested solutions**
n.a.
- [x] I'm sure that this feature hasn't already been referenced.
| 1.0 | Relations with 1-N to same content-type becomes 1-1 relation - **Informations**
- **Node.js version**: 9.9.0
- **npm version**: 5.6.0
- **Strapi version**: 3.0.0-alpha.12.1.3
- **Database**: MongoDB
- **Operating system**: Windows 10
**What is the current behavior?**
Adding a relation from a content-type to itself does not work for 1-N relations. After creation of the content-type, the server restarts and clicking on the content-type shows that the relation is indeed 1-N. But when creating a new object of that content-type, the content-type becomes a 1-1 relation for no apparant reason. N-N relations also work like they should.
**Steps to reproduce the problem**
1) create a new content-type with a name and a 1-N relation with itself
2) navigate to the view of all content-types and add a new object of the content-type with a name, and save it
3) add a second object of that content-type in similar fashion
4) add a third object of that content-type and attempt to add the other 2 objects of that content-type to the object. The dropdown list now only allows to choose 1 of the 2 other objects, not both.
5) returning to the content-type settings, the content-type magically became a 1-1 relation
**What is the expected behavior?**
1-N relations should get saved and not change to 1-1 when creating a new object
**Suggested solutions**
n.a.
- [x] I'm sure that this feature hasn't already been referenced.
| priority | relations with n to same content type becomes relation informations node js version npm version strapi version alpha database mongodb operating system windows what is the current behavior adding a relation from a content type to itself does not work for n relations after creation of the content type the server restarts and clicking on the content type shows that the relation is indeed n but when creating a new object of that content type the content type becomes a relation for no apparant reason n n relations also work like they should steps to reproduce the problem create a new content type with a name and a n relation with itself navigate to the view of all content types and add a new object of the content type with a name and save it add a second object of that content type in similar fashion add a third object of that content type and attempt to add the other objects of that content type to the object the dropdown list now only allows to choose of the other objects not both returning to the content type settings the content type magically became a relation what is the expected behavior n relations should get saved and not change to when creating a new object suggested solutions n a i m sure that this feature hasn t already been referenced | 1 |
426,822 | 12,379,906,568 | IssuesEvent | 2020-05-19 13:15:00 | ooni/probe-engine | https://api.github.com/repos/ooni/probe-engine | closed | jsonapi: disable logging body by default to avoid leaks | effort/S enhancement ooni/probe-engine priority/medium | When finalizing https://github.com/ooni/spec/pull/175, it occurred to me that probably want to guarantee that we aren't leaking information exchanged with orchestra, especially when this information is guaranteed to contain secrets. This is not an issue for now, because we are not using the jsonapi to exchange any secrets. But this may change as we evolve the `tor` experiment. | 1.0 | jsonapi: disable logging body by default to avoid leaks - When finalizing https://github.com/ooni/spec/pull/175, it occurred to me that probably want to guarantee that we aren't leaking information exchanged with orchestra, especially when this information is guaranteed to contain secrets. This is not an issue for now, because we are not using the jsonapi to exchange any secrets. But this may change as we evolve the `tor` experiment. | priority | jsonapi disable logging body by default to avoid leaks when finalizing it occurred to me that probably want to guarantee that we aren t leaking information exchanged with orchestra especially when this information is guaranteed to contain secrets this is not an issue for now because we are not using the jsonapi to exchange any secrets but this may change as we evolve the tor experiment | 1 |
128,377 | 5,060,459,074 | IssuesEvent | 2016-12-22 12:03:00 | HBHWoolacotts/RPii | https://api.github.com/repos/HBHWoolacotts/RPii | opened | Unable to remove email address from this customer's account - keeps coming back! | Label: General RP Bugs and Support Priority - Medium | Every time we remove it, it comes back!
Also, the window doesn't scale correctly as the VAT flag field is overlapping the save & close buttons at the bottom.

| 1.0 | Unable to remove email address from this customer's account - keeps coming back! - Every time we remove it, it comes back!
Also, the window doesn't scale correctly as the VAT flag field is overlapping the save & close buttons at the bottom.

| priority | unable to remove email address from this customer s account keeps coming back every time we remove it it comes back also the window doesn t scale correctly as the vat flag field is overlapping the save close buttons at the bottom | 1 |
608,152 | 18,816,551,070 | IssuesEvent | 2021-11-10 00:22:54 | ouch-org/ouch | https://api.github.com/repos/ouch-org/ouch | closed | Add nested folders to automated compression and decompression tests. | medium priority | Currently, we only test archive formats with multiple regular random files, there's no nested folders testing, so for that formats the tests need to be refactored. | 1.0 | Add nested folders to automated compression and decompression tests. - Currently, we only test archive formats with multiple regular random files, there's no nested folders testing, so for that formats the tests need to be refactored. | priority | add nested folders to automated compression and decompression tests currently we only test archive formats with multiple regular random files there s no nested folders testing so for that formats the tests need to be refactored | 1 |
454,800 | 13,107,693,432 | IssuesEvent | 2020-08-04 15:38:05 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | opened | Remove “Mark task complete” as an available action for BvaDispatchTasks | Feature: generic-queue Priority: Medium Product: caseflow-queue Stakeholder: BVA Team: Echo 🐬 Type: Enhancement | `BvaDispatchTask`s are automatically marked complete when an appeal is outcoded through IDT. The the only way a `BvaDispatchTask` can be considered "complete" is if the appeal is dispatched by outcoding the appeal through IDT. Currently, `MARK_COMPLETE` is an available action for the assignee of a `BvaDispatchTask`. This should not be provided as an option so that users cannot manually close their tasks without completing the work.
### Relevant
[Batteam thread](https://dsva.slack.com/archives/CHX8FMP28/p1596549215130500) that spawned this
Where we determine actions available for these tasks
https://github.com/department-of-veterans-affairs/caseflow/blob/6256e1107c81c0df73ff7466bd0b5e533a797f63/app/models/tasks/bva_dispatch_task.rb#L8-L18
### AC
- [ ] `BvaDispatchTask`s cannot be marked complete by anyone | 1.0 | Remove “Mark task complete” as an available action for BvaDispatchTasks - `BvaDispatchTask`s are automatically marked complete when an appeal is outcoded through IDT. The the only way a `BvaDispatchTask` can be considered "complete" is if the appeal is dispatched by outcoding the appeal through IDT. Currently, `MARK_COMPLETE` is an available action for the assignee of a `BvaDispatchTask`. This should not be provided as an option so that users cannot manually close their tasks without completing the work.
### Relevant
[Batteam thread](https://dsva.slack.com/archives/CHX8FMP28/p1596549215130500) that spawned this
Where we determine actions available for these tasks
https://github.com/department-of-veterans-affairs/caseflow/blob/6256e1107c81c0df73ff7466bd0b5e533a797f63/app/models/tasks/bva_dispatch_task.rb#L8-L18
### AC
- [ ] `BvaDispatchTask`s cannot be marked complete by anyone | priority | remove “mark task complete” as an available action for bvadispatchtasks bvadispatchtask s are automatically marked complete when an appeal is outcoded through idt the the only way a bvadispatchtask can be considered complete is if the appeal is dispatched by outcoding the appeal through idt currently mark complete is an available action for the assignee of a bvadispatchtask this should not be provided as an option so that users cannot manually close their tasks without completing the work relevant that spawned this where we determine actions available for these tasks ac bvadispatchtask s cannot be marked complete by anyone | 1 |
78,297 | 3,509,559,163 | IssuesEvent | 2016-01-08 23:27:37 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Bugged sit (BB #1067) | migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** Ezopko
**Original Date:** 18.07.2015 20:05:39 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1067
<hr>
Bugged sit at all chair
http://ctrlv.cz/CQoh | 1.0 | Bugged sit (BB #1067) - This issue was migrated from bitbucket.
**Original Reporter:** Ezopko
**Original Date:** 18.07.2015 20:05:39 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1067
<hr>
Bugged sit at all chair
http://ctrlv.cz/CQoh | priority | bugged sit bb this issue was migrated from bitbucket original reporter ezopko original date gmt original priority major original type bug original state resolved direct link bugged sit at all chair | 1 |
603,715 | 18,670,634,871 | IssuesEvent | 2021-10-30 16:38:49 | fslaborg/RProvider | https://api.github.com/repos/fslaborg/RProvider | closed | Project build when referencing RProvider hangs forever (ubuntu-latest - github actions) | Type: Bug Status: Available Priority: Medium | **Describe the bug**
Projects that reference RProvider fail to build when using Ubuntu 20.04 as the OS within Github Actions using FAKE. When dotnet build is called for the specific project, it hangs forever.
**To Reproduce**
Steps to reproduce the behavior:
1. Enable ubuntu-latest in the github actions script for RProvider itself or Deedle (see fslaborg/deedle#538)
2. The build will hang forever on Ubuntu agent but complete on a macos or windows agent. For Deedle, this is the RProvider plugin project; for RProvider, this is the test project.
**Expected behavior**
No hangs and successful build.
**Environment (please complete the following information):**
- OS: Github Actions - ubuntu latest (20.04)
- OS Version: 20.04
- Using in script or library: library
- RProvider 2.0.1-beta2
- Installed R Version 4.0.2
| 1.0 | Project build when referencing RProvider hangs forever (ubuntu-latest - github actions) - **Describe the bug**
Projects that reference RProvider fail to build when using Ubuntu 20.04 as the OS within Github Actions using FAKE. When dotnet build is called for the specific project, it hangs forever.
**To Reproduce**
Steps to reproduce the behavior:
1. Enable ubuntu-latest in the github actions script for RProvider itself or Deedle (see fslaborg/deedle#538)
2. The build will hang forever on Ubuntu agent but complete on a macos or windows agent. For Deedle, this is the RProvider plugin project; for RProvider, this is the test project.
**Expected behavior**
No hangs and successful build.
**Environment (please complete the following information):**
- OS: Github Actions - ubuntu latest (20.04)
- OS Version: 20.04
- Using in script or library: library
- RProvider 2.0.1-beta2
- Installed R Version 4.0.2
| priority | project build when referencing rprovider hangs forever ubuntu latest github actions describe the bug projects that reference rprovider fail to build when using ubuntu as the os within github actions using fake when dotnet build is called for the specific project it hangs forever to reproduce steps to reproduce the behavior enable ubuntu latest in the github actions script for rprovider itself or deedle see fslaborg deedle the build will hang forever on ubuntu agent but complete on a macos or windows agent for deedle this is the rprovider plugin project for rprovider this is the test project expected behavior no hangs and successful build environment please complete the following information os github actions ubuntu latest os version using in script or library library rprovider installed r version | 1 |
563,578 | 16,701,118,520 | IssuesEvent | 2021-06-09 02:38:44 | unicode-org/icu4x | https://api.github.com/repos/unicode-org/icu4x | opened | Add PluralRules GetKeywords functionality | S-medium T-core discuss-priority | ICU4X currently has no functionality to return an enumeration of keywords that are supported by a PluralRules object for a given locale and `Cardinal`/`Ordinal` setting.
The equivalent call that we currently use in SpiderMonkey is `uplrules_getKeywords()`
https://unicode-org.github.io/icu-docs/apidoc/dev/icu4c/upluralrules_8h.html#a5f8a89ec2b58323e9f96ac0bf2c13ab5 | 1.0 | Add PluralRules GetKeywords functionality - ICU4X currently has no functionality to return an enumeration of keywords that are supported by a PluralRules object for a given locale and `Cardinal`/`Ordinal` setting.
The equivalent call that we currently use in SpiderMonkey is `uplrules_getKeywords()`
https://unicode-org.github.io/icu-docs/apidoc/dev/icu4c/upluralrules_8h.html#a5f8a89ec2b58323e9f96ac0bf2c13ab5 | priority | add pluralrules getkeywords functionality currently has no functionality to return an enumeration of keywords that are supported by a pluralrules object for a given locale and cardinal ordinal setting the equivalent call that we currently use in spidermonkey is uplrules getkeywords | 1 |
623,745 | 19,677,461,295 | IssuesEvent | 2022-01-11 13:48:37 | haydarsinan/SWE573-2021-Fall | https://api.github.com/repos/haydarsinan/SWE573-2021-Fall | closed | Geolocation Research for Services and Events | High Priority Medium Work Learning Research | Closer events and services should be shown to users firstly depending on their location. Therefore, geolocation feature should be implemented and researched.
| 1.0 | Geolocation Research for Services and Events - Closer events and services should be shown to users firstly depending on their location. Therefore, geolocation feature should be implemented and researched.
| priority | geolocation research for services and events closer events and services should be shown to users firstly depending on their location therefore geolocation feature should be implemented and researched | 1 |
696,688 | 23,911,663,742 | IssuesEvent | 2022-09-09 08:47:42 | Cheos137/ArmorpointsPlusplus | https://api.github.com/repos/Cheos137/ArmorpointsPlusplus | closed | [Feature Request]: option to disable compatibility warnings | feature request minecraft: all supported future update priority: medium | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description
Add a new config option to hide messages warning about incompatibility with other mods.
### Solution
Add a new config option to hide messages warning about incompatibility with other mods.
### Alternatives
_No response_
### Additional Context
_No response_ | 1.0 | [Feature Request]: option to disable compatibility warnings - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description
Add a new config option to hide messages warning about incompatibility with other mods.
### Solution
Add a new config option to hide messages warning about incompatibility with other mods.
### Alternatives
_No response_
### Additional Context
_No response_ | priority | option to disable compatibility warnings is there an existing issue for this i have searched the existing issues description add a new config option to hide messages warning about incompatibility with other mods solution add a new config option to hide messages warning about incompatibility with other mods alternatives no response additional context no response | 1 |
351,344 | 10,516,076,161 | IssuesEvent | 2019-09-28 14:55:48 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | closed | Characters don't get dryad blood | :beetle: bug :beetle: :grey_exclamation: priority medium | **Mod Version**
6fc3865c
**What expansions do you have installed?**
All
**Please explain your issue in as much detail as possible:**
Characters don't get dryad blood
**Steps to reproduce the issue:**
Marry dryad to centaur or frost nymph
**Upload an attachment below: .zip of your save, or screenshots:**
<details>
<summary>Click to expand</summary>


</details> | 1.0 | Characters don't get dryad blood - **Mod Version**
6fc3865c
**What expansions do you have installed?**
All
**Please explain your issue in as much detail as possible:**
Characters don't get dryad blood
**Steps to reproduce the issue:**
Marry dryad to centaur or frost nymph
**Upload an attachment below: .zip of your save, or screenshots:**
<details>
<summary>Click to expand</summary>


</details> | priority | characters don t get dryad blood mod version what expansions do you have installed all please explain your issue in as much detail as possible characters don t get dryad blood steps to reproduce the issue marry dryad to centaur or frost nymph upload an attachment below zip of your save or screenshots click to expand | 1 |
68,358 | 3,286,691,490 | IssuesEvent | 2015-10-29 05:03:41 | cs2103aug2015-w10-4j/main | https://api.github.com/repos/cs2103aug2015-w10-4j/main | closed | As an advanced user, I can reorder the command given to create a task | priority.medium type.story | ...so that I can focus more on the task rather than how to input the task
| 1.0 | As an advanced user, I can reorder the command given to create a task - ...so that I can focus more on the task rather than how to input the task
| priority | as an advanced user i can reorder the command given to create a task so that i can focus more on the task rather than how to input the task | 1 |
208,752 | 7,158,110,450 | IssuesEvent | 2018-01-26 22:43:28 | IfyAniefuna/experiment_metadata | https://api.github.com/repos/IfyAniefuna/experiment_metadata | opened | Include nutrient source units (g/L, etc) for within generated metadata | enhancement medium priority | The units may change over time, therefore the units being asked for by the current webapp needs to be automatically included in the generated metadata,
example: user enters Glucose of 2 (g/L) in the webapp. The generated metadata form has the following:
"carbon-source","Glucose(2g/L)" | 1.0 | Include nutrient source units (g/L, etc) for within generated metadata - The units may change over time, therefore the units being asked for by the current webapp needs to be automatically included in the generated metadata,
example: user enters Glucose of 2 (g/L) in the webapp. The generated metadata form has the following:
"carbon-source","Glucose(2g/L)" | priority | include nutrient source units g l etc for within generated metadata the units may change over time therefore the units being asked for by the current webapp needs to be automatically included in the generated metadata example user enters glucose of g l in the webapp the generated metadata form has the following carbon source glucose l | 1 |
537,364 | 15,727,813,206 | IssuesEvent | 2021-03-29 13:07:47 | SteamedHams722/automatic-octopus | https://api.github.com/repos/SteamedHams722/automatic-octopus | closed | Use Django instead of Heroku for Postgres | Medium Priority | Why
It may make more sense to just use Django for hosting the Postgres db. The heroku hobby versions are very limiting and won't be able to handle the regular data loads.
How
- Investigate how Django will handle this data.
- Set-up a postgres db that can run in Django
- Get rid of the Heroku db. | 1.0 | Use Django instead of Heroku for Postgres - Why
It may make more sense to just use Django for hosting the Postgres db. The heroku hobby versions are very limiting and won't be able to handle the regular data loads.
How
- Investigate how Django will handle this data.
- Set-up a postgres db that can run in Django
- Get rid of the Heroku db. | priority | use django instead of heroku for postgres why it may make more sense to just use django for hosting the postgres db the heroku hobby versions are very limiting and won t be able to handle the regular data loads how investigate how django will handle this data set up a postgres db that can run in django get rid of the heroku db | 1 |
32,672 | 2,757,596,123 | IssuesEvent | 2015-04-27 15:40:15 | pydata/numexpr | https://api.github.com/repos/pydata/numexpr | closed | Numexpr query with empty string does not work (discovered using pytables) | bug imported Priority-Medium | _From [daniel.h...@fmi.ch](https://code.google.com/u/113211097457152931757/) on December 03, 2013 12:37:10_
See https://github.com/PyTables/PyTables/issues/184
_Original issue: http://code.google.com/p/numexpr/issues/detail?id=121_ | 1.0 | Numexpr query with empty string does not work (discovered using pytables) - _From [daniel.h...@fmi.ch](https://code.google.com/u/113211097457152931757/) on December 03, 2013 12:37:10_
See https://github.com/PyTables/PyTables/issues/184
_Original issue: http://code.google.com/p/numexpr/issues/detail?id=121_ | priority | numexpr query with empty string does not work discovered using pytables from on december see original issue | 1 |
765,922 | 26,865,725,822 | IssuesEvent | 2023-02-03 23:24:41 | clt313/SuperballVR | https://api.github.com/repos/clt313/SuperballVR | closed | Update audio to use singleton pattern | priority: medium | Rather than juggling audio sources like we have been, use an AudioManager instead.
Useful video: https://www.youtube.com/watch?v=6OT43pvUyfY&ab_channel=Brackeys | 1.0 | Update audio to use singleton pattern - Rather than juggling audio sources like we have been, use an AudioManager instead.
Useful video: https://www.youtube.com/watch?v=6OT43pvUyfY&ab_channel=Brackeys | priority | update audio to use singleton pattern rather than juggling audio sources like we have been use an audiomanager instead useful video | 1 |
321,342 | 9,797,389,890 | IssuesEvent | 2019-06-11 09:50:25 | nhsconnect/gpconnect-provider-testing | https://api.github.com/repos/nhsconnect/gpconnect-provider-testing | closed | Default date range if only a start or end date parameter passed in getCareRecord | 0.7.0 HTML Medium priority | the scenario where only a start or an end date need to be considered and the GP Connect program needs to update the specification to cater for the possibility of someone only sending a start date or an end date
date banner section uplifted to add requirements on expected date range behaviours
| 1.0 | Default date range if only a start or end date parameter passed in getCareRecord - the scenario where only a start or an end date need to be considered and the GP Connect program needs to update the specification to cater for the possibility of someone only sending a start date or an end date
date banner section uplifted to add requirements on expected date range behaviours
| priority | default date range if only a start or end date parameter passed in getcarerecord the scenario where only a start or an end date need to be considered and the gp connect program needs to update the specification to cater for the possibility of someone only sending a start date or an end date date banner section uplifted to add requirements on expected date range behaviours | 1 |
428,536 | 12,412,966,962 | IssuesEvent | 2020-05-22 11:42:49 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Law spam | Priority: Medium Status: Fixed |
2. Spam from law notification in chat. Step to reproduce:
- create law like this:

- start to place something 3 times. I have very strange notifications (I have civics debug off):

- if you place it very fast it can appear here:

3. Notification from executive office. Step to reproduce:
- create executive action like:

- /civics winelection. This notification doesn't always appear:

- I can't see it in chat only in tooltip:

| 1.0 | Law spam -
2. Spam from law notification in chat. Step to reproduce:
- create law like this:

- start to place something 3 times. I have very strange notifications (I have civics debug off):

- if you place it very fast it can appear here:

3. Notification from executive office. Step to reproduce:
- create executive action like:

- /civics winelection. This notification doesn't always appear:

- I can't see it in chat only in tooltip:

| priority | law spam spam from law notification in chat step to reproduce create law like this start to place something times i have very strange notifications i have civics debug off if you place it very fast it can appear here notification from executive office step to reproduce create executive action like civics winelection this notification doesn t always appear i can t see it in chat only in tooltip | 1 |
645,145 | 20,996,211,346 | IssuesEvent | 2022-03-29 13:42:31 | HabitRPG/habitica-android | https://api.github.com/repos/HabitRPG/habitica-android | closed | External links not working from Guild pages | Type: Bug Help wanted Priority: medium Status: pr: ready for review | User report: (also reproduced by @voidbytes )
"Device: LGE LM-V600
Android Version: 30
AppVersion: Version 3.4.1.1 (3090)
User ID: d2656f22-44e5-4239-ad66-e0d40305090f
Level: 5
Class: warrior
Is in Inn: false
Uses Costume: false
Custom Day Start: 0
Timezone Offset: 300
Details:
Unable to click external links in guild pages including wiki pages. Verified with second user that the issue exists. "
| 1.0 | External links not working from Guild pages - User report: (also reproduced by @voidbytes )
"Device: LGE LM-V600
Android Version: 30
AppVersion: Version 3.4.1.1 (3090)
User ID: d2656f22-44e5-4239-ad66-e0d40305090f
Level: 5
Class: warrior
Is in Inn: false
Uses Costume: false
Custom Day Start: 0
Timezone Offset: 300
Details:
Unable to click external links in guild pages including wiki pages. Verified with second user that the issue exists. "
| priority | external links not working from guild pages user report also reproduced by voidbytes device lge lm android version appversion version user id level class warrior is in inn false uses costume false custom day start timezone offset details unable to click external links in guild pages including wiki pages verified with second user that the issue exists | 1 |
358,733 | 10,631,790,535 | IssuesEvent | 2019-10-15 09:04:56 | ooni/probe-android | https://api.github.com/repos/ooni/probe-android | closed | Publish app apk to github release | enhancement priority/medium | We would like to add a direct link to download the OONI Probe app apk on Android to our website (see: https://github.com/TheTorProject/ooni-web/issues/277).
It would be great if during the release process we add an extra step about uploading the built apk to the relevant github release. | 1.0 | Publish app apk to github release - We would like to add a direct link to download the OONI Probe app apk on Android to our website (see: https://github.com/TheTorProject/ooni-web/issues/277).
It would be great if during the release process we add an extra step about uploading the built apk to the relevant github release. | priority | publish app apk to github release we would like to add a direct link to download the ooni probe app apk on android to our website see it would be great if during the release process we add an extra step about uploading the built apk to the relevant github release | 1 |
643,909 | 20,961,610,727 | IssuesEvent | 2022-03-27 21:49:08 | abedmaatalla/sipdroid | https://api.github.com/repos/abedmaatalla/sipdroid | closed | Allow customized "User-Agent" header | Priority-Medium auto-migrated Type-Enhancement | ```
Some providers use "User-Agent" header to validate registration. Would it
possible to allow customized user agent when setting up a line?
```
Original issue reported on code.google.com by `zhang...@gmail.com` on 26 Nov 2012 at 7:04
| 1.0 | Allow customized "User-Agent" header - ```
Some providers use "User-Agent" header to validate registration. Would it
possible to allow customized user agent when setting up a line?
```
Original issue reported on code.google.com by `zhang...@gmail.com` on 26 Nov 2012 at 7:04
| priority | allow customized user agent header some providers use user agent header to validate registration would it possible to allow customized user agent when setting up a line original issue reported on code google com by zhang gmail com on nov at | 1 |
496,241 | 14,343,581,275 | IssuesEvent | 2020-11-28 10:11:39 | greasemonkey123/Redefine-A | https://api.github.com/repos/greasemonkey123/Redefine-A | opened | Better group admin settings | Medium priority Redefine:A enhancement | ability to add multiple groups to the admin list and a way to change the rank tolerance type from ">=" to "=", etc.
Suggested by ShhhhhhhhhhhBro.
- [ ] Loader edit.
- [ ] Module edit.
- [ ] Allow easily identifiable settings. | 1.0 | Better group admin settings - ability to add multiple groups to the admin list and a way to change the rank tolerance type from ">=" to "=", etc.
Suggested by ShhhhhhhhhhhBro.
- [ ] Loader edit.
- [ ] Module edit.
- [ ] Allow easily identifiable settings. | priority | better group admin settings ability to add multiple groups to the admin list and a way to change the rank tolerance type from to etc suggested by shhhhhhhhhhhbro loader edit module edit allow easily identifiable settings | 1 |
597,445 | 18,163,665,677 | IssuesEvent | 2021-09-27 12:33:52 | Undecided-Discord/Oshavery-Server | https://api.github.com/repos/Undecided-Discord/Oshavery-Server | opened | WebHook対応ほしい、ほしくない? | enhancement Priority: medium | ## 概要
<!-- 概要を書いてください -->
とりあえずGitHubのWebHookがほしいです
## なぜIssueを作ったか
<!-- Issueを作ろうと思った理由 -->
開発の通知がほしいので...
## 参考
<!-- 参考になるものがあれば貼ってください -->

## 必要なこと
<!-- わかるのであればタスク一覧を書いてください -->
- [ ] embedの実装 | 1.0 | WebHook対応ほしい、ほしくない? - ## 概要
<!-- 概要を書いてください -->
とりあえずGitHubのWebHookがほしいです
## なぜIssueを作ったか
<!-- Issueを作ろうと思った理由 -->
開発の通知がほしいので...
## 参考
<!-- 参考になるものがあれば貼ってください -->

## 必要なこと
<!-- わかるのであればタスク一覧を書いてください -->
- [ ] embedの実装 | priority | webhook対応ほしい、ほしくない? 概要 とりあえずgithubのwebhookがほしいです なぜissueを作ったか 開発の通知がほしいので 参考 必要なこと embedの実装 | 1 |
24,385 | 2,667,388,995 | IssuesEvent | 2015-03-22 15:25:34 | NewCreature/EOF | https://api.github.com/repos/NewCreature/EOF | opened | Allow MIDI playback | enhancement imported Priority-Medium | _From [raynebc](https://code.google.com/u/raynebc/) on June 23, 2010 12:22:14_
I was thinking about how work-intensive it is to prepare a new set of vocal tone samples. It seems to be that since Allegro has built-in support for MIDI, it may be worthwhile to allow it to be a user-option to try to initialize the system's default MIDI device so that EOF's output can optionally be sent to a software/hardware synthesizer.
Besides just sending Note On and Off data, we'll need to consider how to allow the user to specify which synth tone to use. When I created the piano tones for EOF, I learned a bit about this process, but haven't mastered it fully. I'm looking into the creation of Cakewalk instrument definition files, which seem to be the most common format of instrument definitions. Anvil Studio will allow you to define information for its internal use if the MIDI package is purchased ($20 or so, but I don't know yet if it will allow an instrument definition file to be created). I'm willing to get this to test creating definition files, and with any luck could write functions to import such definitions into EOF.
This would allow for a good deal of flexibility with customizing the sound/volume of vocal tones. This may also be a good way to begin MIDI integration, should EOF be given the ability to accept MIDI input in the future (ie. piano record mode, if FoF gains the ability to play keyboard tracks).
Besides vocal tones, it could allow the user to supply their own clap, metronome, etc. without having to provide recorded audio samples to EOF.
_Original issue: http://code.google.com/p/editor-on-fire/issues/detail?id=121_ | 1.0 | Allow MIDI playback - _From [raynebc](https://code.google.com/u/raynebc/) on June 23, 2010 12:22:14_
I was thinking about how work-intensive it is to prepare a new set of vocal tone samples. It seems to be that since Allegro has built-in support for MIDI, it may be worthwhile to allow it to be a user-option to try to initialize the system's default MIDI device so that EOF's output can optionally be sent to a software/hardware synthesizer.
Besides just sending Note On and Off data, we'll need to consider how to allow the user to specify which synth tone to use. When I created the piano tones for EOF, I learned a bit about this process, but haven't mastered it fully. I'm looking into the creation of Cakewalk instrument definition files, which seem to be the most common format of instrument definitions. Anvil Studio will allow you to define information for its internal use if the MIDI package is purchased ($20 or so, but I don't know yet if it will allow an instrument definition file to be created). I'm willing to get this to test creating definition files, and with any luck could write functions to import such definitions into EOF.
This would allow for a good deal of flexibility with customizing the sound/volume of vocal tones. This may also be a good way to begin MIDI integration, should EOF be given the ability to accept MIDI input in the future (ie. piano record mode, if FoF gains the ability to play keyboard tracks).
Besides vocal tones, it could allow the user to supply their own clap, metronome, etc. without having to provide recorded audio samples to EOF.
_Original issue: http://code.google.com/p/editor-on-fire/issues/detail?id=121_ | priority | allow midi playback from on june i was thinking about how work intensive it is to prepare a new set of vocal tone samples it seems to be that since allegro has built in support for midi it may be worthwhile to allow it to be a user option to try to initialize the system s default midi device so that eof s output can optionally be sent to a software hardware synthesizer besides just sending note on and off data we ll need to consider how to allow the user to specify which synth tone to use when i created the piano tones for eof i learned a bit about this process but haven t mastered it fully i m looking into the creation of cakewalk instrument definition files which seem to be the most common format of instrument definitions anvil studio will allow you to define information for its internal use if the midi package is purchased or so but i don t know yet if it will allow an instrument definition file to be created i m willing to get this to test creating definition files and with any luck could write functions to import such definitions into eof this would allow for a good deal of flexibility with customizing the sound volume of vocal tones this may also be a good way to begin midi integration should eof be given the ability to accept midi input in the future ie piano record mode if fof gains the ability to play keyboard tracks besides vocal tones it could allow the user to supply their own clap metronome etc without having to provide recorded audio samples to eof original issue | 1 |
394,205 | 11,633,367,715 | IssuesEvent | 2020-02-28 08:04:40 | Repair-DeskPOS/RepairDesk-BUGS-IMPROVEMENTS | https://api.github.com/repos/Repair-DeskPOS/RepairDesk-BUGS-IMPROVEMENTS | closed | Issue when adding inventory items to ticket | Medium Priority enhancement | Cannot add more than 1 item to an inventory item to a ticket after it has been created. We need a solution for if a tech comes across a defective part and how we can take multiple out of iinventory | 1.0 | Issue when adding inventory items to ticket - Cannot add more than 1 item to an inventory item to a ticket after it has been created. We need a solution for if a tech comes across a defective part and how we can take multiple out of iinventory | priority | issue when adding inventory items to ticket cannot add more than item to an inventory item to a ticket after it has been created we need a solution for if a tech comes across a defective part and how we can take multiple out of iinventory | 1 |
40,942 | 2,868,956,368 | IssuesEvent | 2015-06-05 22:11:23 | dart-lang/pub | https://api.github.com/repos/dart-lang/pub | closed | pub serve protocol enhancements | DevStory enhancement NotPlanned Priority-Medium | <a href="https://github.com/devoncarew"><img src="https://avatars.githubusercontent.com/u/1269969?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [devoncarew](https://github.com/devoncarew)**
_Originally opened as dart-lang/sdk#13935_
----
Here's a proposal here: https://gist.github.com/devoncarew/6897835
and a spec that uses a similar protocol:
https://developers.google.com/chrome-developer-tools/docs/protocol/tot/page | 1.0 | pub serve protocol enhancements - <a href="https://github.com/devoncarew"><img src="https://avatars.githubusercontent.com/u/1269969?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [devoncarew](https://github.com/devoncarew)**
_Originally opened as dart-lang/sdk#13935_
----
Here's a proposal here: https://gist.github.com/devoncarew/6897835
and a spec that uses a similar protocol:
https://developers.google.com/chrome-developer-tools/docs/protocol/tot/page | priority | pub serve protocol enhancements issue by originally opened as dart lang sdk here s a proposal here and a spec that uses a similar protocol | 1 |
711,490 | 24,465,520,629 | IssuesEvent | 2022-10-07 14:42:14 | canonical/ubuntu.com | https://api.github.com/repos/canonical/ubuntu.com | closed | Dead link and typographical error on ubuntu.com/about/packages | Priority: Medium | ## Summary
In the **Charms are packages for cloud software operations** section on ubuntu.com/about/packages, there appears to be a misplaced box at the end of the paragraph and the *Get started making a charm* link takes you to a non-existent Discourse page.
## Process
1. Navigate to https://ubuntu.com/about/packages
2. Scroll down to **Charms are packages for cloud software operations** section (see the misplaced box)
3. Click on *Get started making a charm* hyperlink.
## Current and expected result
Noticed the box and did not expect it to be there. When I clicked *Get started making a charm* link I was taken to an "Oops not found" page rather than being taken to introductory documentation for making charms.
## Screenshot
### Misplaced box

### Non-existent Discourse page

## Browser details
Vivaldi | 5.3.2679.61 (Stable channel) stable (64-bit)
-- | --
Revision | fcd44035934deba382cb933a4c62bd4fa7690078
OS | Linux
JavaScript | V8 10.2.154.10
User Agent | Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/102.0.5005.136 Safari/537.36
Command Line | /usr/bin/vivaldi-stable --new-window --enable-crashpad --flag-switches-begin --flag-switches-end --save-page-as-mhtml
Executable Path | /opt/vivaldi/vivaldi
Profile Path | /home/nucci/.config/vivaldi/Default
I also have the uBlock origin extension enabled.
| 1.0 | Dead link and typographical error on ubuntu.com/about/packages - ## Summary
In the **Charms are packages for cloud software operations** section on ubuntu.com/about/packages, there appears to be a misplaced box at the end of the paragraph and the *Get started making a charm* link takes you to a non-existent Discourse page.
## Process
1. Navigate to https://ubuntu.com/about/packages
2. Scroll down to **Charms are packages for cloud software operations** section (see the misplaced box)
3. Click on *Get started making a charm* hyperlink.
## Current and expected result
Noticed the box and did not expect it to be there. When I clicked *Get started making a charm* link I was taken to an "Oops not found" page rather than being taken to introductory documentation for making charms.
## Screenshot
### Misplaced box

### Non-existent Discourse page

## Browser details
Vivaldi | 5.3.2679.61 (Stable channel) stable (64-bit)
-- | --
Revision | fcd44035934deba382cb933a4c62bd4fa7690078
OS | Linux
JavaScript | V8 10.2.154.10
User Agent | Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/102.0.5005.136 Safari/537.36
Command Line | /usr/bin/vivaldi-stable --new-window --enable-crashpad --flag-switches-begin --flag-switches-end --save-page-as-mhtml
Executable Path | /opt/vivaldi/vivaldi
Profile Path | /home/nucci/.config/vivaldi/Default
I also have the uBlock origin extension enabled.
| priority | dead link and typographical error on ubuntu com about packages summary in the charms are packages for cloud software operations section on ubuntu com about packages there appears to be a misplaced box at the end of the paragraph and the get started making a charm link takes you to a non existent discourse page process navigate to scroll down to charms are packages for cloud software operations section see the misplaced box click on get started making a charm hyperlink current and expected result noticed the box and did not expect it to be there when i clicked get started making a charm link i was taken to an oops not found page rather than being taken to introductory documentation for making charms screenshot misplaced box non existent discourse page browser details vivaldi stable channel stable bit revision os linux javascript user agent mozilla linux applewebkit khtml like gecko chrome safari command line usr bin vivaldi stable new window enable crashpad flag switches begin flag switches end save page as mhtml executable path opt vivaldi vivaldi profile path home nucci config vivaldi default i also have the ublock origin extension enabled | 1 |
420,510 | 12,238,937,023 | IssuesEvent | 2020-05-04 20:43:04 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | Allow "Likes" in Activity > Comments/Replies | bug component: activity priority: medium | **Describe the bug**
Allow "Like" functionality in Activity > Replies.
This is existing on BuddyBoss Wall with Boss Theme.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Activity, post something
2. Comment something to that post
3. You can only see Reply and Delete, but no Like option
**Expected behavior**
In BuddyBoss Wall, users have the option to Like the comments and replies on activity feed.
This must be existing on BuddyBoss Platform as well.
**Screenshots**


**Support ticket links**
https://secure.helpscout.net/conversation/1148788883/70864?folderId=3701241
| 1.0 | Allow "Likes" in Activity > Comments/Replies - **Describe the bug**
Allow "Like" functionality in Activity > Replies.
This is existing on BuddyBoss Wall with Boss Theme.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Activity, post something
2. Comment something to that post
3. You can only see Reply and Delete, but no Like option
**Expected behavior**
In BuddyBoss Wall, users have the option to Like the comments and replies on activity feed.
This must be existing on BuddyBoss Platform as well.
**Screenshots**


**Support ticket links**
https://secure.helpscout.net/conversation/1148788883/70864?folderId=3701241
| priority | allow likes in activity comments replies describe the bug allow like functionality in activity replies this is existing on buddyboss wall with boss theme to reproduce steps to reproduce the behavior go to activity post something comment something to that post you can only see reply and delete but no like option expected behavior in buddyboss wall users have the option to like the comments and replies on activity feed this must be existing on buddyboss platform as well screenshots support ticket links | 1 |
365,414 | 10,781,161,450 | IssuesEvent | 2019-11-04 14:25:31 | salesagility/SuiteCRM | https://api.github.com/repos/salesagility/SuiteCRM | closed | function isOwner() in SugarBean.php is not checking permissions ok | Bug Fix Proposed Medium Priority Resolved: Next Release | <!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
<!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. --->
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug -->
Investigating why an user with "owner permissions" in 'Notes' module could not enter the DetailView of a record that was created by him, I found that we have a problem with the method [`isOwner()` in SugarBean.php](https://github.com/salesagility/SuiteCRM/blob/master/data/SugarBean.php#L1673)
The `isOwner()` method is [checking: `isset($this->assigned_user_id)`](https://github.com/salesagility/SuiteCRM/blob/master/data/SugarBean.php#L1685)
```php
} elseif (isset($this->assigned_user_id)) {
if ($this->assigned_user_id == $user_id) {
return true;
}
return false;
}
```
But if `$this->assigned_user_id` is `''` the function [`isset`](https://www.php.net/isset) will return true, so the execution will enter into the `elseif` and will check `$this->assigned_user_id == $user_id` and in other words will check something like this: `'' == '7cb40869-3ca6-7c88-006c-5da61ac6dc6e'` which is `false`.
This behavior prevents the execution from reaching the next check that will return `true` because the record was created by the user, so `$this->created_by == $user_id` is `true`
```php
//other wise if there is a created_by that is the owner
if (isset($this->created_by) && $this->created_by == $user_id) {
return true;
}
```
#### Expected Behavior
<!--- Tell us what should happen -->
`isOwner()` checks permissions in a good way ;-)
#### Actual Behavior
<!--- Tell us what happens instead -->
<!--- Also please check relevant logs (suitecrm.log, php error.log etc.) -->
A user that creates a 'Note' cannot go to the DetailView of that record.
The `isOwner()` method is not checking permissions in a bad way.
#### Possible Fix
<!--- Not obligatory, but suggest a fix or reason for the bug -->
We should use [`!empty()`](https://www.php.net/empty#refsect1-function.empty-description) instead of [`isset()`](https://www.php.net/isset)
#### Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug include code to reproduce, if relevant -->
1. Create and Role and assign an user to this Role.
2. Set the followings permissions to the module Notes for this Role:
```
access: 89
delete: -99
edit: 75
export: -99
import: -99
list: 75
view: 75
```
3. Login with this user
4. Create a Note
5. Try to go to the DetailView.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used: 7.10.20
* Environment name and version (e.g. MySQL, PHP 7): MySQL 8.0.x / PHP 7.3.x
| 1.0 | function isOwner() in SugarBean.php is not checking permissions ok - <!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
<!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. --->
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug -->
Investigating why an user with "owner permissions" in 'Notes' module could not enter the DetailView of a record that was created by him, I found that we have a problem with the method [`isOwner()` in SugarBean.php](https://github.com/salesagility/SuiteCRM/blob/master/data/SugarBean.php#L1673)
The `isOwner()` method is [checking: `isset($this->assigned_user_id)`](https://github.com/salesagility/SuiteCRM/blob/master/data/SugarBean.php#L1685)
```php
} elseif (isset($this->assigned_user_id)) {
if ($this->assigned_user_id == $user_id) {
return true;
}
return false;
}
```
But if `$this->assigned_user_id` is `''` the function [`isset`](https://www.php.net/isset) will return true, so the execution will enter into the `elseif` and will check `$this->assigned_user_id == $user_id` and in other words will check something like this: `'' == '7cb40869-3ca6-7c88-006c-5da61ac6dc6e'` which is `false`.
This behavior prevents the execution from reaching the next check that will return `true` because the record was created by the user, so `$this->created_by == $user_id` is `true`
```php
//other wise if there is a created_by that is the owner
if (isset($this->created_by) && $this->created_by == $user_id) {
return true;
}
```
#### Expected Behavior
<!--- Tell us what should happen -->
`isOwner()` checks permissions in a good way ;-)
#### Actual Behavior
<!--- Tell us what happens instead -->
<!--- Also please check relevant logs (suitecrm.log, php error.log etc.) -->
A user that creates a 'Note' cannot go to the DetailView of that record.
The `isOwner()` method is not checking permissions in a bad way.
#### Possible Fix
<!--- Not obligatory, but suggest a fix or reason for the bug -->
We should use [`!empty()`](https://www.php.net/empty#refsect1-function.empty-description) instead of [`isset()`](https://www.php.net/isset)
#### Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug include code to reproduce, if relevant -->
1. Create and Role and assign an user to this Role.
2. Set the followings permissions to the module Notes for this Role:
```
access: 89
delete: -99
edit: 75
export: -99
import: -99
list: 75
view: 75
```
3. Login with this user
4. Create a Note
5. Try to go to the DetailView.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used: 7.10.20
* Environment name and version (e.g. MySQL, PHP 7): MySQL 8.0.x / PHP 7.3.x
| priority | function isowner in sugarbean php is not checking permissions ok issue investigating why an user with owner permissions in notes module could not enter the detailview of a record that was created by him i found that we have a problem with the method the isowner method is php elseif isset this assigned user id if this assigned user id user id return true return false but if this assigned user id is the function will return true so the execution will enter into the elseif and will check this assigned user id user id and in other words will check something like this which is false this behavior prevents the execution from reaching the next check that will return true because the record was created by the user so this created by user id is true php other wise if there is a created by that is the owner if isset this created by this created by user id return true expected behavior isowner checks permissions in a good way actual behavior a user that creates a note cannot go to the detailview of that record the isowner method is not checking permissions in a bad way possible fix we should use instead of steps to reproduce create and role and assign an user to this role set the followings permissions to the module notes for this role access delete edit export import list view login with this user create a note try to go to the detailview your environment suitecrm version used environment name and version e g mysql php mysql x php x | 1 |
16,259 | 2,612,939,294 | IssuesEvent | 2015-02-27 17:37:55 | skelterjohn/gomatrix | https://api.github.com/repos/skelterjohn/gomatrix | opened | support for native linear algebra libraries | auto-migrated Priority-Medium Type-Enhancement | ```
Such as LAPACK, LINPACK or others.
```
Original issue reported on code.google.com by `jasm...@gmail.com` on 25 Nov 2009 at 10:04 | 1.0 | support for native linear algebra libraries - ```
Such as LAPACK, LINPACK or others.
```
Original issue reported on code.google.com by `jasm...@gmail.com` on 25 Nov 2009 at 10:04 | priority | support for native linear algebra libraries such as lapack linpack or others original issue reported on code google com by jasm gmail com on nov at | 1 |
786,663 | 27,661,544,131 | IssuesEvent | 2023-03-12 15:27:37 | AY2223S2-CS2113-T12-1/tp | https://api.github.com/repos/AY2223S2-CS2113-T12-1/tp | opened | [Task] Edit expenses and income | type.Task priority.Medium | Add functionality for user to edit existing expense and income.
Edits the item at the specified INDEX, where INDEX must be accessible by the respective arrayList.
Fiels provided will be updated to the new inputs.
| 1.0 | [Task] Edit expenses and income - Add functionality for user to edit existing expense and income.
Edits the item at the specified INDEX, where INDEX must be accessible by the respective arrayList.
Fiels provided will be updated to the new inputs.
| priority | edit expenses and income add functionality for user to edit existing expense and income edits the item at the specified index where index must be accessible by the respective arraylist fiels provided will be updated to the new inputs | 1 |
356,035 | 10,587,744,270 | IssuesEvent | 2019-10-08 23:09:59 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio] Create APIs to support "Publishing status dashboard" | enhancement priority: medium | **In relation to #3352...**
We need new APIs to be able to support #3352's requests.
**Describe the solution you'd like**
Required APIs would be
- Get packages - retrieves the list of packages on the publishing queue. Each package should have at a minimum scheduled date, user who submitted, status, environment, comment and package id. It should support several filters namely a path expression, environments, and states.
- Get package details - retrieves the list of files associated with the package by package id. If this is not expensive to do at once on the `get packages` call, please do fit it there and we wouldn't need this second call.
- Cancel packages - given a list of package ids, it'd _cancel_ all.
 | 1.0 | [studio] Create APIs to support "Publishing status dashboard" - **In relation to #3352...**
We need new APIs to be able to support #3352's requests.
**Describe the solution you'd like**
Required APIs would be
- Get packages - retrieves the list of packages on the publishing queue. Each package should have at a minimum scheduled date, user who submitted, status, environment, comment and package id. It should support several filters namely a path expression, environments, and states.
- Get package details - retrieves the list of files associated with the package by package id. If this is not expensive to do at once on the `get packages` call, please do fit it there and we wouldn't need this second call.
- Cancel packages - given a list of package ids, it'd _cancel_ all.
 | priority | create apis to support publishing status dashboard in relation to we need new apis to be able to support s requests describe the solution you d like required apis would be get packages retrieves the list of packages on the publishing queue each package should have at a minimum scheduled date user who submitted status environment comment and package id it should support several filters namely a path expression environments and states get package details retrieves the list of files associated with the package by package id if this is not expensive to do at once on the get packages call please do fit it there and we wouldn t need this second call cancel packages given a list of package ids it d cancel all | 1 |
675,778 | 23,105,347,132 | IssuesEvent | 2022-07-27 08:19:53 | owncloud/web | https://api.github.com/repos/owncloud/web | closed | Left Sidebar: Active navitem pretends to be not clickable | Type:Bug Priority:p3-medium | ### Steps to reproduce
1. open ownCloud Web and hover over your active (blue) item in the left sidebar
2. Cursor-style is arrow (indicates "not clickable")

### Expected behaviour
Cursor style should indicate that its clickbale (Pointer)

| 1.0 | Left Sidebar: Active navitem pretends to be not clickable - ### Steps to reproduce
1. open ownCloud Web and hover over your active (blue) item in the left sidebar
2. Cursor-style is arrow (indicates "not clickable")

### Expected behaviour
Cursor style should indicate that its clickbale (Pointer)

| priority | left sidebar active navitem pretends to be not clickable steps to reproduce open owncloud web and hover over your active blue item in the left sidebar cursor style is arrow indicates not clickable expected behaviour cursor style should indicate that its clickbale pointer | 1 |
346,969 | 10,422,298,123 | IssuesEvent | 2019-09-16 08:42:24 | getkirby/kirby | https://api.github.com/repos/getkirby/kirby | closed | Switching pages causes select/checkbox/radio fields to briefly highlight as invalid | priority: medium 🔜 type: bug 🐛 | ### Describe the bug
When switching pages in the panel (for example by using the left/right navigation to switch between sibling pages or by navigation to parent or child page) all fields of type radio/checkbox/select are highlighted as "invalid" just before the pages switches. So it's only there for a brief moment.
Important to mention that this happens **although** the fields are not required and filled out.
Also, when switching away and back to the page, the fields will still not show up as invalid – because they aren't. The "flashing" only appears in the split second before switching away.
### To Reproduce
Steps to reproduce the behavior:
1. Add any field with the afflicted field types (select / checkbox / radio) to a page blueprint
2. Navigate away from the page, e.g. to a sibling page with same blueprint
3. See "flashing" of afflicted fields with "error" style (see screenshot above)
### Expected behavior
Prior to Kirby 3.2.(3) switching pages did not result in the behavior described above. The fields should not validate as "invalid" if they aren't.
### Screenshots

### Kirby Version
3.2.3
| 1.0 | Switching pages causes select/checkbox/radio fields to briefly highlight as invalid - ### Describe the bug
When switching pages in the panel (for example by using the left/right navigation to switch between sibling pages or by navigation to parent or child page) all fields of type radio/checkbox/select are highlighted as "invalid" just before the pages switches. So it's only there for a brief moment.
Important to mention that this happens **although** the fields are not required and filled out.
Also, when switching away and back to the page, the fields will still not show up as invalid – because they aren't. The "flashing" only appears in the split second before switching away.
### To Reproduce
Steps to reproduce the behavior:
1. Add any field with the afflicted field types (select / checkbox / radio) to a page blueprint
2. Navigate away from the page, e.g. to a sibling page with same blueprint
3. See "flashing" of afflicted fields with "error" style (see screenshot above)
### Expected behavior
Prior to Kirby 3.2.(3) switching pages did not result in the behavior described above. The fields should not validate as "invalid" if they aren't.
### Screenshots

### Kirby Version
3.2.3
| priority | switching pages causes select checkbox radio fields to briefly highlight as invalid describe the bug when switching pages in the panel for example by using the left right navigation to switch between sibling pages or by navigation to parent or child page all fields of type radio checkbox select are highlighted as invalid just before the pages switches so it s only there for a brief moment important to mention that this happens although the fields are not required and filled out also when switching away and back to the page the fields will still not show up as invalid – because they aren t the flashing only appears in the split second before switching away to reproduce steps to reproduce the behavior add any field with the afflicted field types select checkbox radio to a page blueprint navigate away from the page e g to a sibling page with same blueprint see flashing of afflicted fields with error style see screenshot above expected behavior prior to kirby switching pages did not result in the behavior described above the fields should not validate as invalid if they aren t screenshots kirby version | 1 |
779,616 | 27,360,243,762 | IssuesEvent | 2023-02-27 15:25:02 | orden-gg/fireball | https://api.github.com/repos/orden-gg/fireball | closed | Add prettier as dev dependency | enhancement priority: medium | It is dev only issue.
We have missing `prettier` dependency in `package.json`. | 1.0 | Add prettier as dev dependency - It is dev only issue.
We have missing `prettier` dependency in `package.json`. | priority | add prettier as dev dependency it is dev only issue we have missing prettier dependency in package json | 1 |
117,022 | 4,710,115,384 | IssuesEvent | 2016-10-14 09:01:06 | bespokeinteractive/mchapp | https://api.github.com/repos/bespokeinteractive/mchapp | closed | MCH Stores: User cannot clear search done for date i.e. from and to dates. | bug Medium Priority | User cannot clear search done for date -from and to dates.
Suggestion:
Add clear search button to allow clearing of search items within MCH store tabs. | 1.0 | MCH Stores: User cannot clear search done for date i.e. from and to dates. - User cannot clear search done for date -from and to dates.
Suggestion:
Add clear search button to allow clearing of search items within MCH store tabs. | priority | mch stores user cannot clear search done for date i e from and to dates user cannot clear search done for date from and to dates suggestion add clear search button to allow clearing of search items within mch store tabs | 1 |
473,517 | 13,643,449,770 | IssuesEvent | 2020-09-25 17:08:38 | LBL-EESA/TECA | https://api.github.com/repos/LBL-EESA/TECA | opened | app tests hard coded to use 2 mpi ranks | 2_medium_priority | **Describe the bug**
The app tests are hard coded to use 2 mpi ranks. For most of the test the mpi concurrency should be based on the cmake variable(s) TEST_CORES and/or HALF_TEST_CORES so that when there are more than 2 cores the tests run faster and MPI parallelism is stressed. | 1.0 | app tests hard coded to use 2 mpi ranks - **Describe the bug**
The app tests are hard coded to use 2 mpi ranks. For most of the test the mpi concurrency should be based on the cmake variable(s) TEST_CORES and/or HALF_TEST_CORES so that when there are more than 2 cores the tests run faster and MPI parallelism is stressed. | priority | app tests hard coded to use mpi ranks describe the bug the app tests are hard coded to use mpi ranks for most of the test the mpi concurrency should be based on the cmake variable s test cores and or half test cores so that when there are more than cores the tests run faster and mpi parallelism is stressed | 1 |
275,604 | 8,577,542,563 | IssuesEvent | 2018-11-13 00:30:09 | minio/minio | https://api.github.com/repos/minio/minio | closed | Storage backend has reached its minimum free disk threshold | community priority: medium working as intended | <!--- Provide a general summary of the issue in the Title above -->
## Expected Behavior
I want to put object up to end of my free space.
## Current Behavior
I try to put object with size about 10 Mib at the first time. But I cannot put the same object again. Minio throw an exception:
```
Error executing "PutObject" on "http://core.example.local:4567/prv0.example.local/mobile-app/packages/app_1.0.2.apk";
AWS HTTP error: Server error:
`PUT http://core.example.local:4567/prv0.example.local/mobile-app/packages/app_1.0.2.apk` resulted in a `507 Insufficient Storage` response:
<?xml version="1.0" encoding="UTF-8"?>
<Error><Code>XMinioStorageFull</Code><Message>Storage backend has reached its min (truncated...)
XMinioStorageFull (server): Storage backend has reached its minimum free disk threshold. Please delete a few objects to proceed. - <?xml version="1.0" encoding="UTF-8"?>
<Error><Code>XMinioStorageFull</Code><Message>Storage backend has reached its minimum free disk threshold. Please delete a few objects to proceed.</Message><Key></Key><BucketName></BucketName><Resource>/prv0.example.local/mobile-app/packages/app_1.0.2.apk</Resource><RequestId>3L137</RequestId><HostId>3L137</HostId></Error> (File: /vagrant/vendor/aws/aws-sdk-php/src/WrappedHttpHandler.php, Line: 191):
```
It's only one file in only one bucket. I have only one bucket at all. Total size of all files is 10 Mib.
## Possible Solution
I don't know. How to change threshold size of Minio?
I have enough free space:
```
vagrant@example:/var/www/html$ df
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/sda1 9620408 8127212 981452 90% /
udev 10240 0 10240 0% /dev
tmpfs 204876 4452 200424 3% /run
tmpfs 512184 0 512184 0% /dev/shm
tmpfs 5120 0 5120 0% /run/lock
tmpfs 512184 0 512184 0% /sys/fs/cgroup
192.168.20.1:/Volumes/Project/example 28945984 7908088 21037896 28% /vagrant
```

## Steps to Reproduce (for bugs)
```
/** @var \Aws\S3\S3Client $s3 **/
$s3->putObject([
'Key' => 'destination/file.apk',
'Bucket' => 'prv0.example.local',
'SourceFile' => '/path/to/file.apk',
'ContentType' => 'application/vnd.android.package-archive'
]);
```
## Context
Everything is general, nothing special.
## Regression
I don't know, may be it was before.
## Your Environment
* Version used (`minio version`):
```
Version: 2018-07-13T00:09:07Z
Release-Tag: RELEASE.2018-07-13T00-09-07Z
Commit-ID: 2aa18cafc674faf97201a973504cb4caa3a3e67c
```
* Environment name and version (e.g. nginx 1.9.1):
```
Vagrant 2.1.2
Nginx 1.6.2
PHP 7.1.10-1+0~20170929170631.9+jessie~1.gbp501135 (cli) (built: Sep 29 2017 17:33:58) ( NTS )
```
* Server type and version:
```
Linux example 3.16.0-4-amd64 #1 SMP Debian 3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux
Debian Jessie x64
```
* Operating System and version (`uname -a`):
```
Darwin macbook.local 16.7.0 Darwin Kernel Version 16.7.0: Thu Jun 21 20:07:39 PDT 2018; root:xnu-3789.73.14~1/RELEASE_X86_64 x86_64
```
* Link to your project:
It's private project. | 1.0 | Storage backend has reached its minimum free disk threshold - <!--- Provide a general summary of the issue in the Title above -->
## Expected Behavior
I want to put object up to end of my free space.
## Current Behavior
I try to put object with size about 10 Mib at the first time. But I cannot put the same object again. Minio throw an exception:
```
Error executing "PutObject" on "http://core.example.local:4567/prv0.example.local/mobile-app/packages/app_1.0.2.apk";
AWS HTTP error: Server error:
`PUT http://core.example.local:4567/prv0.example.local/mobile-app/packages/app_1.0.2.apk` resulted in a `507 Insufficient Storage` response:
<?xml version="1.0" encoding="UTF-8"?>
<Error><Code>XMinioStorageFull</Code><Message>Storage backend has reached its min (truncated...)
XMinioStorageFull (server): Storage backend has reached its minimum free disk threshold. Please delete a few objects to proceed. - <?xml version="1.0" encoding="UTF-8"?>
<Error><Code>XMinioStorageFull</Code><Message>Storage backend has reached its minimum free disk threshold. Please delete a few objects to proceed.</Message><Key></Key><BucketName></BucketName><Resource>/prv0.example.local/mobile-app/packages/app_1.0.2.apk</Resource><RequestId>3L137</RequestId><HostId>3L137</HostId></Error> (File: /vagrant/vendor/aws/aws-sdk-php/src/WrappedHttpHandler.php, Line: 191):
```
It's only one file in only one bucket. I have only one bucket at all. Total size of all files is 10 Mib.
## Possible Solution
I don't know. How to change threshold size of Minio?
I have enough free space:
```
vagrant@example:/var/www/html$ df
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/sda1 9620408 8127212 981452 90% /
udev 10240 0 10240 0% /dev
tmpfs 204876 4452 200424 3% /run
tmpfs 512184 0 512184 0% /dev/shm
tmpfs 5120 0 5120 0% /run/lock
tmpfs 512184 0 512184 0% /sys/fs/cgroup
192.168.20.1:/Volumes/Project/example 28945984 7908088 21037896 28% /vagrant
```

## Steps to Reproduce (for bugs)
```
/** @var \Aws\S3\S3Client $s3 **/
$s3->putObject([
'Key' => 'destination/file.apk',
'Bucket' => 'prv0.example.local',
'SourceFile' => '/path/to/file.apk',
'ContentType' => 'application/vnd.android.package-archive'
]);
```
## Context
Everything is general, nothing special.
## Regression
I don't know, may be it was before.
## Your Environment
* Version used (`minio version`):
```
Version: 2018-07-13T00:09:07Z
Release-Tag: RELEASE.2018-07-13T00-09-07Z
Commit-ID: 2aa18cafc674faf97201a973504cb4caa3a3e67c
```
* Environment name and version (e.g. nginx 1.9.1):
```
Vagrant 2.1.2
Nginx 1.6.2
PHP 7.1.10-1+0~20170929170631.9+jessie~1.gbp501135 (cli) (built: Sep 29 2017 17:33:58) ( NTS )
```
* Server type and version:
```
Linux example 3.16.0-4-amd64 #1 SMP Debian 3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux
Debian Jessie x64
```
* Operating System and version (`uname -a`):
```
Darwin macbook.local 16.7.0 Darwin Kernel Version 16.7.0: Thu Jun 21 20:07:39 PDT 2018; root:xnu-3789.73.14~1/RELEASE_X86_64 x86_64
```
* Link to your project:
It's private project. | priority | storage backend has reached its minimum free disk threshold expected behavior i want to put object up to end of my free space current behavior i try to put object with size about mib at the first time but i cannot put the same object again minio throw an exception error executing putobject on aws http error server error put resulted in a insufficient storage response xminiostoragefull storage backend has reached its min truncated xminiostoragefull server storage backend has reached its minimum free disk threshold please delete a few objects to proceed xminiostoragefull storage backend has reached its minimum free disk threshold please delete a few objects to proceed example local mobile app packages app apk file vagrant vendor aws aws sdk php src wrappedhttphandler php line it s only one file in only one bucket i have only one bucket at all total size of all files is mib possible solution i don t know how to change threshold size of minio i have enough free space vagrant example var www html df filesystem blocks used available use mounted on dev udev dev tmpfs run tmpfs dev shm tmpfs run lock tmpfs sys fs cgroup volumes project example vagrant steps to reproduce for bugs var aws putobject key destination file apk bucket example local sourcefile path to file apk contenttype application vnd android package archive context everything is general nothing special regression i don t know may be it was before your environment version used minio version version release tag release commit id environment name and version e g nginx vagrant nginx php jessie cli built sep nts server type and version linux example smp debian gnu linux debian jessie operating system and version uname a darwin macbook local darwin kernel version thu jun pdt root xnu release link to your project it s private project | 1 |
314,456 | 9,597,324,302 | IssuesEvent | 2019-05-09 20:58:44 | minio/minio | https://api.github.com/repos/minio/minio | closed | [gateway] b2 - unable to create s3ql filesystem | community priority: medium working as intended | <!--- Provide a general summary of the issue in the Title above -->
## Expected Behavior
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
Expect mkfs.s3ql to create a filesystem.
## Current Behavior
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
```
2019-05-06 11:49:34.210 14 INFO MainThread s3ql.mkfs.main: Generating random encryption key...
2019-05-06 11:49:34.215 14 DEBUG MainThread s3ql.backends.s3c.open_write: started with s3ql_passphrase
2019-05-06 11:49:34.217 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase
2019-05-06 11:49:34.218 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None
2019-05-06 11:49:34.219 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:36.330 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase
2019-05-06 11:49:36.330 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None
2019-05-06 11:49:36.335 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:37.484 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.02 Hz
2019-05-06 11:49:37.484 14 DEBUG MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f115226b22e7e6c2c_d20190506_m114935_c001_v0001039_t0027-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 1)...
2019-05-06 11:49:37.511 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase
2019-05-06 11:49:37.512 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None
2019-05-06 11:49:37.513 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:40.137 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase
2019-05-06 11:49:40.137 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None
2019-05-06 11:49:40.138 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:41.718 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.03 Hz
2019-05-06 11:49:41.718 14 DEBUG MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f1025162da09e4b6b_d20190506_m114938_c001_v0001117_t0052-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 2)...
2019-05-06 11:49:41.775 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase
2019-05-06 11:49:41.776 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None
2019-05-06 11:49:41.777 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:45.836 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase
2019-05-06 11:49:45.836 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None
2019-05-06 11:49:45.837 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:46.991 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.05 Hz
2019-05-06 11:49:46.992 14 INFO MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f116d93e2350333e3_d20190506_m114942_c001_v0001101_t0013-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 3)...
```
etc...
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
mkfs.s3ql works fine when minio is acting as a server. The B2 gateway seems to be broken.
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. minio gateway b2
2. mkfs.s3ql --backend-options no-ssl s3c://minio:9000/s3ql
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio version`):
* Environment name and version (e.g. nginx 1.9.1):
* Server type and version:
* Operating System and version (`uname -a`):
* Link to your project:
minio/minio docker container:
/ # minio version
Version: 2019-05-02T19:07:09Z
Release-Tag: RELEASE.2019-05-02T19-07-09Z
Commit-ID: 4b858b562a0887e10bfd0414dc87e68f1af31c3a
s3ql docker container:
kronostechnologies/s3ql
```
# dpkg -l s3ql
Desired=Unknown/Install/Remove/Purge/Hold
| Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
|/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
||/ Name Version Architecture Description
+++-==============-============-============-=================================================
ii s3ql 3.0+dfsg-1 amd64 Full-featured file system for online data storage
``` | 1.0 | [gateway] b2 - unable to create s3ql filesystem - <!--- Provide a general summary of the issue in the Title above -->
## Expected Behavior
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
Expect mkfs.s3ql to create a filesystem.
## Current Behavior
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
```
2019-05-06 11:49:34.210 14 INFO MainThread s3ql.mkfs.main: Generating random encryption key...
2019-05-06 11:49:34.215 14 DEBUG MainThread s3ql.backends.s3c.open_write: started with s3ql_passphrase
2019-05-06 11:49:34.217 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase
2019-05-06 11:49:34.218 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None
2019-05-06 11:49:34.219 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:36.330 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase
2019-05-06 11:49:36.330 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None
2019-05-06 11:49:36.335 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:37.484 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.02 Hz
2019-05-06 11:49:37.484 14 DEBUG MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f115226b22e7e6c2c_d20190506_m114935_c001_v0001039_t0027-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 1)...
2019-05-06 11:49:37.511 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase
2019-05-06 11:49:37.512 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None
2019-05-06 11:49:37.513 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:40.137 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase
2019-05-06 11:49:40.137 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None
2019-05-06 11:49:40.138 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:41.718 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.03 Hz
2019-05-06 11:49:41.718 14 DEBUG MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f1025162da09e4b6b_d20190506_m114938_c001_v0001117_t0052-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 2)...
2019-05-06 11:49:41.775 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase
2019-05-06 11:49:41.776 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None
2019-05-06 11:49:41.777 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:45.836 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase
2019-05-06 11:49:45.836 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None
2019-05-06 11:49:45.837 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase
2019-05-06 11:49:46.991 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.05 Hz
2019-05-06 11:49:46.992 14 INFO MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f116d93e2350333e3_d20190506_m114942_c001_v0001101_t0013-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 3)...
```
etc...
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
mkfs.s3ql works fine when minio is acting as a server. The B2 gateway seems to be broken.
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. minio gateway b2
2. mkfs.s3ql --backend-options no-ssl s3c://minio:9000/s3ql
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio version`):
* Environment name and version (e.g. nginx 1.9.1):
* Server type and version:
* Operating System and version (`uname -a`):
* Link to your project:
minio/minio docker container:
/ # minio version
Version: 2019-05-02T19:07:09Z
Release-Tag: RELEASE.2019-05-02T19-07-09Z
Commit-ID: 4b858b562a0887e10bfd0414dc87e68f1af31c3a
s3ql docker container:
kronostechnologies/s3ql
```
# dpkg -l s3ql
Desired=Unknown/Install/Remove/Purge/Hold
| Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
|/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
||/ Name Version Architecture Description
+++-==============-============-============-=================================================
ii s3ql 3.0+dfsg-1 amd64 Full-featured file system for online data storage
``` | priority | unable to create filesystem expected behavior expect mkfs to create a filesystem current behavior info mainthread mkfs main generating random encryption key debug mainthread backends open write started with passphrase debug mainthread backends close started with passphrase debug mainthread backends do request started with put passphrase none qs none debug mainthread backends send request sending put passphrase debug mainthread backends delete started with passphrase debug mainthread backends do request started with delete passphrase none qs none debug mainthread backends send request sending delete passphrase debug mainthread backends common wrapped average retry rate hz debug mainthread backends common wrapped encountered baddigesterror baddigest mismatch for passphrase received sent retrying objectw close attempt debug mainthread backends close started with passphrase debug mainthread backends do request started with put passphrase none qs none debug mainthread backends send request sending put passphrase debug mainthread backends delete started with passphrase debug mainthread backends do request started with delete passphrase none qs none debug mainthread backends send request sending delete passphrase debug mainthread backends common wrapped average retry rate hz debug mainthread backends common wrapped encountered baddigesterror baddigest mismatch for passphrase received sent retrying objectw close attempt debug mainthread backends close started with passphrase debug mainthread backends do request started with put passphrase none qs none debug mainthread backends send request sending put passphrase debug mainthread backends delete started with passphrase debug mainthread backends do request started with delete passphrase none qs none debug mainthread backends send request sending delete passphrase debug mainthread backends common wrapped average retry rate hz info mainthread backends common wrapped encountered baddigesterror baddigest mismatch for passphrase received sent retrying objectw close attempt etc possible solution mkfs works fine when minio is acting as a server the gateway seems to be broken steps to reproduce for bugs minio gateway mkfs backend options no ssl minio your environment version used minio version environment name and version e g nginx server type and version operating system and version uname a link to your project minio minio docker container minio version version release tag release commit id docker container kronostechnologies dpkg l desired unknown install remove purge hold status not inst conf files unpacked half conf half inst trig await trig pend err none reinst required status err uppercase bad name version architecture description ii dfsg full featured file system for online data storage | 1 |
88,646 | 3,783,485,984 | IssuesEvent | 2016-03-19 05:27:56 | parishod/viewer-api | https://api.github.com/repos/parishod/viewer-api | closed | Copy to clipboard feature for social sharing modal. | Priority: Medium Status: In Progress Status: Review Needed Type: Enhancement | Copy to clipboard button should be added along with the already existing social buttons.
Clicking on copy to clipboard button should copy to clipboard if possible and should also display text box filled with url and pre selected. | 1.0 | Copy to clipboard feature for social sharing modal. - Copy to clipboard button should be added along with the already existing social buttons.
Clicking on copy to clipboard button should copy to clipboard if possible and should also display text box filled with url and pre selected. | priority | copy to clipboard feature for social sharing modal copy to clipboard button should be added along with the already existing social buttons clicking on copy to clipboard button should copy to clipboard if possible and should also display text box filled with url and pre selected | 1 |
709,597 | 24,384,255,451 | IssuesEvent | 2022-10-04 10:22:17 | trustwallet/wallet-core | https://api.github.com/repos/trustwallet/wallet-core | opened | Support custom EVM chain without code change, but runtime data | enhancement priority:low size:medium improvements | **Is your feature request related to a problem? Please describe.**
Support for a new EVM-clone chain has been simplified (see #2214), but still requires code change and new release in wallet-core. It would be nice if new custom chain can be added to a wallet dynamically at runtime, but still some support from wallet core.
**Describe the solution you'd like**
Two possible solutions in the wallet app:
1. A custom EVM-clone chain can be added by the user, and then app uses Ethereum coin for address derivation, and Ethereum coin and custom chainId for signing. No involvement from wallet core.
2. A custom EVM-clone chain can be added by the user. App 'registers' the new chain properties with wallet code. Wallet core extends its relevant coin-specific methods to also support the dynamically registered chains (additionally to the ones statically defined in `registry.json`).
**Limitation**
Since dynamically registered types are not known at build time, generated sources will not contain these coins.
**Describe alternatives you've considered**
See above number 1).
**Checklist**
<!--- Group checklist per issue needed, one specific feature of your goal -->
<!--- Each big task can have subtask, doesn't hesitate to split into small pull request to simplify the review process -->
- [ ] task number 1
- [ ] subtask number 1
- [ ] subtask number 2
- [ ] task number 2
- [ ] task number 3
**Resources**
<!--- Link resources this way: [My Resource Title](link) -->
| 1.0 | Support custom EVM chain without code change, but runtime data - **Is your feature request related to a problem? Please describe.**
Support for a new EVM-clone chain has been simplified (see #2214), but still requires code change and new release in wallet-core. It would be nice if new custom chain can be added to a wallet dynamically at runtime, but still some support from wallet core.
**Describe the solution you'd like**
Two possible solutions in the wallet app:
1. A custom EVM-clone chain can be added by the user, and then app uses Ethereum coin for address derivation, and Ethereum coin and custom chainId for signing. No involvement from wallet core.
2. A custom EVM-clone chain can be added by the user. App 'registers' the new chain properties with wallet code. Wallet core extends its relevant coin-specific methods to also support the dynamically registered chains (additionally to the ones statically defined in `registry.json`).
**Limitation**
Since dynamically registered types are not known at build time, generated sources will not contain these coins.
**Describe alternatives you've considered**
See above number 1).
**Checklist**
<!--- Group checklist per issue needed, one specific feature of your goal -->
<!--- Each big task can have subtask, doesn't hesitate to split into small pull request to simplify the review process -->
- [ ] task number 1
- [ ] subtask number 1
- [ ] subtask number 2
- [ ] task number 2
- [ ] task number 3
**Resources**
<!--- Link resources this way: [My Resource Title](link) -->
| priority | support custom evm chain without code change but runtime data is your feature request related to a problem please describe support for a new evm clone chain has been simplified see but still requires code change and new release in wallet core it would be nice if new custom chain can be added to a wallet dynamically at runtime but still some support from wallet core describe the solution you d like two possible solutions in the wallet app a custom evm clone chain can be added by the user and then app uses ethereum coin for address derivation and ethereum coin and custom chainid for signing no involvement from wallet core a custom evm clone chain can be added by the user app registers the new chain properties with wallet code wallet core extends its relevant coin specific methods to also support the dynamically registered chains additionally to the ones statically defined in registry json limitation since dynamically registered types are not known at build time generated sources will not contain these coins describe alternatives you ve considered see above number checklist task number subtask number subtask number task number task number resources | 1 |
568,990 | 16,992,239,909 | IssuesEvent | 2021-06-30 22:29:12 | nexB/scancode.io | https://api.github.com/repos/nexB/scancode.io | closed | meta-issue: design-needed: Finding and acting on issues and exceptions to review in scan | enhancement medium priority | Once a scan is completed, I want to be able to efficiently review "issues and exception" in my scan such as:
1. files or packages under a certain license (e.g. possibly in relation with policies #90 )
2. inconclusive detections (such as unknown licenses and similar)
and possibly a few more.
It could take the form of filtered lists or some dashboard that I could "zoom" to and needs design.
The outcome is to identify a number of related files or packages that share some origin and license exception, issues or policy problem.
Once I "zoom in" a problem, I would to be able to act on this such as:
1. changing or setting some status that would state that these resources or packages have been reviewed
2. changing or setting and updated license and or copyright or origin
3. mark a bunch files with some detected clues are having non-relevant clues (say for instance because of an http://example.com URLs)
4. tag files or packages for future extra review or action (something like a TODO list)
and possibly a few more.
| 1.0 | meta-issue: design-needed: Finding and acting on issues and exceptions to review in scan - Once a scan is completed, I want to be able to efficiently review "issues and exception" in my scan such as:
1. files or packages under a certain license (e.g. possibly in relation with policies #90 )
2. inconclusive detections (such as unknown licenses and similar)
and possibly a few more.
It could take the form of filtered lists or some dashboard that I could "zoom" to and needs design.
The outcome is to identify a number of related files or packages that share some origin and license exception, issues or policy problem.
Once I "zoom in" a problem, I would to be able to act on this such as:
1. changing or setting some status that would state that these resources or packages have been reviewed
2. changing or setting and updated license and or copyright or origin
3. mark a bunch files with some detected clues are having non-relevant clues (say for instance because of an http://example.com URLs)
4. tag files or packages for future extra review or action (something like a TODO list)
and possibly a few more.
| priority | meta issue design needed finding and acting on issues and exceptions to review in scan once a scan is completed i want to be able to efficiently review issues and exception in my scan such as files or packages under a certain license e g possibly in relation with policies inconclusive detections such as unknown licenses and similar and possibly a few more it could take the form of filtered lists or some dashboard that i could zoom to and needs design the outcome is to identify a number of related files or packages that share some origin and license exception issues or policy problem once i zoom in a problem i would to be able to act on this such as changing or setting some status that would state that these resources or packages have been reviewed changing or setting and updated license and or copyright or origin mark a bunch files with some detected clues are having non relevant clues say for instance because of an urls tag files or packages for future extra review or action something like a todo list and possibly a few more | 1 |
26,280 | 2,684,288,338 | IssuesEvent | 2015-03-28 20:51:56 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | opened | Auto update nightmare | 2–5 stars bug imported Priority-Medium | _From [sspirido...@gmail.com](https://code.google.com/u/106209959253029514009/) on February 08, 2013 15:39:50_
Required information! OS version: Win8 x64 ConEmu version: 2013.02.05 and all previous
Far version (if you are using Far Manager): Far Manager, version 3.0 (build 3011) x64 *Bug description* Mess on auto update ConEmu setup with automatically update on dev. releases.
1. ComEmu (with FAR) ask to upload new release - Yes
2. ComEmu (with FAR) ask to update - Yes
3. New console has been opened
4. ConEmu (with new console) ask to upload new release - No (Grr..)
5. Installator reports about open FAR and console (Grrr...)
6. If select close it can't close new console (Grrrr....)
7. If select do not close and reboot after the installator do something
8. Installator ask to reboot - No, I d'like to do it later.
9. Installator crash with some error (Grrrr!!!).
10. If I run the FAR (with conEmu) after that quest WITHOUT reboot it will updated and works correctly on the first look.
I am lost. *Steps to reproduction* 1. See above
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=930_ | 1.0 | Auto update nightmare - _From [sspirido...@gmail.com](https://code.google.com/u/106209959253029514009/) on February 08, 2013 15:39:50_
Required information! OS version: Win8 x64 ConEmu version: 2013.02.05 and all previous
Far version (if you are using Far Manager): Far Manager, version 3.0 (build 3011) x64 *Bug description* Mess on auto update ConEmu setup with automatically update on dev. releases.
1. ComEmu (with FAR) ask to upload new release - Yes
2. ComEmu (with FAR) ask to update - Yes
3. New console has been opened
4. ConEmu (with new console) ask to upload new release - No (Grr..)
5. Installator reports about open FAR and console (Grrr...)
6. If select close it can't close new console (Grrrr....)
7. If select do not close and reboot after the installator do something
8. Installator ask to reboot - No, I d'like to do it later.
9. Installator crash with some error (Grrrr!!!).
10. If I run the FAR (with conEmu) after that quest WITHOUT reboot it will updated and works correctly on the first look.
I am lost. *Steps to reproduction* 1. See above
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=930_ | priority | auto update nightmare from on february required information os version conemu version and all previous far version if you are using far manager far manager version build bug description mess on auto update conemu setup with automatically update on dev releases comemu with far ask to upload new release yes comemu with far ask to update yes new console has been opened conemu with new console ask to upload new release no grr installator reports about open far and console grrr if select close it can t close new console grrrr if select do not close and reboot after the installator do something installator ask to reboot no i d like to do it later installator crash with some error grrrr if i run the far with conemu after that quest without reboot it will updated and works correctly on the first look i am lost steps to reproduction see above original issue | 1 |
208,927 | 7,163,083,995 | IssuesEvent | 2018-01-29 05:08:47 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [6.4] Console is not opening until server comes online | Medium Priority | John asked me to post this as apparently it is not the behavior wanted:
The Console does not open up on desktop until after the world generation/server load is compete and the port opens to make game available. Once the game is online the console will also load but it sort of removed a major tool for system admins to know what the server is doing/what state it is in other then just by watching CPU behavior in the task manager, especially at this time when we are having many different issues trying to get the servers to load. | 1.0 | [6.4] Console is not opening until server comes online - John asked me to post this as apparently it is not the behavior wanted:
The Console does not open up on desktop until after the world generation/server load is compete and the port opens to make game available. Once the game is online the console will also load but it sort of removed a major tool for system admins to know what the server is doing/what state it is in other then just by watching CPU behavior in the task manager, especially at this time when we are having many different issues trying to get the servers to load. | priority | console is not opening until server comes online john asked me to post this as apparently it is not the behavior wanted the console does not open up on desktop until after the world generation server load is compete and the port opens to make game available once the game is online the console will also load but it sort of removed a major tool for system admins to know what the server is doing what state it is in other then just by watching cpu behavior in the task manager especially at this time when we are having many different issues trying to get the servers to load | 1 |
609,691 | 18,884,930,333 | IssuesEvent | 2021-11-15 06:18:22 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | closed | Message Improvement | feature: enhancement priority: medium | Some code refactoring required to fix some performance-related core logic
- With the Message thread ajax, We sending all recipients lists for every thread. it'll be a little heavy with a large network
We should send recipients with some limit like 10 and we also need to send total recipients with thread objects.
- We need to provide separate ajax to get message thread recipients so we can support pagination.
- One major refactoring we can do is, we can move our message logic based on Rest endpoint instead of ajax, by this change we no need to maintain two chunks of code.
-----
Customer Query for the Reference
Slow SQL queries from group message causing the private messages and message notification to load so slow and sometimes causing 500 internal error, it happens when there are thousands of group members in a group message thread.
This is the slow SQL query reported by client: https://prnt.sc/wiqzai
--------
https://secure.helpscout.net/conversation/1470954591/134010
Test Group and with only one member the Group message had no delay and acted as one would expect.
some other Group - with over 49 members it has this massive delay.
Related HS Tickets:
https://secure.helpscout.net/conversation/1371027857/115494/
https://secure.helpscout.net/conversation/1388206152/118467/
https://secure.helpscout.net/conversation/1470954591/134010/
https://secure.helpscout.net/conversation/1567963949/152658/ | 1.0 | Message Improvement - Some code refactoring required to fix some performance-related core logic
- With the Message thread ajax, We sending all recipients lists for every thread. it'll be a little heavy with a large network
We should send recipients with some limit like 10 and we also need to send total recipients with thread objects.
- We need to provide separate ajax to get message thread recipients so we can support pagination.
- One major refactoring we can do is, we can move our message logic based on Rest endpoint instead of ajax, by this change we no need to maintain two chunks of code.
-----
Customer Query for the Reference
Slow SQL queries from group message causing the private messages and message notification to load so slow and sometimes causing 500 internal error, it happens when there are thousands of group members in a group message thread.
This is the slow SQL query reported by client: https://prnt.sc/wiqzai
--------
https://secure.helpscout.net/conversation/1470954591/134010
Test Group and with only one member the Group message had no delay and acted as one would expect.
some other Group - with over 49 members it has this massive delay.
Related HS Tickets:
https://secure.helpscout.net/conversation/1371027857/115494/
https://secure.helpscout.net/conversation/1388206152/118467/
https://secure.helpscout.net/conversation/1470954591/134010/
https://secure.helpscout.net/conversation/1567963949/152658/ | priority | message improvement some code refactoring required to fix some performance related core logic with the message thread ajax we sending all recipients lists for every thread it ll be a little heavy with a large network we should send recipients with some limit like and we also need to send total recipients with thread objects we need to provide separate ajax to get message thread recipients so we can support pagination one major refactoring we can do is we can move our message logic based on rest endpoint instead of ajax by this change we no need to maintain two chunks of code customer query for the reference slow sql queries from group message causing the private messages and message notification to load so slow and sometimes causing internal error it happens when there are thousands of group members in a group message thread this is the slow sql query reported by client test group and with only one member the group message had no delay and acted as one would expect some other group with over members it has this massive delay related hs tickets | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.