Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
957
labels
stringlengths
4
795
body
stringlengths
1
259k
index
stringclasses
12 values
text_combine
stringlengths
96
259k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
118,378
4,738,328,886
IssuesEvent
2016-10-20 03:37:29
CS2103AUG2016-T10-C2/main
https://api.github.com/repos/CS2103AUG2016-T10-C2/main
closed
Code Cleaning
priority.medium type.enhancement
Change approach for prepareEdit in Parser. Refer to #92 for the suggestions provided by @kabir-plod
1.0
Code Cleaning - Change approach for prepareEdit in Parser. Refer to #92 for the suggestions provided by @kabir-plod
priority
code cleaning change approach for prepareedit in parser refer to for the suggestions provided by kabir plod
1
171,108
6,479,096,078
IssuesEvent
2017-08-18 09:41:24
pest-parser/pest
https://api.github.com/repos/pest-parser/pest
closed
Matchers that do not make progress are difficult to use in error handling
Priority: Medium Status: Pending Type: Feature/Enhancement
Let's say I have a rule like this: ```rust identifier = { !keyword ~ alpha ~ alphanumeric* } keyword = { ["foo"] } ``` If I try to parse the word `foo`, it doesn't parse (as expected), but I don't get enough information to provide a useful error to the user. `parser.expected()` returns `[identifier]` and all the other methods return `[]`. So the error I would report to the user is: ``` Expected an identifier ``` Notice that the rest of the rule is `alpha ~ alphanumeric*`. The word `foo` clearly matches that criteria, but the identifier rule fails to parse because of the `!keyword` part. It would be more useful to print: ``` Expected an identifier, found a keyword instead. ``` Or something else letting them know what really happened. This applies to all matchers that do not make progress like `!` and `&`.
1.0
Matchers that do not make progress are difficult to use in error handling - Let's say I have a rule like this: ```rust identifier = { !keyword ~ alpha ~ alphanumeric* } keyword = { ["foo"] } ``` If I try to parse the word `foo`, it doesn't parse (as expected), but I don't get enough information to provide a useful error to the user. `parser.expected()` returns `[identifier]` and all the other methods return `[]`. So the error I would report to the user is: ``` Expected an identifier ``` Notice that the rest of the rule is `alpha ~ alphanumeric*`. The word `foo` clearly matches that criteria, but the identifier rule fails to parse because of the `!keyword` part. It would be more useful to print: ``` Expected an identifier, found a keyword instead. ``` Or something else letting them know what really happened. This applies to all matchers that do not make progress like `!` and `&`.
priority
matchers that do not make progress are difficult to use in error handling let s say i have a rule like this rust identifier keyword alpha alphanumeric keyword if i try to parse the word foo it doesn t parse as expected but i don t get enough information to provide a useful error to the user parser expected returns and all the other methods return so the error i would report to the user is expected an identifier notice that the rest of the rule is alpha alphanumeric the word foo clearly matches that criteria but the identifier rule fails to parse because of the keyword part it would be more useful to print expected an identifier found a keyword instead or something else letting them know what really happened this applies to all matchers that do not make progress like and
1
519,419
15,050,932,295
IssuesEvent
2021-02-03 13:31:15
rsherrera/project-caja
https://api.github.com/repos/rsherrera/project-caja
opened
Judicial. Pago Rendición. Cancela Aporte. Orden de las cuotas listadas.
Priority-Medium bug
El sistema muestra desordenado el listado de cuotas en pantalla Cancela Aporte. Debe ordenarse desde la cuota mas antigua a la mas nueva.
1.0
Judicial. Pago Rendición. Cancela Aporte. Orden de las cuotas listadas. - El sistema muestra desordenado el listado de cuotas en pantalla Cancela Aporte. Debe ordenarse desde la cuota mas antigua a la mas nueva.
priority
judicial pago rendición cancela aporte orden de las cuotas listadas el sistema muestra desordenado el listado de cuotas en pantalla cancela aporte debe ordenarse desde la cuota mas antigua a la mas nueva
1
558,152
16,527,493,140
IssuesEvent
2021-05-26 22:27:09
actually-colab/editor
https://api.github.com/repos/actually-colab/editor
opened
Emit websocket errors
AWS bug client difficulty: hard priority: medium server socket
API Gateway Route Responses do not properly support emitting errors. Come up with an alternative such as emitting errors through messages.
1.0
Emit websocket errors - API Gateway Route Responses do not properly support emitting errors. Come up with an alternative such as emitting errors through messages.
priority
emit websocket errors api gateway route responses do not properly support emitting errors come up with an alternative such as emitting errors through messages
1
87,705
3,757,176,058
IssuesEvent
2016-03-13 20:40:18
svof/svof
https://api.github.com/repos/svof/svof
opened
Consider new tree scenario against 2 handed
enhancement low priority medium difficulty up for grabs
reported by Aerek > A Touch Tree scenario for curing when you ONLY have 2H fractures and no other affs would be useful. I can't figure out how to script this, myself
1.0
Consider new tree scenario against 2 handed - reported by Aerek > A Touch Tree scenario for curing when you ONLY have 2H fractures and no other affs would be useful. I can't figure out how to script this, myself
priority
consider new tree scenario against handed reported by aerek a touch tree scenario for curing when you only have fractures and no other affs would be useful i can t figure out how to script this myself
1
275,591
8,577,325,183
IssuesEvent
2018-11-12 23:29:17
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
segfault in python multithreaded setting
bug medium priority
Zihang Dai reports (and I've reproduced) that the autograd engine is not thread-safe. Here's a repro script: https://gist.github.com/zihangdai/fc8f76fbb8a0f6323a6b31e6d98ceb50 Run it a few times, occassionally it segfaults. Segfault is from a much different location, when cleaning up imports: ``` Thread 1 "python" received signal SIGSEGV, Segmentation fault. malloc_consolidate (av=av@entry=0x7ffff70a8b20 <main_arena>) at malloc.c:4167 4167 malloc.c: No such file or directory. (gdb) where #0 malloc_consolidate (av=av@entry=0x7ffff70a8b20 <main_arena>) at malloc.c:4167 #1 0x00007ffff6d64678 in _int_free (av=0x7ffff70a8b20 <main_arena>, p=<optimized out>, have_lock=0) at malloc.c:4075 #2 0x00007ffff6d6853c in __GI___libc_free (mem=<optimized out>) at malloc.c:2968 #3 0x00007ffff7a6a427 in dict_dealloc (mp=0x7ffff7ed0c58) at Objects/dictobject.c:1044 #4 0x00007ffff7a682c7 in insertdict_by_entry (mp=0x7ffff7ed0d70, key='build_time_vars', hash=<optimized out>, ep=<optimized out>, value=<optimized out>) at Objects/dictobject.c:519 #5 0x00007ffff7a6b79c in insertdict (value=None, hash=-295987683324531010, key='build_time_vars', mp=0x7ffff7ed0d70) at Objects/dictobject.c:556 #6 dict_set_item_by_hash_or_entry (value=None, ep=0x0, hash=-295987683324531010, key='build_time_vars', op={'__builtins__': {'bytearray': <type at remote 0x7ffff7d7c300>, 'IndexError': <type at remote 0x7ffff7d82bc0>, 'all': <built-in function all>, 'help': <_Helper at remote 0x7ffff7ed8c50>, 'vars': <built-in function vars>, 'SyntaxError': <type at remote 0x7ffff7d82540>, 'unicode': <type at remote 0x7ffff7d99040>, 'UnicodeDecodeError': <type at remote 0x7ffff7d833e0>, 'memoryview': <type at remote 0x7ffff7d8d900>, 'isinstance': <built-in function isinstance>, 'copyright': <_Printer(_Printer__data='Copyright (c) 2001-2016 Python Software Foundation.\nAll Rights Reserved.\n\nCopyright (c) 2000 BeOpen.com.\nAll Rights Reserved.\n\nCopyright (c) 1995-2001 Corporation for National Research Initiatives.\nAll Rights Reserved.\n\nCopyright (c) 1991-1995 Stichting Mathematisch Centrum, Amsterdam.\nAll Rights Reserved.', _Printer__lines=None, _Printer__name='copyright', _Printer__dirs=(), _Printer__files=(...)) at remote 0x7ffff7ed8a10>, 'NameError': <type at remote 0x7ffff7d82060>, 'BytesWarning': <type at remote 0x7ffff...(truncated)) at Objects/dictobject.c:795 #7 PyDict_SetItem ( op={'__builtins__': {'bytearray': <type at remote 0x7ffff7d7c300>, 'IndexError': <type at remote 0x7ffff7d82bc0>, 'all': <built-in function all>, 'help': <_Helper at remote 0x7ffff7ed8c50>, 'vars': <built-in function vars>, 'SyntaxError': <type at remote 0x7ffff7d82540>, 'unicode': <type at remote 0x7ffff7d99040>, 'UnicodeDecodeError': <type at remote 0x7ffff7d833e0>, 'memoryview': <type at remote 0x7ffff7d8d900>, 'isinstance': <built-in function isinstance>, 'copyright': <_Printer(_Printer__data='Copyright (c) 2001-2016 Python Software Foundation.\nAll Rights Reserved.\n\nCopyright (c) 2000 BeOpen.com.\nAll Rights Reserved.\n\nCopyright (c) 1995-2001 Corporation for National Research Initiatives.\nAll Rights Reserved.\n\nCopyright (c) 1991-1995 Stichting Mathematisch Centrum, Amsterdam.\nAll Rights Reserved.', _Printer__lines=None, _Printer__name='copyright', _Printer__dirs=(), _Printer__files=(...)) at remote 0x7ffff7ed8a10>, 'NameError': <type at remote 0x7ffff7d82060>, 'BytesWarning': <type at remote 0x7ffff...(truncated), key='build_time_vars', value=None) at Objects/dictobject.c:848 #8 0x00007ffff7a6ea8d in _PyModule_Clear (m=<optimized out>) at Objects/moduleobject.c:139 #9 0x00007ffff7aec4a1 in PyImport_Cleanup () at Python/import.c:512 #10 0x00007ffff7af957b in Py_Finalize () at Python/pythonrun.c:458 #11 0x00007ffff7b0f8e5 in Py_Main (argc=<optimized out>, argv=<optimized out>) at Modules/main.c:670 #12 0x00007ffff6d04830 in __libc_start_main (main=0x4007f0 <main>, argc=2, argv=0x7fffffffe008, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffdff8) at ../csu/libc-start.c:291 #13 0x0000000000400729 in _start () ```
1.0
segfault in python multithreaded setting - Zihang Dai reports (and I've reproduced) that the autograd engine is not thread-safe. Here's a repro script: https://gist.github.com/zihangdai/fc8f76fbb8a0f6323a6b31e6d98ceb50 Run it a few times, occassionally it segfaults. Segfault is from a much different location, when cleaning up imports: ``` Thread 1 "python" received signal SIGSEGV, Segmentation fault. malloc_consolidate (av=av@entry=0x7ffff70a8b20 <main_arena>) at malloc.c:4167 4167 malloc.c: No such file or directory. (gdb) where #0 malloc_consolidate (av=av@entry=0x7ffff70a8b20 <main_arena>) at malloc.c:4167 #1 0x00007ffff6d64678 in _int_free (av=0x7ffff70a8b20 <main_arena>, p=<optimized out>, have_lock=0) at malloc.c:4075 #2 0x00007ffff6d6853c in __GI___libc_free (mem=<optimized out>) at malloc.c:2968 #3 0x00007ffff7a6a427 in dict_dealloc (mp=0x7ffff7ed0c58) at Objects/dictobject.c:1044 #4 0x00007ffff7a682c7 in insertdict_by_entry (mp=0x7ffff7ed0d70, key='build_time_vars', hash=<optimized out>, ep=<optimized out>, value=<optimized out>) at Objects/dictobject.c:519 #5 0x00007ffff7a6b79c in insertdict (value=None, hash=-295987683324531010, key='build_time_vars', mp=0x7ffff7ed0d70) at Objects/dictobject.c:556 #6 dict_set_item_by_hash_or_entry (value=None, ep=0x0, hash=-295987683324531010, key='build_time_vars', op={'__builtins__': {'bytearray': <type at remote 0x7ffff7d7c300>, 'IndexError': <type at remote 0x7ffff7d82bc0>, 'all': <built-in function all>, 'help': <_Helper at remote 0x7ffff7ed8c50>, 'vars': <built-in function vars>, 'SyntaxError': <type at remote 0x7ffff7d82540>, 'unicode': <type at remote 0x7ffff7d99040>, 'UnicodeDecodeError': <type at remote 0x7ffff7d833e0>, 'memoryview': <type at remote 0x7ffff7d8d900>, 'isinstance': <built-in function isinstance>, 'copyright': <_Printer(_Printer__data='Copyright (c) 2001-2016 Python Software Foundation.\nAll Rights Reserved.\n\nCopyright (c) 2000 BeOpen.com.\nAll Rights Reserved.\n\nCopyright (c) 1995-2001 Corporation for National Research Initiatives.\nAll Rights Reserved.\n\nCopyright (c) 1991-1995 Stichting Mathematisch Centrum, Amsterdam.\nAll Rights Reserved.', _Printer__lines=None, _Printer__name='copyright', _Printer__dirs=(), _Printer__files=(...)) at remote 0x7ffff7ed8a10>, 'NameError': <type at remote 0x7ffff7d82060>, 'BytesWarning': <type at remote 0x7ffff...(truncated)) at Objects/dictobject.c:795 #7 PyDict_SetItem ( op={'__builtins__': {'bytearray': <type at remote 0x7ffff7d7c300>, 'IndexError': <type at remote 0x7ffff7d82bc0>, 'all': <built-in function all>, 'help': <_Helper at remote 0x7ffff7ed8c50>, 'vars': <built-in function vars>, 'SyntaxError': <type at remote 0x7ffff7d82540>, 'unicode': <type at remote 0x7ffff7d99040>, 'UnicodeDecodeError': <type at remote 0x7ffff7d833e0>, 'memoryview': <type at remote 0x7ffff7d8d900>, 'isinstance': <built-in function isinstance>, 'copyright': <_Printer(_Printer__data='Copyright (c) 2001-2016 Python Software Foundation.\nAll Rights Reserved.\n\nCopyright (c) 2000 BeOpen.com.\nAll Rights Reserved.\n\nCopyright (c) 1995-2001 Corporation for National Research Initiatives.\nAll Rights Reserved.\n\nCopyright (c) 1991-1995 Stichting Mathematisch Centrum, Amsterdam.\nAll Rights Reserved.', _Printer__lines=None, _Printer__name='copyright', _Printer__dirs=(), _Printer__files=(...)) at remote 0x7ffff7ed8a10>, 'NameError': <type at remote 0x7ffff7d82060>, 'BytesWarning': <type at remote 0x7ffff...(truncated), key='build_time_vars', value=None) at Objects/dictobject.c:848 #8 0x00007ffff7a6ea8d in _PyModule_Clear (m=<optimized out>) at Objects/moduleobject.c:139 #9 0x00007ffff7aec4a1 in PyImport_Cleanup () at Python/import.c:512 #10 0x00007ffff7af957b in Py_Finalize () at Python/pythonrun.c:458 #11 0x00007ffff7b0f8e5 in Py_Main (argc=<optimized out>, argv=<optimized out>) at Modules/main.c:670 #12 0x00007ffff6d04830 in __libc_start_main (main=0x4007f0 <main>, argc=2, argv=0x7fffffffe008, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffdff8) at ../csu/libc-start.c:291 #13 0x0000000000400729 in _start () ```
priority
segfault in python multithreaded setting zihang dai reports and i ve reproduced that the autograd engine is not thread safe here s a repro script run it a few times occassionally it segfaults segfault is from a much different location when cleaning up imports thread python received signal sigsegv segmentation fault malloc consolidate av av entry at malloc c malloc c no such file or directory gdb where malloc consolidate av av entry at malloc c in int free av p have lock at malloc c in gi libc free mem at malloc c in dict dealloc mp at objects dictobject c in insertdict by entry mp key build time vars hash ep value at objects dictobject c in insertdict value none hash key build time vars mp at objects dictobject c dict set item by hash or entry value none ep hash key build time vars op builtins bytearray indexerror all help vars syntaxerror unicode unicodedecodeerror memoryview isinstance copyright nameerror byteswarning type at remote truncated at objects dictobject c pydict setitem op builtins bytearray indexerror all help vars syntaxerror unicode unicodedecodeerror memoryview isinstance copyright nameerror byteswarning type at remote truncated key build time vars value none at objects dictobject c in pymodule clear m at objects moduleobject c in pyimport cleanup at python import c in py finalize at python pythonrun c in py main argc argv at modules main c in libc start main main argc argv init fini rtld fini stack end at csu libc start c in start
1
205,572
7,103,388,179
IssuesEvent
2018-01-16 04:40:43
morpheus65535/bazarr
https://api.github.com/repos/morpheus65535/bazarr
closed
Option to not append language to subtitle filename
enhancement priority:medium
Bazarr appends the language of the subtitle file into it's filename - e.g. `sometitle.en.srt` - making it easy to organise multiple external subtitle tracks for the same video. However, some DLNA servers and players only serve external subtitles if they have _exactly_ the same filename. Thus, subtitle files which have the language appended to them do not get served by these devices. For users who are only interested in one language of subtitle for a video, it would nice to have an option not to append the language in the filename, in order to make them compatible with a wider number of players. Alternatively, implementation of #21 would probably solve this use-case equally well.
1.0
Option to not append language to subtitle filename - Bazarr appends the language of the subtitle file into it's filename - e.g. `sometitle.en.srt` - making it easy to organise multiple external subtitle tracks for the same video. However, some DLNA servers and players only serve external subtitles if they have _exactly_ the same filename. Thus, subtitle files which have the language appended to them do not get served by these devices. For users who are only interested in one language of subtitle for a video, it would nice to have an option not to append the language in the filename, in order to make them compatible with a wider number of players. Alternatively, implementation of #21 would probably solve this use-case equally well.
priority
option to not append language to subtitle filename bazarr appends the language of the subtitle file into it s filename e g sometitle en srt making it easy to organise multiple external subtitle tracks for the same video however some dlna servers and players only serve external subtitles if they have exactly the same filename thus subtitle files which have the language appended to them do not get served by these devices for users who are only interested in one language of subtitle for a video it would nice to have an option not to append the language in the filename in order to make them compatible with a wider number of players alternatively implementation of would probably solve this use case equally well
1
605,368
18,734,308,142
IssuesEvent
2021-11-04 04:08:01
AY2122S1-CS2103T-W13-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-W13-2/tp
closed
[PE-D] Help Window does not show User Guide on subsequent attempts
type.Bug priority.Medium severity.Low
I was successful in opening the User Guide through the Help Window after keying in `help`. I then navigated to other pages by clicking the links to your team's GitHub repository via the Help Window and closed this window. Thereafter, I re-entered `help` into the CommandBox but was presented with the webpage I last visited before I closed the window (i.e.: your team's GitHub repo) instead of the User Guide. ![image.png](https://raw.githubusercontent.com/stanley-1/ped/main/files/49953a4e-913c-4305-88a5-6d4a51aa3062.png) <!--session: 1635494331267-38517f32-21de-436c-8367-a94de66a554e--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.High` `type.FeatureFlaw` original: stanley-1/ped#3
1.0
[PE-D] Help Window does not show User Guide on subsequent attempts - I was successful in opening the User Guide through the Help Window after keying in `help`. I then navigated to other pages by clicking the links to your team's GitHub repository via the Help Window and closed this window. Thereafter, I re-entered `help` into the CommandBox but was presented with the webpage I last visited before I closed the window (i.e.: your team's GitHub repo) instead of the User Guide. ![image.png](https://raw.githubusercontent.com/stanley-1/ped/main/files/49953a4e-913c-4305-88a5-6d4a51aa3062.png) <!--session: 1635494331267-38517f32-21de-436c-8367-a94de66a554e--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.High` `type.FeatureFlaw` original: stanley-1/ped#3
priority
help window does not show user guide on subsequent attempts i was successful in opening the user guide through the help window after keying in help i then navigated to other pages by clicking the links to your team s github repository via the help window and closed this window thereafter i re entered help into the commandbox but was presented with the webpage i last visited before i closed the window i e your team s github repo instead of the user guide labels severity high type featureflaw original stanley ped
1
510,415
14,790,326,971
IssuesEvent
2021-01-12 11:52:27
canonical-web-and-design/vanilla-framework
https://api.github.com/repos/canonical-web-and-design/vanilla-framework
closed
Proposing `p-separator` for spaced <hr>
Priority: Medium
In [many sites](https://github.com/search?q=org%3Acanonical-web-and-design+p-separator&type=code), we use an in row border to separate content in the same strip. The way we do that currently is by having a local pattern called `p-separator` which when attached to an `hr` creates a horizontal divider with a few rem padding top and bottom. ## Visual ![image](https://user-images.githubusercontent.com/1413534/100770829-9bdeb500-33f5-11eb-9153-41786676c8ff.png) ## Context This is used mainly on marketing sites. The style has been copied or worse inlined on the `hr` element which means it is not resonsive. ## State None that I can think of. ## Progressive enhancement The padding should scale with the viewport.
1.0
Proposing `p-separator` for spaced <hr> - In [many sites](https://github.com/search?q=org%3Acanonical-web-and-design+p-separator&type=code), we use an in row border to separate content in the same strip. The way we do that currently is by having a local pattern called `p-separator` which when attached to an `hr` creates a horizontal divider with a few rem padding top and bottom. ## Visual ![image](https://user-images.githubusercontent.com/1413534/100770829-9bdeb500-33f5-11eb-9153-41786676c8ff.png) ## Context This is used mainly on marketing sites. The style has been copied or worse inlined on the `hr` element which means it is not resonsive. ## State None that I can think of. ## Progressive enhancement The padding should scale with the viewport.
priority
proposing p separator for spaced in we use an in row border to separate content in the same strip the way we do that currently is by having a local pattern called p separator which when attached to an hr creates a horizontal divider with a few rem padding top and bottom visual context this is used mainly on marketing sites the style has been copied or worse inlined on the hr element which means it is not resonsive state none that i can think of progressive enhancement the padding should scale with the viewport
1
303,680
9,309,504,192
IssuesEvent
2019-03-25 16:37:02
conan-io/conan-package-tools
https://api.github.com/repos/conan-io/conan-package-tools
closed
Support custom docker args for docker run command
complex: low priority: medium stage: queue type: feature
For example: `CONAN_DOCKER_ARGS` and an equivalent constructor argument. We've discovered an issue in my company where the default settings for Windows Containers is to allocate 1GB of RAM. When building particularly large visual studio projects, we get a very random and intermittant compiler error. This occurs when using multiple versions of Visual Studio Build Tools 15.6.7, 15.7.6, and 15.8.6, and the error references a random C++ template parameter each time, from a variety of different libraries we use. ``` error C1001: An internal error has occurred in the compiler. compiler file 'msc1.cpp', line 1511 ``` There are many reports about this error, and Microsoft explains that it's a very generic error that tells them almost nothing. For us, we've found that it doesn't happen unless we're in Docker, and only on one of our docker hosts, and only on this really big project. So, we discovered a workaround to be to pass the following additional argument to Docker: `-m 2gb` This doubles the RAM allocated to each container, and the problem seems to have gone away. However, it's very likely that simply applying this to all builds as part of CPT would have unexpected consequences for other users or even our other projects. Additionally, there are countless other docker flags which might be necessary to set under certain circumstances. As a result, this request is for a primitive catch all string variable that would give users some control over the docker process they are launching, without complicating CPT too much.
1.0
Support custom docker args for docker run command - For example: `CONAN_DOCKER_ARGS` and an equivalent constructor argument. We've discovered an issue in my company where the default settings for Windows Containers is to allocate 1GB of RAM. When building particularly large visual studio projects, we get a very random and intermittant compiler error. This occurs when using multiple versions of Visual Studio Build Tools 15.6.7, 15.7.6, and 15.8.6, and the error references a random C++ template parameter each time, from a variety of different libraries we use. ``` error C1001: An internal error has occurred in the compiler. compiler file 'msc1.cpp', line 1511 ``` There are many reports about this error, and Microsoft explains that it's a very generic error that tells them almost nothing. For us, we've found that it doesn't happen unless we're in Docker, and only on one of our docker hosts, and only on this really big project. So, we discovered a workaround to be to pass the following additional argument to Docker: `-m 2gb` This doubles the RAM allocated to each container, and the problem seems to have gone away. However, it's very likely that simply applying this to all builds as part of CPT would have unexpected consequences for other users or even our other projects. Additionally, there are countless other docker flags which might be necessary to set under certain circumstances. As a result, this request is for a primitive catch all string variable that would give users some control over the docker process they are launching, without complicating CPT too much.
priority
support custom docker args for docker run command for example conan docker args and an equivalent constructor argument we ve discovered an issue in my company where the default settings for windows containers is to allocate of ram when building particularly large visual studio projects we get a very random and intermittant compiler error this occurs when using multiple versions of visual studio build tools and and the error references a random c template parameter each time from a variety of different libraries we use error an internal error has occurred in the compiler compiler file cpp line there are many reports about this error and microsoft explains that it s a very generic error that tells them almost nothing for us we ve found that it doesn t happen unless we re in docker and only on one of our docker hosts and only on this really big project so we discovered a workaround to be to pass the following additional argument to docker m this doubles the ram allocated to each container and the problem seems to have gone away however it s very likely that simply applying this to all builds as part of cpt would have unexpected consequences for other users or even our other projects additionally there are countless other docker flags which might be necessary to set under certain circumstances as a result this request is for a primitive catch all string variable that would give users some control over the docker process they are launching without complicating cpt too much
1
279,996
8,676,780,459
IssuesEvent
2018-11-30 15:03:48
openshiftio/openshift.io
https://api.github.com/repos/openshiftio/openshift.io
closed
If a user's attempt to reset his OSIO environment fails, no root cause, or suggested actions are provided
SEV3-medium area/user/tenant issue/intermittent priority/P3 team/platform train/heather type/bug
Steps to recreate: * Reset a user's environment The reset often fails, but a root cause for the failure is not communicated to the user. Generally, it is possible to reset the environment if a 2nd (or 3rd) attempt is made. We should present error information, and a suggestion for the user's next steps. ![screenshot 79](https://user-images.githubusercontent.com/642621/33886993-52a8d94c-df16-11e7-8bb7-520be2130987.png)
1.0
If a user's attempt to reset his OSIO environment fails, no root cause, or suggested actions are provided - Steps to recreate: * Reset a user's environment The reset often fails, but a root cause for the failure is not communicated to the user. Generally, it is possible to reset the environment if a 2nd (or 3rd) attempt is made. We should present error information, and a suggestion for the user's next steps. ![screenshot 79](https://user-images.githubusercontent.com/642621/33886993-52a8d94c-df16-11e7-8bb7-520be2130987.png)
priority
if a user s attempt to reset his osio environment fails no root cause or suggested actions are provided steps to recreate reset a user s environment the reset often fails but a root cause for the failure is not communicated to the user generally it is possible to reset the environment if a or attempt is made we should present error information and a suggestion for the user s next steps
1
595,090
18,059,695,808
IssuesEvent
2021-09-20 12:43:52
AY2122S1-CS2103-W14-4/tp
https://api.github.com/repos/AY2122S1-CS2103-W14-4/tp
opened
Sort groups & students by their tP / iP progress (by ascending/descending order)
type.Story priority.Medium
As a user I can sort groups & students by their tP / iP progress (by ascending/descending order) so that I can see who needs help.
1.0
Sort groups & students by their tP / iP progress (by ascending/descending order) - As a user I can sort groups & students by their tP / iP progress (by ascending/descending order) so that I can see who needs help.
priority
sort groups students by their tp ip progress by ascending descending order as a user i can sort groups students by their tp ip progress by ascending descending order so that i can see who needs help
1
356,113
10,588,985,248
IssuesEvent
2019-10-09 04:16:33
AY1920S1-CS2103T-T09-1/main
https://api.github.com/repos/AY1920S1-CS2103T-T09-1/main
opened
As a fast typist, I can be auto-corrected for my typos
priority.Medium type.Story
so that I won't break my train of thought while planning
1.0
As a fast typist, I can be auto-corrected for my typos - so that I won't break my train of thought while planning
priority
as a fast typist i can be auto corrected for my typos so that i won t break my train of thought while planning
1
289,144
8,855,198,425
IssuesEvent
2019-01-09 05:15:37
visit-dav/issues-test
https://api.github.com/repos/visit-dav/issues-test
closed
XRay Image query crashes if you give it bad variable names for the absorptivity and emissivity.
bug crash likelihood medium priority reviewed severity high wrong results
Paul Amala was using the XRay Image query and forgot to set the absorptivity and emmissivity variable names and VisIt crashed because they were wrong. The engine crashed and then the viewer was in such a state that even after restarting the engine, the engine would immediately crash again. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2310 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: XRay Image query crashes if you give it bad variable names for the absorptivity and emissivity. Assigned to: Kevin Griffin Category: - Target version: 2.12.0 Author: Eric Brugger Start: 06/24/2015 Due date: % Done: 100% Estimated time: Created: 06/24/2015 03:41 pm Updated: 10/25/2016 01:41 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.9.2 Impact: Expected Use: OS: All Support Group: Any Description: Paul Amala was using the XRay Image query and forgot to set the absorptivity and emmissivity variable names and VisIt crashed because they were wrong. The engine crashed and then the viewer was in such a state that even after restarting the engine, the engine would immediately crash again. Comments: Hello:I added a check missing variables in the Xray Image Query and also made the error messages a little more verbose (Bug #2310). Will update the release notes once the relnotes2.11.1.html file is created.2.11RC:Sending avt/Filters/avtXRayFilter.CSending avt/Queries/Queries/avtXRayImageQuery.CSending avt/Queries/Queries/avtXRayImageQuery.hTransmitting file data ...doneCommitting transaction...Committed revision 29337.TRUNK:Sending avt/Filters/avtXRayFilter.CSending avt/Queries/Queries/avtXRayImageQuery.CSending avt/Queries/Queries/avtXRayImageQuery.hTransmitting file data ...doneCommitting transaction...Committed revision 29339.-Kevin
1.0
XRay Image query crashes if you give it bad variable names for the absorptivity and emissivity. - Paul Amala was using the XRay Image query and forgot to set the absorptivity and emmissivity variable names and VisIt crashed because they were wrong. The engine crashed and then the viewer was in such a state that even after restarting the engine, the engine would immediately crash again. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2310 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: XRay Image query crashes if you give it bad variable names for the absorptivity and emissivity. Assigned to: Kevin Griffin Category: - Target version: 2.12.0 Author: Eric Brugger Start: 06/24/2015 Due date: % Done: 100% Estimated time: Created: 06/24/2015 03:41 pm Updated: 10/25/2016 01:41 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.9.2 Impact: Expected Use: OS: All Support Group: Any Description: Paul Amala was using the XRay Image query and forgot to set the absorptivity and emmissivity variable names and VisIt crashed because they were wrong. The engine crashed and then the viewer was in such a state that even after restarting the engine, the engine would immediately crash again. Comments: Hello:I added a check missing variables in the Xray Image Query and also made the error messages a little more verbose (Bug #2310). Will update the release notes once the relnotes2.11.1.html file is created.2.11RC:Sending avt/Filters/avtXRayFilter.CSending avt/Queries/Queries/avtXRayImageQuery.CSending avt/Queries/Queries/avtXRayImageQuery.hTransmitting file data ...doneCommitting transaction...Committed revision 29337.TRUNK:Sending avt/Filters/avtXRayFilter.CSending avt/Queries/Queries/avtXRayImageQuery.CSending avt/Queries/Queries/avtXRayImageQuery.hTransmitting file data ...doneCommitting transaction...Committed revision 29339.-Kevin
priority
xray image query crashes if you give it bad variable names for the absorptivity and emissivity paul amala was using the xray image query and forgot to set the absorptivity and emmissivity variable names and visit crashed because they were wrong the engine crashed and then the viewer was in such a state that even after restarting the engine the engine would immediately crash again redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject xray image query crashes if you give it bad variable names for the absorptivity and emissivity assigned to kevin griffin category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description paul amala was using the xray image query and forgot to set the absorptivity and emmissivity variable names and visit crashed because they were wrong the engine crashed and then the viewer was in such a state that even after restarting the engine the engine would immediately crash again comments hello i added a check missing variables in the xray image query and also made the error messages a little more verbose bug will update the release notes once the html file is created sending avt filters avtxrayfilter csending avt queries queries avtxrayimagequery csending avt queries queries avtxrayimagequery htransmitting file data donecommitting transaction committed revision trunk sending avt filters avtxrayfilter csending avt queries queries avtxrayimagequery csending avt queries queries avtxrayimagequery htransmitting file data donecommitting transaction committed revision kevin
1
22,151
2,645,693,733
IssuesEvent
2015-03-13 01:10:52
prikhi/evoluspencil
https://api.github.com/repos/prikhi/evoluspencil
closed
Please add the ability to create a new page size @ a specific size
1 star bug imported Priority-Medium
_From [tony.sev...@gmail.com](https://code.google.com/u/111188139539534463642/) on August 25, 2008 13:57:31_ New docs seem to take on the document window size at the time of creation but it would be good to to define the document size in specific px. _Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=43_
1.0
Please add the ability to create a new page size @ a specific size - _From [tony.sev...@gmail.com](https://code.google.com/u/111188139539534463642/) on August 25, 2008 13:57:31_ New docs seem to take on the document window size at the time of creation but it would be good to to define the document size in specific px. _Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=43_
priority
please add the ability to create a new page size a specific size from on august new docs seem to take on the document window size at the time of creation but it would be good to to define the document size in specific px original issue
1
109,243
4,383,163,184
IssuesEvent
2016-08-07 10:49:28
madewithlove/laravel-cqrs-es
https://api.github.com/repos/madewithlove/laravel-cqrs-es
closed
Reset read models when doing event-store:replay
Priority: Medium Status: Completed Type: Enhancement
Currently when doing `event-store:replay` it will still keep old read models around. Ideally these should be cleared. Since read models can be stored in different drivers we should write the logic ourselves but instead provide an API so that people can hook into these events. My proposal is to trigger the following application events: `event-store.replay.before` `event-store.replay.after` This would allow package consumers to listen for these events and executed the necessary logic.
1.0
Reset read models when doing event-store:replay - Currently when doing `event-store:replay` it will still keep old read models around. Ideally these should be cleared. Since read models can be stored in different drivers we should write the logic ourselves but instead provide an API so that people can hook into these events. My proposal is to trigger the following application events: `event-store.replay.before` `event-store.replay.after` This would allow package consumers to listen for these events and executed the necessary logic.
priority
reset read models when doing event store replay currently when doing event store replay it will still keep old read models around ideally these should be cleared since read models can be stored in different drivers we should write the logic ourselves but instead provide an api so that people can hook into these events my proposal is to trigger the following application events event store replay before event store replay after this would allow package consumers to listen for these events and executed the necessary logic
1
255,602
8,125,821,051
IssuesEvent
2018-08-16 22:29:23
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Provide non-mesa linux as well as with-mesa distros.
Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal
Our linux with-mesa distributions (started with 2.9.0) have caused problems for some users (Doug Reeder is one). https://elist.ornl.gov/pipermail/visit-users/2015-June/017430.html When I provided Doug with a non-mesa version, his problems went away. Perhaps we should provide both mesa and non-mesa versions. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2312 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: Provide non-mesa linux as well as with-mesa distros. Assigned to: Eric Brugger Category: Target version: 2.10 Author: Kathleen Biagas Start: 06/26/2015 Due date: % Done: 100 Estimated time: 6.0 Created: 06/26/2015 02:04 pm Updated: 10/07/2015 02:57 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: Our linux with-mesa distributions (started with 2.9.0) have caused problems for some users (Doug Reeder is one). https://elist.ornl.gov/pipermail/visit-users/2015-June/017430.html When I provided Doug with a non-mesa version, his problems went away. Perhaps we should provide both mesa and non-mesa versions. Comments: I committed revisions 27281 and 27283 to the 2.10 RC and trunk with thefollowing change:1) I made a number of changes to the open build and install scripts. I removed sidious. I updated them to use the 2.10.0 version of cmake. I modified the kickit build to build a version with mesa and a version without.M svn_bin/visit-build-openM svn_bin/visit-install-openI committed revisions 27377 and 27379 to the 2.10 RC and trunk with thefollowing change:1) I rebuilt the third party libraries for 2.10 on ubuntu1104_64. I also enhanced the open build script to build both with and without mesa versions of the distribution.M config-site/ubuntu1104-64.cmakeM svn_bin/visit-build-openM svn_bin/visit-install-openI committed revisions 27419 and 27421 to the 2.10 RC and trunk with thefollowing change:1) I rebuilt the third party libraries for 2.10 on ubuntu1404_64. I also enhanced the open build script to build both with and without mesa versions of the distribution. This together with checkins 27281 and 27377 (2.10RC) and 27283 and 27379 (trunk) resolves #2312.M config-site/ubuntu1404-64.sqalab-vmware.box.cmakeM svn_bin/visit-build-openM svn_bin/visit-install-open
1.0
Provide non-mesa linux as well as with-mesa distros. - Our linux with-mesa distributions (started with 2.9.0) have caused problems for some users (Doug Reeder is one). https://elist.ornl.gov/pipermail/visit-users/2015-June/017430.html When I provided Doug with a non-mesa version, his problems went away. Perhaps we should provide both mesa and non-mesa versions. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2312 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: Provide non-mesa linux as well as with-mesa distros. Assigned to: Eric Brugger Category: Target version: 2.10 Author: Kathleen Biagas Start: 06/26/2015 Due date: % Done: 100 Estimated time: 6.0 Created: 06/26/2015 02:04 pm Updated: 10/07/2015 02:57 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: Our linux with-mesa distributions (started with 2.9.0) have caused problems for some users (Doug Reeder is one). https://elist.ornl.gov/pipermail/visit-users/2015-June/017430.html When I provided Doug with a non-mesa version, his problems went away. Perhaps we should provide both mesa and non-mesa versions. Comments: I committed revisions 27281 and 27283 to the 2.10 RC and trunk with thefollowing change:1) I made a number of changes to the open build and install scripts. I removed sidious. I updated them to use the 2.10.0 version of cmake. I modified the kickit build to build a version with mesa and a version without.M svn_bin/visit-build-openM svn_bin/visit-install-openI committed revisions 27377 and 27379 to the 2.10 RC and trunk with thefollowing change:1) I rebuilt the third party libraries for 2.10 on ubuntu1104_64. I also enhanced the open build script to build both with and without mesa versions of the distribution.M config-site/ubuntu1104-64.cmakeM svn_bin/visit-build-openM svn_bin/visit-install-openI committed revisions 27419 and 27421 to the 2.10 RC and trunk with thefollowing change:1) I rebuilt the third party libraries for 2.10 on ubuntu1404_64. I also enhanced the open build script to build both with and without mesa versions of the distribution. This together with checkins 27281 and 27377 (2.10RC) and 27283 and 27379 (trunk) resolves #2312.M config-site/ubuntu1404-64.sqalab-vmware.box.cmakeM svn_bin/visit-build-openM svn_bin/visit-install-open
priority
provide non mesa linux as well as with mesa distros our linux with mesa distributions started with have caused problems for some users doug reeder is one when i provided doug with a non mesa version his problems went away perhaps we should provide both mesa and non mesa versions redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject provide non mesa linux as well as with mesa distros assigned to eric brugger category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood severity found in version impact medium expected use occasional os all support group any description our linux with mesa distributions started with have caused problems for some users doug reeder is one when i provided doug with a non mesa version his problems went away perhaps we should provide both mesa and non mesa versions comments i committed revisions and to the rc and trunk with thefollowing change i made a number of changes to the open build and install scripts i removed sidious i updated them to use the version of cmake i modified the kickit build to build a version with mesa and a version without m svn bin visit build openm svn bin visit install openi committed revisions and to the rc and trunk with thefollowing change i rebuilt the third party libraries for on i also enhanced the open build script to build both with and without mesa versions of the distribution m config site cmakem svn bin visit build openm svn bin visit install openi committed revisions and to the rc and trunk with thefollowing change i rebuilt the third party libraries for on i also enhanced the open build script to build both with and without mesa versions of the distribution this together with checkins and and and trunk resolves m config site sqalab vmware box cmakem svn bin visit build openm svn bin visit install open
1
349,475
10,469,993,043
IssuesEvent
2019-09-23 01:11:33
ssec/sift
https://api.github.com/repos/ssec/sift
closed
Fix numba warnings when opening SIFT
optimization priority: medium
This is likely a change in numba which shows warnings by default, but @katherinekolman noticed them. Numba is now telling us that our numba.jit functions are not as optimized as they should be. We at least want to get rid of the warnings, but making them optimized would be good too.
1.0
Fix numba warnings when opening SIFT - This is likely a change in numba which shows warnings by default, but @katherinekolman noticed them. Numba is now telling us that our numba.jit functions are not as optimized as they should be. We at least want to get rid of the warnings, but making them optimized would be good too.
priority
fix numba warnings when opening sift this is likely a change in numba which shows warnings by default but katherinekolman noticed them numba is now telling us that our numba jit functions are not as optimized as they should be we at least want to get rid of the warnings but making them optimized would be good too
1
170,787
6,471,567,897
IssuesEvent
2017-08-17 11:59:59
systers/powerup-android
https://api.github.com/repos/systers/powerup-android
closed
Developer needs UI elements for the Home Scenario -1
Priority: MEDIUM Program: GSoC17 Type: Enhancement
## Description As a developer, I want to see the UIs of home party scene, so that I can code the stage Home Scenario- 1. ## Mocks N/A ## Acceptance Criteria -Background ## Definition of Done -Passes user testing per acceptance criteria items. -Featured in design meeting. -Approved by UI team. (Dil & May) ## Estimation
1.0
Developer needs UI elements for the Home Scenario -1 - ## Description As a developer, I want to see the UIs of home party scene, so that I can code the stage Home Scenario- 1. ## Mocks N/A ## Acceptance Criteria -Background ## Definition of Done -Passes user testing per acceptance criteria items. -Featured in design meeting. -Approved by UI team. (Dil & May) ## Estimation
priority
developer needs ui elements for the home scenario description as a developer i want to see the uis of home party scene so that i can code the stage home scenario mocks n a acceptance criteria background definition of done passes user testing per acceptance criteria items featured in design meeting approved by ui team dil may estimation
1
721,762
24,836,956,816
IssuesEvent
2022-10-26 09:35:23
AY2223S1-CS2103T-T12-1/tp
https://api.github.com/repos/AY2223S1-CS2103T-T12-1/tp
closed
Allow model to access unfiltered list of students.
enhancement priority.MEDIUM type.Enhancement
Perhaps it would be better if we can move `deleteModuleClass` logic to `Model` to improve abstraction. _Originally posted by @Berted in https://github.com/AY2223S1-CS2103T-T12-1/tp/pull/137#discussion_r995335084_
1.0
Allow model to access unfiltered list of students. - Perhaps it would be better if we can move `deleteModuleClass` logic to `Model` to improve abstraction. _Originally posted by @Berted in https://github.com/AY2223S1-CS2103T-T12-1/tp/pull/137#discussion_r995335084_
priority
allow model to access unfiltered list of students perhaps it would be better if we can move deletemoduleclass logic to model to improve abstraction originally posted by berted in
1
484,628
13,943,032,025
IssuesEvent
2020-10-22 22:09:16
cloudfoundry-incubator/kubecf
https://api.github.com/repos/cloudfoundry-incubator/kubecf
closed
Unable to deploy KubeCF 2.5.6/7/8
Priority: Medium Status: Validation Type: Bug
**Describe the bug** Unable to deploy KubeCF 2.5.6/7/8 on AWS EKS. The cf-apps-dns pod remains stuck in the pod initialization stage forever. Describing the pod shows the following log MountVolume.SetUp failed for volume "client-tls" : secret "var-cf-app-sd-client-tls" not found **To Reproduce** Use a simple overrides file for values.yml and deploy cf-operator and kubecf. #####overrides.yaml###### ` system_domain: <mydomain> kube: storage_class: gp2 credentials: cf_admin_password: <strong_pass> uaa_admin_client_secret: <yet_another_strong_pass> features: autoscaler: enabled: false ` ########### helm3 install cf-operator \ --namespace cf-operator \ --set "global.operator.watchNamespace=kubecf" \ https://s3.amazonaws.com/cf-operators/release/helm-charts/cf-operator-6.1.15%2B0.g89a56300.tgz helm3 install kubecf \ --namespace kubecf \ --values overrides.yaml \ ./kubecf-v2.5.6.tgz **Environment** - KubeCF Version 2.5.6 - cf-operator 6.1.15 - AWS EKS (Kubernetes 1.17.9) - OS Image -- Amazon Linux 2 - Container runtime -- docker://19.3.6 **Additional context** **Please note:- KubeCF 2.2.2 deployment works and always succeeds on the same cluster.**
1.0
Unable to deploy KubeCF 2.5.6/7/8 - **Describe the bug** Unable to deploy KubeCF 2.5.6/7/8 on AWS EKS. The cf-apps-dns pod remains stuck in the pod initialization stage forever. Describing the pod shows the following log MountVolume.SetUp failed for volume "client-tls" : secret "var-cf-app-sd-client-tls" not found **To Reproduce** Use a simple overrides file for values.yml and deploy cf-operator and kubecf. #####overrides.yaml###### ` system_domain: <mydomain> kube: storage_class: gp2 credentials: cf_admin_password: <strong_pass> uaa_admin_client_secret: <yet_another_strong_pass> features: autoscaler: enabled: false ` ########### helm3 install cf-operator \ --namespace cf-operator \ --set "global.operator.watchNamespace=kubecf" \ https://s3.amazonaws.com/cf-operators/release/helm-charts/cf-operator-6.1.15%2B0.g89a56300.tgz helm3 install kubecf \ --namespace kubecf \ --values overrides.yaml \ ./kubecf-v2.5.6.tgz **Environment** - KubeCF Version 2.5.6 - cf-operator 6.1.15 - AWS EKS (Kubernetes 1.17.9) - OS Image -- Amazon Linux 2 - Container runtime -- docker://19.3.6 **Additional context** **Please note:- KubeCF 2.2.2 deployment works and always succeeds on the same cluster.**
priority
unable to deploy kubecf describe the bug unable to deploy kubecf on aws eks the cf apps dns pod remains stuck in the pod initialization stage forever describing the pod shows the following log mountvolume setup failed for volume client tls secret var cf app sd client tls not found to reproduce use a simple overrides file for values yml and deploy cf operator and kubecf overrides yaml system domain kube storage class credentials cf admin password uaa admin client secret features autoscaler enabled false install cf operator namespace cf operator set global operator watchnamespace kubecf install kubecf namespace kubecf values overrides yaml kubecf tgz environment kubecf version cf operator aws eks kubernetes os image amazon linux container runtime docker additional context please note kubecf deployment works and always succeeds on the same cluster
1
752,606
26,292,185,367
IssuesEvent
2023-01-08 14:54:12
deadw00d/AROS
https://api.github.com/repos/deadw00d/AROS
opened
"Floating" option in Wanderer Prefs causes wrong icons rendering
type:bug priority:medium
"By integrating "20200413" on AROS One x86 on "WMWare SVGA", I had encountered the same problem of corrupted icon graphics as as with Scalos (see screenshot), as if by magic the problem is solved by reconfiguring from scratch Wanderer Prefs" https://ae.amigalife.org/index.php?topic=1063.0.msg14016#msg14016
1.0
"Floating" option in Wanderer Prefs causes wrong icons rendering - "By integrating "20200413" on AROS One x86 on "WMWare SVGA", I had encountered the same problem of corrupted icon graphics as as with Scalos (see screenshot), as if by magic the problem is solved by reconfiguring from scratch Wanderer Prefs" https://ae.amigalife.org/index.php?topic=1063.0.msg14016#msg14016
priority
floating option in wanderer prefs causes wrong icons rendering by integrating on aros one on wmware svga i had encountered the same problem of corrupted icon graphics as as with scalos see screenshot as if by magic the problem is solved by reconfiguring from scratch wanderer prefs
1
113,325
4,546,781,798
IssuesEvent
2016-09-12 00:11:03
Dolibarr/dolibarr
https://api.github.com/repos/Dolibarr/dolibarr
closed
3.7 does not fill edit image alt attribute
Bug Priority Medium version 3.7.1 version 3.8.0-beta version develop
Title tag is filled instead. Both attributes are similar, but the "alt" attribute is shown when the image cannot be loaded, while the title attribute is shown to the user as a complementary text to the image. Example (from https://www.w3.org/wiki/HTML/Elements/img): ``` <img src="sales.gif" title="Sales graph" alt="From 1998 to 2005, sales increased by the following percentages with each year: 624%, 75%, 138%, 40%, 35%, 9%, 21%"> ``` As you can see in that example, the `alt` attribute describes the image while the `title` attribute does not. ![captura de pantalla 2015-07-27 a las 13 20 42](https://cloud.githubusercontent.com/assets/168744/8904751/4fc3dd3c-3462-11e5-8670-d44591184409.png)
1.0
3.7 does not fill edit image alt attribute - Title tag is filled instead. Both attributes are similar, but the "alt" attribute is shown when the image cannot be loaded, while the title attribute is shown to the user as a complementary text to the image. Example (from https://www.w3.org/wiki/HTML/Elements/img): ``` <img src="sales.gif" title="Sales graph" alt="From 1998 to 2005, sales increased by the following percentages with each year: 624%, 75%, 138%, 40%, 35%, 9%, 21%"> ``` As you can see in that example, the `alt` attribute describes the image while the `title` attribute does not. ![captura de pantalla 2015-07-27 a las 13 20 42](https://cloud.githubusercontent.com/assets/168744/8904751/4fc3dd3c-3462-11e5-8670-d44591184409.png)
priority
does not fill edit image alt attribute title tag is filled instead both attributes are similar but the alt attribute is shown when the image cannot be loaded while the title attribute is shown to the user as a complementary text to the image example from img src sales gif title sales graph alt from to sales increased by the following percentages with each year as you can see in that example the alt attribute describes the image while the title attribute does not
1
364,737
10,772,714,461
IssuesEvent
2019-11-02 16:20:17
SOSML/SOSML
https://api.github.com/repos/SOSML/SOSML
closed
Polymorphic declarations in let broken
p5: medium priority s:elaboration t:squid
In some cases, val declarations in let expressions are (seemingly) treated as ambiguous declarations: `val (a, b) = let val x = fn y => y in (x 5, x 9.0) end` does not pass elaboration although it should (and the code does work in mosml and smlnj).
1.0
Polymorphic declarations in let broken - In some cases, val declarations in let expressions are (seemingly) treated as ambiguous declarations: `val (a, b) = let val x = fn y => y in (x 5, x 9.0) end` does not pass elaboration although it should (and the code does work in mosml and smlnj).
priority
polymorphic declarations in let broken in some cases val declarations in let expressions are seemingly treated as ambiguous declarations val a b let val x fn y y in x x end does not pass elaboration although it should and the code does work in mosml and smlnj
1
73,103
3,407,210,493
IssuesEvent
2015-12-04 01:11:14
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Possible regression in waiting for DOMContentLoaded
area-dart2js area-html Priority-Medium
Opening a new bug for possible regression reported in https://github.com/dart-lang/sdk/issues/885#issuecomment-161689055 @vsmenon did the original fix. Comment from there by @joseluis is On version 1.13.0 this doesn't seem true anymore. Minimal example: querySelector('#someid'); returns null when the element is placed after the script tag that loads the Dart script compiled with dart2js. In order to make it work I must use the script defer attribute, like this: ```html <html> <head> <meta charset="utf-8"> </head> <body> <script src="dart-program.js" defer></script> <script id="someid"></script> </body> </html> ```
1.0
Possible regression in waiting for DOMContentLoaded - Opening a new bug for possible regression reported in https://github.com/dart-lang/sdk/issues/885#issuecomment-161689055 @vsmenon did the original fix. Comment from there by @joseluis is On version 1.13.0 this doesn't seem true anymore. Minimal example: querySelector('#someid'); returns null when the element is placed after the script tag that loads the Dart script compiled with dart2js. In order to make it work I must use the script defer attribute, like this: ```html <html> <head> <meta charset="utf-8"> </head> <body> <script src="dart-program.js" defer></script> <script id="someid"></script> </body> </html> ```
priority
possible regression in waiting for domcontentloaded opening a new bug for possible regression reported in vsmenon did the original fix comment from there by joseluis is on version this doesn t seem true anymore minimal example queryselector someid returns null when the element is placed after the script tag that loads the dart script compiled with in order to make it work i must use the script defer attribute like this html
1
265,651
8,357,298,275
IssuesEvent
2018-10-02 21:06:52
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
tests/kernel/fifo/fifo_timeout fails on nrf51_pca10028 and nrf52_pca10040
bug in progress nRF priority: medium
Looks like this bug has been failing on this platform since the test was introduced ``` ***** delaying boot 1000ms (per build configuration) ***** ***** Booting Zephyr OS 1.12.0-rc2 (delayed boot 1000ms) ***** Running test suite test_fifo_timeout =================================================================== starting test - test_timeout_empty_fifo PASS - test_timeout_empty_fifo =================================================================== starting test - test_timeout_non_empty_fifo PASS - test_timeout_non_empty_fifo =================================================================== starting test - test_timeout_fifo_thread PASS - test_timeout_fifo_thread =================================================================== starting test - test_timeout_threads_pend_on_fifo thread (q order: 2, t/o: 0, fifo 0x20000000) thread (q order: 3, t/o: 10, fifo 0x20000000) thread (q order: 0, t/o: 20, fifo 0x20000000) thread (q order: 4, t/o: 30, fifo 0x20000000) thread (q order: 1, t/o: 40, fifo 0x20000000) PASS - test_timeout_threads_pend_on_fifo =================================================================== starting test - test_timeout_threads_pend_on_dual_fifos thread (q order: 0, t/o: 0, fifo 0x20000010) thread (q order: 5, t/o: 10, fifo 0x20000000) FAIL - test_multiple_threads_pending@156. *** thread 3 woke up, expected 2 Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:386: test_timeout_threads_pend_on_dual_fifos: rv not equal to TC_PASS FAIL - test_timeout_threads_pend_on_dual_fifos =================================================================== starting test - test_timeout_threads_pend_fail_on_fifo FAIL - test_multiple_threads_get_data@207. *** thread 7 woke up, expected 0 Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:401: test_timeout_threads_pend_fail_on_fifo: rv not equal to TC_PASS FAIL - test_timeout_threads_pend_fail_on_fifo =================================================================== =================================================================== RunID: ci-180601-1931-1716:k8if PROJECT EXECUTION FAILED Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false ``` This test also fails with "prj_poll.conf" that is CONFIG_POLL=y in the latest commit: 5b8e4ae4df378d0da32ce36266246fb530a6309d ``` ***** delaying boot 1000ms (per build configuration) ***** ***** Booting Zephyr OS v1.12.0-831-g5b8e4ae (delayed boot 1000ms) ***** Running test suite test_fifo_timeout =================================================================== starting test - test_timeout_empty_fifo PASS - test_timeout_empty_fifo =================================================================== starting test - test_timeout_non_empty_fifo PASS - test_timeout_non_empty_fifo =================================================================== starting test - test_timeout_fifo_thread PASS - test_timeout_fifo_thread =================================================================== starting test - test_timeout_threads_pend_on_fifo thread (q order: 2, t/o: 0, fifo 0x20000000) thread (q order: 3, t/o: 10, fifo 0x20000000) thread (q order: 0, t/o: 20, fifo 0x20000000) thread (q order: 4, t/o: 30, fifo 0x20000000) thread (q order: 1, t/o: 40, fifo 0x20000000) PASS - test_timeout_threads_pend_on_fifo =================================================================== starting test - test_timeout_threads_pend_on_dual_fifos thread (q order: 0, t/o: 0, fifo 0x20000010) thread (q order: 5, t/o: 10, fifo 0x20000000) FAIL - test_multiple_threads_pending@156. *** thread 3 woke up, expected 2 Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:396: test_timeout_threads_pend_on_dual_fifos: rv not equal to TC_PASS FAIL - test_timeout_threads_pend_on_dual_fifos =================================================================== starting test - test_timeout_threads_pend_fail_on_fifo FAIL - test_multiple_threads_get_data@207. *** thread 7 woke up, expected 0 Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:411: test_timeout_threads_pend_fail_on_fifo: rv not equal to TC_PASS Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:129: test_thread_pend_and_timeout: packet == NULL is false FAIL - test_timeout_threads_pend_fail_on_fifo =================================================================== =================================================================== RunID: :wjv2 PROJECT EXECUTION FAILED Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false ``` This failure(with CONFIG_POLL=y) was not seen in earlier commits.
1.0
tests/kernel/fifo/fifo_timeout fails on nrf51_pca10028 and nrf52_pca10040 - Looks like this bug has been failing on this platform since the test was introduced ``` ***** delaying boot 1000ms (per build configuration) ***** ***** Booting Zephyr OS 1.12.0-rc2 (delayed boot 1000ms) ***** Running test suite test_fifo_timeout =================================================================== starting test - test_timeout_empty_fifo PASS - test_timeout_empty_fifo =================================================================== starting test - test_timeout_non_empty_fifo PASS - test_timeout_non_empty_fifo =================================================================== starting test - test_timeout_fifo_thread PASS - test_timeout_fifo_thread =================================================================== starting test - test_timeout_threads_pend_on_fifo thread (q order: 2, t/o: 0, fifo 0x20000000) thread (q order: 3, t/o: 10, fifo 0x20000000) thread (q order: 0, t/o: 20, fifo 0x20000000) thread (q order: 4, t/o: 30, fifo 0x20000000) thread (q order: 1, t/o: 40, fifo 0x20000000) PASS - test_timeout_threads_pend_on_fifo =================================================================== starting test - test_timeout_threads_pend_on_dual_fifos thread (q order: 0, t/o: 0, fifo 0x20000010) thread (q order: 5, t/o: 10, fifo 0x20000000) FAIL - test_multiple_threads_pending@156. *** thread 3 woke up, expected 2 Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:386: test_timeout_threads_pend_on_dual_fifos: rv not equal to TC_PASS FAIL - test_timeout_threads_pend_on_dual_fifos =================================================================== starting test - test_timeout_threads_pend_fail_on_fifo FAIL - test_multiple_threads_get_data@207. *** thread 7 woke up, expected 0 Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:401: test_timeout_threads_pend_fail_on_fifo: rv not equal to TC_PASS FAIL - test_timeout_threads_pend_fail_on_fifo =================================================================== =================================================================== RunID: ci-180601-1931-1716:k8if PROJECT EXECUTION FAILED Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/jenkins/workspace/zephyr-master-tcf-v0.11-branch/LABEL/verify/SHARD/3-3/ZEPHYR_GCC_VARIANT/zephyr/zephyr.git/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false ``` This test also fails with "prj_poll.conf" that is CONFIG_POLL=y in the latest commit: 5b8e4ae4df378d0da32ce36266246fb530a6309d ``` ***** delaying boot 1000ms (per build configuration) ***** ***** Booting Zephyr OS v1.12.0-831-g5b8e4ae (delayed boot 1000ms) ***** Running test suite test_fifo_timeout =================================================================== starting test - test_timeout_empty_fifo PASS - test_timeout_empty_fifo =================================================================== starting test - test_timeout_non_empty_fifo PASS - test_timeout_non_empty_fifo =================================================================== starting test - test_timeout_fifo_thread PASS - test_timeout_fifo_thread =================================================================== starting test - test_timeout_threads_pend_on_fifo thread (q order: 2, t/o: 0, fifo 0x20000000) thread (q order: 3, t/o: 10, fifo 0x20000000) thread (q order: 0, t/o: 20, fifo 0x20000000) thread (q order: 4, t/o: 30, fifo 0x20000000) thread (q order: 1, t/o: 40, fifo 0x20000000) PASS - test_timeout_threads_pend_on_fifo =================================================================== starting test - test_timeout_threads_pend_on_dual_fifos thread (q order: 0, t/o: 0, fifo 0x20000010) thread (q order: 5, t/o: 10, fifo 0x20000000) FAIL - test_multiple_threads_pending@156. *** thread 3 woke up, expected 2 Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:396: test_timeout_threads_pend_on_dual_fifos: rv not equal to TC_PASS FAIL - test_timeout_threads_pend_on_dual_fifos =================================================================== starting test - test_timeout_threads_pend_fail_on_fifo FAIL - test_multiple_threads_get_data@207. *** thread 7 woke up, expected 0 Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:411: test_timeout_threads_pend_fail_on_fifo: rv not equal to TC_PASS Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:129: test_thread_pend_and_timeout: packet == NULL is false FAIL - test_timeout_threads_pend_fail_on_fifo =================================================================== =================================================================== RunID: :wjv2 PROJECT EXECUTION FAILED Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false Assertion failed at /home/pswarnak/workspace/1.12_execution/zephyr/tests/kernel/fifo/fifo_timeout/src/main.c:171: test_thread_pend_and_get_data: packet != NULL is false ``` This failure(with CONFIG_POLL=y) was not seen in earlier commits.
priority
tests kernel fifo fifo timeout fails on and looks like this bug has been failing on this platform since the test was introduced delaying boot per build configuration booting zephyr os delayed boot running test suite test fifo timeout starting test test timeout empty fifo pass test timeout empty fifo starting test test timeout non empty fifo pass test timeout non empty fifo starting test test timeout fifo thread pass test timeout fifo thread starting test test timeout threads pend on fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo pass test timeout threads pend on fifo starting test test timeout threads pend on dual fifos thread q order t o fifo thread q order t o fifo fail test multiple threads pending thread woke up expected assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test timeout threads pend on dual fifos rv not equal to tc pass fail test timeout threads pend on dual fifos starting test test timeout threads pend fail on fifo fail test multiple threads get data thread woke up expected assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test timeout threads pend fail on fifo rv not equal to tc pass fail test timeout threads pend fail on fifo runid ci project execution failed assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home jenkins workspace zephyr master tcf branch label verify shard zephyr gcc variant zephyr zephyr git tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false this test also fails with prj poll conf that is config poll y in the latest commit delaying boot per build configuration booting zephyr os delayed boot running test suite test fifo timeout starting test test timeout empty fifo pass test timeout empty fifo starting test test timeout non empty fifo pass test timeout non empty fifo starting test test timeout fifo thread pass test timeout fifo thread starting test test timeout threads pend on fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo thread q order t o fifo pass test timeout threads pend on fifo starting test test timeout threads pend on dual fifos thread q order t o fifo thread q order t o fifo fail test multiple threads pending thread woke up expected assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test timeout threads pend on dual fifos rv not equal to tc pass fail test timeout threads pend on dual fifos starting test test timeout threads pend fail on fifo fail test multiple threads get data thread woke up expected assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test timeout threads pend fail on fifo rv not equal to tc pass assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and timeout packet null is false fail test timeout threads pend fail on fifo runid project execution failed assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false assertion failed at home pswarnak workspace execution zephyr tests kernel fifo fifo timeout src main c test thread pend and get data packet null is false this failure with config poll y was not seen in earlier commits
1
597,787
18,171,823,838
IssuesEvent
2021-09-27 20:58:16
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Clean up react warnings
type:task priority:medium component:ui state:needs_review
##### SUMMARY Clean up react warnings. Go to console. See: ``` Warning: Can't perform a React state update on an unmounted component. This is a no-op, but it indicates a memory leak in your application. To fix, cancel all subscriptions and asynchronous tasks in a useEffect cleanup function. in InventorySourceAdd (at InventorySources.js:12) in Route (at InventorySources.js:11) index.js:1``` ``` ``` Warning: React does not recognize the `isValid` prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase `isvalid` instead. If you accidentally passed it from a parent component, remove it from the DOM element. in ul (created by Context.Consumer) in SelectMenu (created by ForwardRef) in ForwardRef (created by Select) in div (created by Select) in GenerateId (created by Select) in Select (at SCMSubForm.js:113) in div (created by FormGroup) in div (created by FormGroup) in FormGroup (at SCMSubForm.js:94) in SCMSubForm (at InventorySourceForm.js:204) in div (created by styled.div) in styled.div (at InventorySourceForm.js:155) in div (created by styled.div) in styled.div (at InventorySourceForm.js:151) in InventorySourceFormFields (at InventorySourceForm.js:298) in div (created by styled.div) in styled.div (at InventorySourceForm.js:297) ``` One more - Logged out and see ``` Warning: Cannot update a component (`SessionProvider`) while rendering a different component (`ProtectedRoute`). To locate the bad setState() call inside `ProtectedRoute`, follow the stack trace as described in https://fb.me/setstate-in-render in ProtectedRoute (at App.js:166) in Switch (at App.js:156) in SessionProvider (at App.js:155) in Unknown (at App.js:154) in I18nProvider (at App.js:153) in App (at App.js:180) in Router (created by HashRouter) in HashRouter (at App.js:179) in Unknown (at src/index.js:10) in StrictMode (at src/index.js:9) ```
1.0
Clean up react warnings - ##### SUMMARY Clean up react warnings. Go to console. See: ``` Warning: Can't perform a React state update on an unmounted component. This is a no-op, but it indicates a memory leak in your application. To fix, cancel all subscriptions and asynchronous tasks in a useEffect cleanup function. in InventorySourceAdd (at InventorySources.js:12) in Route (at InventorySources.js:11) index.js:1``` ``` ``` Warning: React does not recognize the `isValid` prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase `isvalid` instead. If you accidentally passed it from a parent component, remove it from the DOM element. in ul (created by Context.Consumer) in SelectMenu (created by ForwardRef) in ForwardRef (created by Select) in div (created by Select) in GenerateId (created by Select) in Select (at SCMSubForm.js:113) in div (created by FormGroup) in div (created by FormGroup) in FormGroup (at SCMSubForm.js:94) in SCMSubForm (at InventorySourceForm.js:204) in div (created by styled.div) in styled.div (at InventorySourceForm.js:155) in div (created by styled.div) in styled.div (at InventorySourceForm.js:151) in InventorySourceFormFields (at InventorySourceForm.js:298) in div (created by styled.div) in styled.div (at InventorySourceForm.js:297) ``` One more - Logged out and see ``` Warning: Cannot update a component (`SessionProvider`) while rendering a different component (`ProtectedRoute`). To locate the bad setState() call inside `ProtectedRoute`, follow the stack trace as described in https://fb.me/setstate-in-render in ProtectedRoute (at App.js:166) in Switch (at App.js:156) in SessionProvider (at App.js:155) in Unknown (at App.js:154) in I18nProvider (at App.js:153) in App (at App.js:180) in Router (created by HashRouter) in HashRouter (at App.js:179) in Unknown (at src/index.js:10) in StrictMode (at src/index.js:9) ```
priority
clean up react warnings summary clean up react warnings go to console see warning can t perform a react state update on an unmounted component this is a no op but it indicates a memory leak in your application to fix cancel all subscriptions and asynchronous tasks in a useeffect cleanup function in inventorysourceadd at inventorysources js in route at inventorysources js index js warning react does not recognize the isvalid prop on a dom element if you intentionally want it to appear in the dom as a custom attribute spell it as lowercase isvalid instead if you accidentally passed it from a parent component remove it from the dom element in ul created by context consumer in selectmenu created by forwardref in forwardref created by select in div created by select in generateid created by select in select at scmsubform js in div created by formgroup in div created by formgroup in formgroup at scmsubform js in scmsubform at inventorysourceform js in div created by styled div in styled div at inventorysourceform js in div created by styled div in styled div at inventorysourceform js in inventorysourceformfields at inventorysourceform js in div created by styled div in styled div at inventorysourceform js one more logged out and see warning cannot update a component sessionprovider while rendering a different component protectedroute to locate the bad setstate call inside protectedroute follow the stack trace as described in in protectedroute at app js in switch at app js in sessionprovider at app js in unknown at app js in at app js in app at app js in router created by hashrouter in hashrouter at app js in unknown at src index js in strictmode at src index js
1
69,839
3,315,861,921
IssuesEvent
2015-11-06 14:28:24
craftjk/issue-migration-test4
https://api.github.com/repos/craftjk/issue-migration-test4
opened
Visualize the Hierarchical Type IIs Assembly Tree Process
Priority: Medium Type: Feature
<a href="https://github.com/KeeganW"><img src="https://avatars.githubusercontent.com/u/7226822?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [KeeganW](https://github.com/KeeganW)** _Tuesday Jun 10, 2014 at 23:44 GMT_ _Originally opened as https://github.com/TeselaGen/ve/issues/80_ ---- Tim Thimmaiah commented May 09, 2013 10:52 (edited May 09, 2013 11:04) Along with accommodating hierarchical design we will build the corresponding UI tool for the visualization of the assembly tree. RavenCAD + Pidgeon are doing this fairly well right now. We'd want the same thing conceptually, but do it better and more polished.
1.0
Visualize the Hierarchical Type IIs Assembly Tree Process - <a href="https://github.com/KeeganW"><img src="https://avatars.githubusercontent.com/u/7226822?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [KeeganW](https://github.com/KeeganW)** _Tuesday Jun 10, 2014 at 23:44 GMT_ _Originally opened as https://github.com/TeselaGen/ve/issues/80_ ---- Tim Thimmaiah commented May 09, 2013 10:52 (edited May 09, 2013 11:04) Along with accommodating hierarchical design we will build the corresponding UI tool for the visualization of the assembly tree. RavenCAD + Pidgeon are doing this fairly well right now. We'd want the same thing conceptually, but do it better and more polished.
priority
visualize the hierarchical type iis assembly tree process issue by tuesday jun at gmt originally opened as tim thimmaiah commented may edited may along with accommodating hierarchical design we will build the corresponding ui tool for the visualization of the assembly tree ravencad pidgeon are doing this fairly well right now we d want the same thing conceptually but do it better and more polished
1
780,327
27,390,004,694
IssuesEvent
2023-02-28 15:41:57
ansible-collections/azure
https://api.github.com/repos/ansible-collections/azure
closed
azure_rm dynamic inventory error: [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: name 'azure_cloud' is not defined
inventory medium_priority work in
I've searched all over the internet and cannot find a solution to this problem. OS: Ubuntu 22.04.2 $ ansible --version ansible [core 2.14.2] config file = /home/nwmadmin/.ansible.cfg configured module search path = ['/home/nwmadmin/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3/dist-packages/ansible ansible collection location = /home/nwmadmin/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible python version = 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] (/usr/bin/python3) jinja version = 3.0.3 libyaml = True $ python3 --version Python 3.10.6 I'm trying to get the dynamic inventory for my Azure subscription working. Per the Azure documentation, I've created the file azure_rm.yaml with the following contents: plugin: azure.azcollection.azure_rm include_vm_resource_groups: - '*' auth_source: auto Using the command: $ ansible-inventory -i azure_rm.yaml --graph -vvv ansible-inventory [core 2.14.2] config file = /home/nwmadmin/.ansible.cfg configured module search path = ['/home/nwmadmin/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3/dist-packages/ansible ansible collection location = /home/nwmadmin/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible-inventory python version = 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] (/usr/bin/python3) jinja version = 3.0.3 libyaml = True Using /home/nwmadmin/.ansible.cfg as config file redirecting (type: inventory) ansible.builtin.azure_rm to azure.azcollection.azure_rm host_list declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method script declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method Using inventory plugin 'ansible_collections.azure.azcollection.plugins.inventory.azure_rm' to process inventory source '/home/nwmadmin/AWX/azure_rm.yaml' toml declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: name 'azure_cloud' is not defined File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source plugin.parse(self._inventory, self._loader, source, cache=cache) File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/auto.py", line 59, in parse plugin.parse(inventory, loader, path, cache=cache) File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 220, in parse self._credential_setup() File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 241, in _credential_setup self.azure_auth = AzureRMAuth(**auth_options) File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py", line 1496, in __init__ self._cloud_environment = azure_cloud.AZURE_PUBLIC_CLOUD # SDK default [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with yaml plugin: Plugin configuration YAML file, not YAML inventory File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source plugin.parse(self._inventory, self._loader, source, cache=cache) File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/yaml.py", line 114, in parse raise AnsibleParserError('Plugin configuration YAML file, not YAML inventory') [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with ini plugin: Invalid host pattern 'plugin:' supplied, ending in ':' is not allowed, this character is reserved to provide a port. File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source plugin.parse(self._inventory, self._loader, source, cache=cache) File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/ini.py", line 137, in parse raise AnsibleParserError(e) [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with ansible_collections.azure.azcollection.plugins.inventory.azure_rm plugin: name 'azure_cloud' is not defined File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source plugin.parse(self._inventory, self._loader, source, cache=cache) File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 220, in parse self._credential_setup() File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 241, in _credential_setup self.azure_auth = AzureRMAuth(**auth_options) File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py", line 1496, in __init__ self._cloud_environment = azure_cloud.AZURE_PUBLIC_CLOUD # SDK default [WARNING]: Unable to parse /home/nwmadmin/AWX/azure_rm.yaml as an inventory source [WARNING]: No inventory was parsed, only implicit localhost is available @all: |--@ungrouped: I have also tried using a ~/.azure/credentials file with the <value> replaced with my correct identifiers: [default] subscription_id=<value> client_id=<value> secret=<value> tenant=<value> I then change the azure_rm.yaml file to use auth_source: credential_file The output errors are identical to what is shown above. I then try first logging into the azure subscription with: $ az login -u myusername $ az account show $ az account show { "environmentName": "AzureCloud", "homeTenantId": "<value>", "id": "<value>", "isDefault": true, "managedByTenants": [], "name": "<value>", "state": "Enabled", "tenantId": "<value>", "user": { "name": "<myusername>", "type": "user" } } If I replace the auth_source in my azure_rm.yaml file with: auth_source: cli and then run the command: $ ansible-inventory -i azure_rm.yaml --graph -vvv The warning for the auto plugin is: [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: Failed to import the required Python library (azure-cli) on ansibleautomationvm's Python /usr/bin/python3. This is required for `cli` auth_source. Please read the module documentation and install it in the appropriate location. If the required library is installed, but Ansible is using the wrong Python interpreter, please consult the documentation on ansible_python_interpreter I've read the module documentation located here: https://docs.ansible.com/ansible/latest/collections/azure/azcollection/azure_rm_inventory.html I see no reference in that documentation on how to or where to install the (azure-cli) python library.
1.0
azure_rm dynamic inventory error: [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: name 'azure_cloud' is not defined - I've searched all over the internet and cannot find a solution to this problem. OS: Ubuntu 22.04.2 $ ansible --version ansible [core 2.14.2] config file = /home/nwmadmin/.ansible.cfg configured module search path = ['/home/nwmadmin/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3/dist-packages/ansible ansible collection location = /home/nwmadmin/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible python version = 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] (/usr/bin/python3) jinja version = 3.0.3 libyaml = True $ python3 --version Python 3.10.6 I'm trying to get the dynamic inventory for my Azure subscription working. Per the Azure documentation, I've created the file azure_rm.yaml with the following contents: plugin: azure.azcollection.azure_rm include_vm_resource_groups: - '*' auth_source: auto Using the command: $ ansible-inventory -i azure_rm.yaml --graph -vvv ansible-inventory [core 2.14.2] config file = /home/nwmadmin/.ansible.cfg configured module search path = ['/home/nwmadmin/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3/dist-packages/ansible ansible collection location = /home/nwmadmin/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible-inventory python version = 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] (/usr/bin/python3) jinja version = 3.0.3 libyaml = True Using /home/nwmadmin/.ansible.cfg as config file redirecting (type: inventory) ansible.builtin.azure_rm to azure.azcollection.azure_rm host_list declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method script declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method Using inventory plugin 'ansible_collections.azure.azcollection.plugins.inventory.azure_rm' to process inventory source '/home/nwmadmin/AWX/azure_rm.yaml' toml declined parsing /home/nwmadmin/AWX/azure_rm.yaml as it did not pass its verify_file() method [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: name 'azure_cloud' is not defined File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source plugin.parse(self._inventory, self._loader, source, cache=cache) File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/auto.py", line 59, in parse plugin.parse(inventory, loader, path, cache=cache) File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 220, in parse self._credential_setup() File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 241, in _credential_setup self.azure_auth = AzureRMAuth(**auth_options) File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py", line 1496, in __init__ self._cloud_environment = azure_cloud.AZURE_PUBLIC_CLOUD # SDK default [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with yaml plugin: Plugin configuration YAML file, not YAML inventory File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source plugin.parse(self._inventory, self._loader, source, cache=cache) File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/yaml.py", line 114, in parse raise AnsibleParserError('Plugin configuration YAML file, not YAML inventory') [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with ini plugin: Invalid host pattern 'plugin:' supplied, ending in ':' is not allowed, this character is reserved to provide a port. File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source plugin.parse(self._inventory, self._loader, source, cache=cache) File "/usr/lib/python3/dist-packages/ansible/plugins/inventory/ini.py", line 137, in parse raise AnsibleParserError(e) [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with ansible_collections.azure.azcollection.plugins.inventory.azure_rm plugin: name 'azure_cloud' is not defined File "/usr/lib/python3/dist-packages/ansible/inventory/manager.py", line 293, in parse_source plugin.parse(self._inventory, self._loader, source, cache=cache) File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 220, in parse self._credential_setup() File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/inventory/azure_rm.py", line 241, in _credential_setup self.azure_auth = AzureRMAuth(**auth_options) File "/usr/lib/python3/dist-packages/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py", line 1496, in __init__ self._cloud_environment = azure_cloud.AZURE_PUBLIC_CLOUD # SDK default [WARNING]: Unable to parse /home/nwmadmin/AWX/azure_rm.yaml as an inventory source [WARNING]: No inventory was parsed, only implicit localhost is available @all: |--@ungrouped: I have also tried using a ~/.azure/credentials file with the <value> replaced with my correct identifiers: [default] subscription_id=<value> client_id=<value> secret=<value> tenant=<value> I then change the azure_rm.yaml file to use auth_source: credential_file The output errors are identical to what is shown above. I then try first logging into the azure subscription with: $ az login -u myusername $ az account show $ az account show { "environmentName": "AzureCloud", "homeTenantId": "<value>", "id": "<value>", "isDefault": true, "managedByTenants": [], "name": "<value>", "state": "Enabled", "tenantId": "<value>", "user": { "name": "<myusername>", "type": "user" } } If I replace the auth_source in my azure_rm.yaml file with: auth_source: cli and then run the command: $ ansible-inventory -i azure_rm.yaml --graph -vvv The warning for the auto plugin is: [WARNING]: * Failed to parse /home/nwmadmin/AWX/azure_rm.yaml with auto plugin: Failed to import the required Python library (azure-cli) on ansibleautomationvm's Python /usr/bin/python3. This is required for `cli` auth_source. Please read the module documentation and install it in the appropriate location. If the required library is installed, but Ansible is using the wrong Python interpreter, please consult the documentation on ansible_python_interpreter I've read the module documentation located here: https://docs.ansible.com/ansible/latest/collections/azure/azcollection/azure_rm_inventory.html I see no reference in that documentation on how to or where to install the (azure-cli) python library.
priority
azure rm dynamic inventory error failed to parse home nwmadmin awx azure rm yaml with auto plugin name azure cloud is not defined i ve searched all over the internet and cannot find a solution to this problem os ubuntu ansible version ansible config file home nwmadmin ansible cfg configured module search path ansible python module location usr lib dist packages ansible ansible collection location home nwmadmin ansible collections usr share ansible collections executable location usr bin ansible python version main nov usr bin jinja version libyaml true version python i m trying to get the dynamic inventory for my azure subscription working per the azure documentation i ve created the file azure rm yaml with the following contents plugin azure azcollection azure rm include vm resource groups auth source auto using the command ansible inventory i azure rm yaml graph vvv ansible inventory config file home nwmadmin ansible cfg configured module search path ansible python module location usr lib dist packages ansible ansible collection location home nwmadmin ansible collections usr share ansible collections executable location usr bin ansible inventory python version main nov usr bin jinja version libyaml true using home nwmadmin ansible cfg as config file redirecting type inventory ansible builtin azure rm to azure azcollection azure rm host list declined parsing home nwmadmin awx azure rm yaml as it did not pass its verify file method script declined parsing home nwmadmin awx azure rm yaml as it did not pass its verify file method using inventory plugin ansible collections azure azcollection plugins inventory azure rm to process inventory source home nwmadmin awx azure rm yaml toml declined parsing home nwmadmin awx azure rm yaml as it did not pass its verify file method failed to parse home nwmadmin awx azure rm yaml with auto plugin name azure cloud is not defined file usr lib dist packages ansible inventory manager py line in parse source plugin parse self inventory self loader source cache cache file usr lib dist packages ansible plugins inventory auto py line in parse plugin parse inventory loader path cache cache file usr lib dist packages ansible collections azure azcollection plugins inventory azure rm py line in parse self credential setup file usr lib dist packages ansible collections azure azcollection plugins inventory azure rm py line in credential setup self azure auth azurermauth auth options file usr lib dist packages ansible collections azure azcollection plugins module utils azure rm common py line in init self cloud environment azure cloud azure public cloud sdk default failed to parse home nwmadmin awx azure rm yaml with yaml plugin plugin configuration yaml file not yaml inventory file usr lib dist packages ansible inventory manager py line in parse source plugin parse self inventory self loader source cache cache file usr lib dist packages ansible plugins inventory yaml py line in parse raise ansibleparsererror plugin configuration yaml file not yaml inventory failed to parse home nwmadmin awx azure rm yaml with ini plugin invalid host pattern plugin supplied ending in is not allowed this character is reserved to provide a port file usr lib dist packages ansible inventory manager py line in parse source plugin parse self inventory self loader source cache cache file usr lib dist packages ansible plugins inventory ini py line in parse raise ansibleparsererror e failed to parse home nwmadmin awx azure rm yaml with ansible collections azure azcollection plugins inventory azure rm plugin name azure cloud is not defined file usr lib dist packages ansible inventory manager py line in parse source plugin parse self inventory self loader source cache cache file usr lib dist packages ansible collections azure azcollection plugins inventory azure rm py line in parse self credential setup file usr lib dist packages ansible collections azure azcollection plugins inventory azure rm py line in credential setup self azure auth azurermauth auth options file usr lib dist packages ansible collections azure azcollection plugins module utils azure rm common py line in init self cloud environment azure cloud azure public cloud sdk default unable to parse home nwmadmin awx azure rm yaml as an inventory source no inventory was parsed only implicit localhost is available all ungrouped i have also tried using a azure credentials file with the replaced with my correct identifiers subscription id client id secret tenant i then change the azure rm yaml file to use auth source credential file the output errors are identical to what is shown above i then try first logging into the azure subscription with az login u myusername az account show az account show environmentname azurecloud hometenantid id isdefault true managedbytenants name state enabled tenantid user name type user if i replace the auth source in my azure rm yaml file with auth source cli and then run the command ansible inventory i azure rm yaml graph vvv the warning for the auto plugin is failed to parse home nwmadmin awx azure rm yaml with auto plugin failed to import the required python library azure cli on ansibleautomationvm s python usr bin this is required for cli auth source please read the module documentation and install it in the appropriate location if the required library is installed but ansible is using the wrong python interpreter please consult the documentation on ansible python interpreter i ve read the module documentation located here i see no reference in that documentation on how to or where to install the azure cli python library
1
771,171
27,073,055,748
IssuesEvent
2023-02-14 08:38:56
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Setting a Poetry constraint causes lockfile update to fail
type:bug priority-3-medium manager:poetry status:in-progress
### How are you running Renovate? Mend Renovate hosted app on github.com ### If you're self-hosting Renovate, tell us what version of Renovate you run. Present on self-hosted and github app ### If you're self-hosting Renovate, select which platform you are using. gitlab.com ### If you're self-hosting Renovate, tell us what version of the platform you run. Also observed on gitlab.com 15.8 ### Was this something which used to work for you, and then stopped? I never saw this working ### Describe the bug When setting a python and poetry constraint, the poetry lock update command fails because renovate attempts to pip install poetry using the wrong format for pip version constraints. See the repro here: https://github.com/WarpRat/renovate-poetry-contraint-repro/pull/6 ### Relevant debug logs <details><summary>Logs</summary> ``` DEBUG: Failed to update poetry.lock file(branch="renovate/pylint-2.x") { "err": { "name": "ExecError", "cmd": "/bin/sh -c docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"", "stderr": "ERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n", "stdout": "installing v2 tool python v3.9.9\nlinking tool python v3.9.9\nPython 3.9.9\npip 22.3.1 from /opt/buildpack/tools/python/3.9.9/lib/python3.9/site-packages/pip (python 3.9)\nInstalled v2 /usr/local/buildpack/tools/v2/python.sh in 24 seconds\nskip cleanup, not a docker build: 8d08c5ae1170\n", "options": { "cwd": "/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro", "encoding": "utf-8", "env": { "PIP_CACHE_DIR": "/tmp/renovate-cache/others/pip", "HOME": "/home/ubuntu", "PATH": "/home/ubuntu/.local/bin:/home/ubuntu/bin:/opt/buildpack/tools/python/3.9.3/bin:/home/ubuntu/.npm-global/bin:/home/ubuntu/renovateapp/node_modules/.bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "LC_ALL": "C.UTF-8", "LANG": "C.UTF-8", "BUILDPACK_CACHE_DIR": "/tmp/containerbase", "CONTAINERBASE_CACHE_DIR": "/tmp/containerbase" }, "maxBuffer": 10485760, "timeout": 900000 }, "exitCode": 1, "message": "Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"\nERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n", "stack": "ExecError: Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"\nERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n\n at ChildProcess.<anonymous> (/home/ubuntu/renovateapp/node_modules/renovate/dist/util/exec/common.js:87:24)\n at ChildProcess.emit (node:events:525:35)\n at ChildProcess.emit (node:domain:489:12)\n at Process.ChildProcess._handle.onexit (node:internal/child_process:293:12)" } } ``` </details> ### Have you created a minimal reproduction repository? I have linked to a minimal reproduction repository in the bug description
1.0
Setting a Poetry constraint causes lockfile update to fail - ### How are you running Renovate? Mend Renovate hosted app on github.com ### If you're self-hosting Renovate, tell us what version of Renovate you run. Present on self-hosted and github app ### If you're self-hosting Renovate, select which platform you are using. gitlab.com ### If you're self-hosting Renovate, tell us what version of the platform you run. Also observed on gitlab.com 15.8 ### Was this something which used to work for you, and then stopped? I never saw this working ### Describe the bug When setting a python and poetry constraint, the poetry lock update command fails because renovate attempts to pip install poetry using the wrong format for pip version constraints. See the repro here: https://github.com/WarpRat/renovate-poetry-contraint-repro/pull/6 ### Relevant debug logs <details><summary>Logs</summary> ``` DEBUG: Failed to update poetry.lock file(branch="renovate/pylint-2.x") { "err": { "name": "ExecError", "cmd": "/bin/sh -c docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"", "stderr": "ERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n", "stdout": "installing v2 tool python v3.9.9\nlinking tool python v3.9.9\nPython 3.9.9\npip 22.3.1 from /opt/buildpack/tools/python/3.9.9/lib/python3.9/site-packages/pip (python 3.9)\nInstalled v2 /usr/local/buildpack/tools/v2/python.sh in 24 seconds\nskip cleanup, not a docker build: 8d08c5ae1170\n", "options": { "cwd": "/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro", "encoding": "utf-8", "env": { "PIP_CACHE_DIR": "/tmp/renovate-cache/others/pip", "HOME": "/home/ubuntu", "PATH": "/home/ubuntu/.local/bin:/home/ubuntu/bin:/opt/buildpack/tools/python/3.9.3/bin:/home/ubuntu/.npm-global/bin:/home/ubuntu/renovateapp/node_modules/.bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "LC_ALL": "C.UTF-8", "LANG": "C.UTF-8", "BUILDPACK_CACHE_DIR": "/tmp/containerbase", "CONTAINERBASE_CACHE_DIR": "/tmp/containerbase" }, "maxBuffer": 10485760, "timeout": 900000 }, "exitCode": 1, "message": "Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"\nERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n", "stack": "ExecError: Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\":\"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/WarpRat/renovate-poetry-contraint-repro\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.9.9 && pip install --user poetry1.3.2 && poetry update --lock --no-interaction pylint\"\nERROR: Could not find a version that satisfies the requirement poetry1.3.2 (from versions: none)\nERROR: No matching distribution found for poetry1.3.2\n\n at ChildProcess.<anonymous> (/home/ubuntu/renovateapp/node_modules/renovate/dist/util/exec/common.js:87:24)\n at ChildProcess.emit (node:events:525:35)\n at ChildProcess.emit (node:domain:489:12)\n at Process.ChildProcess._handle.onexit (node:internal/child_process:293:12)" } } ``` </details> ### Have you created a minimal reproduction repository? I have linked to a minimal reproduction repository in the bug description
priority
setting a poetry constraint causes lockfile update to fail how are you running renovate mend renovate hosted app on github com if you re self hosting renovate tell us what version of renovate you run present on self hosted and github app if you re self hosting renovate select which platform you are using gitlab com if you re self hosting renovate tell us what version of the platform you run also observed on gitlab com was this something which used to work for you and then stopped i never saw this working describe the bug when setting a python and poetry constraint the poetry lock update command fails because renovate attempts to pip install poetry using the wrong format for pip version constraints see the repro here relevant debug logs logs debug failed to update poetry lock file branch renovate pylint x err name execerror cmd bin sh c docker run rm name renovate sidecar label renovate child v mnt renovate gh warprat renovate poetry contraint repro mnt renovate gh warprat renovate poetry contraint repro v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir e containerbase cache dir w mnt renovate gh warprat renovate poetry contraint repro docker io renovate sidecar bash l c install tool python pip install user poetry update lock no interaction pylint stderr error could not find a version that satisfies the requirement from versions none nerror no matching distribution found for n stdout installing tool python nlinking tool python npython npip from opt buildpack tools python lib site packages pip python ninstalled usr local buildpack tools python sh in seconds nskip cleanup not a docker build n options cwd mnt renovate gh warprat renovate poetry contraint repro encoding utf env pip cache dir tmp renovate cache others pip home home ubuntu path home ubuntu local bin home ubuntu bin opt buildpack tools python bin home ubuntu npm global bin home ubuntu renovateapp node modules bin usr local sbin usr local bin usr sbin usr bin sbin bin lc all c utf lang c utf buildpack cache dir tmp containerbase containerbase cache dir tmp containerbase maxbuffer timeout exitcode message command failed docker run rm name renovate sidecar label renovate child v mnt renovate gh warprat renovate poetry contraint repro mnt renovate gh warprat renovate poetry contraint repro v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir e containerbase cache dir w mnt renovate gh warprat renovate poetry contraint repro docker io renovate sidecar bash l c install tool python pip install user poetry update lock no interaction pylint nerror could not find a version that satisfies the requirement from versions none nerror no matching distribution found for n stack execerror command failed docker run rm name renovate sidecar label renovate child v mnt renovate gh warprat renovate poetry contraint repro mnt renovate gh warprat renovate poetry contraint repro v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir e containerbase cache dir w mnt renovate gh warprat renovate poetry contraint repro docker io renovate sidecar bash l c install tool python pip install user poetry update lock no interaction pylint nerror could not find a version that satisfies the requirement from versions none nerror no matching distribution found for n n at childprocess home ubuntu renovateapp node modules renovate dist util exec common js n at childprocess emit node events n at childprocess emit node domain n at process childprocess handle onexit node internal child process have you created a minimal reproduction repository i have linked to a minimal reproduction repository in the bug description
1
88,251
3,775,408,747
IssuesEvent
2016-03-17 13:28:19
marklogic/marklogic-data-hub
https://api.github.com/repos/marklogic/marklogic-data-hub
closed
Input Flow doesn't report back failures
Component:QuickStart priority:medium
When I run an input flow that fails it doesn't report the errors or the failure back to the quick-start UI
1.0
Input Flow doesn't report back failures - When I run an input flow that fails it doesn't report the errors or the failure back to the quick-start UI
priority
input flow doesn t report back failures when i run an input flow that fails it doesn t report the errors or the failure back to the quick start ui
1
255,607
8,125,822,992
IssuesEvent
2018-08-16 22:29:58
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Add tests for 2D and 3D Line annotations
Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal
Subject says it all. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2337 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: Add tests for 2D and 3D Line annotations Assigned to: Kathleen Biagas Category: Target version: 2.10 Author: Kathleen Biagas Start: 07/14/2015 Due date: % Done: 100 Estimated time: Created: 07/14/2015 12:53 am Updated: 08/27/2015 08:22 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: Subject says it all. Comments: SVN Revision 26740A tests/rendering/line2d.pyA baseline/rendering/line2d/line2d00.pngA baseline/rendering/line2d/line2d10.pngA baseline/rendering/line2d/line2d01.pngA baseline/rendering/line2d/line2d02.pngA baseline/rendering/line2d/line2d03.pngA baseline/rendering/line2d/line2d04.pngA baseline/rendering/line2d/line2d05.pngA baseline/rendering/line2d/line2d06.pngA baseline/rendering/line2d/line2d07.pngA baseline/rendering/line2d/line2d08.pngA baseline/rendering/line2d/line2d09.pngSVN Revision 27056A baseline/rendering/line3d/line3d00.pngA baseline/rendering/line3d/Line3d01.pngA baseline/rendering/line3d/line3d10.pngA baseline/rendering/line3d/Line3d02.pngA baseline/rendering/line3d/line3d11.pngA baseline/rendering/line3d/Line3d03.pngA baseline/rendering/line3d/line3d12.pngA baseline/rendering/line3d/Line3d04.pngA baseline/rendering/line3d/line3d13.pngA baseline/rendering/line3d/line3d05.pngA baseline/rendering/line3d/line3d06.pngA baseline/rendering/line3d/line3d07.pngA baseline/rendering/line3d/line3d08.pngA baseline/rendering/line3d/line3d09.png
1.0
Add tests for 2D and 3D Line annotations - Subject says it all. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2337 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: Add tests for 2D and 3D Line annotations Assigned to: Kathleen Biagas Category: Target version: 2.10 Author: Kathleen Biagas Start: 07/14/2015 Due date: % Done: 100 Estimated time: Created: 07/14/2015 12:53 am Updated: 08/27/2015 08:22 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: Subject says it all. Comments: SVN Revision 26740A tests/rendering/line2d.pyA baseline/rendering/line2d/line2d00.pngA baseline/rendering/line2d/line2d10.pngA baseline/rendering/line2d/line2d01.pngA baseline/rendering/line2d/line2d02.pngA baseline/rendering/line2d/line2d03.pngA baseline/rendering/line2d/line2d04.pngA baseline/rendering/line2d/line2d05.pngA baseline/rendering/line2d/line2d06.pngA baseline/rendering/line2d/line2d07.pngA baseline/rendering/line2d/line2d08.pngA baseline/rendering/line2d/line2d09.pngSVN Revision 27056A baseline/rendering/line3d/line3d00.pngA baseline/rendering/line3d/Line3d01.pngA baseline/rendering/line3d/line3d10.pngA baseline/rendering/line3d/Line3d02.pngA baseline/rendering/line3d/line3d11.pngA baseline/rendering/line3d/Line3d03.pngA baseline/rendering/line3d/line3d12.pngA baseline/rendering/line3d/Line3d04.pngA baseline/rendering/line3d/line3d13.pngA baseline/rendering/line3d/line3d05.pngA baseline/rendering/line3d/line3d06.pngA baseline/rendering/line3d/line3d07.pngA baseline/rendering/line3d/line3d08.pngA baseline/rendering/line3d/line3d09.png
priority
add tests for and line annotations subject says it all redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject add tests for and line annotations assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created am updated pm likelihood severity found in version impact medium expected use occasional os all support group any description subject says it all comments svn revision tests rendering pya baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pngsvn revision baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering png
1
417,799
12,179,396,154
IssuesEvent
2020-04-28 10:36:13
Arquisoft/viade_en1a
https://api.github.com/repos/Arquisoft/viade_en1a
closed
You can not delete shared routes
bug medium priority
The button of deleting routes someone shared with you does not work. You should be able to dismiss shared routes you do not want.
1.0
You can not delete shared routes - The button of deleting routes someone shared with you does not work. You should be able to dismiss shared routes you do not want.
priority
you can not delete shared routes the button of deleting routes someone shared with you does not work you should be able to dismiss shared routes you do not want
1
512,587
14,901,413,177
IssuesEvent
2021-01-21 16:26:21
staxrip/staxrip
https://api.github.com/repos/staxrip/staxrip
closed
Please add NvEnc --vpp-colorspace options.
added/fixed/done feature request priority medium
HDR looks terrible on my TVs with washed out blacks and faded color. Both TVs support HDR and show it as enabled when HDR content is present. However, the colors and blacks still look terrible compared to normal SDR which looks great. This HDR issue is well confirmed on the internet and Tone mapping HDR to SDR has become the current best solution. I use Staxrip with the Nvidia H265 encoder as this is much faster than x265 and retains excellent quality. Nvidia encoder supports fast hardware tone mapping using "--vpp-colorspace hdr2sdr=reinhard" command line option. Currently we have to add this option manually in Encoder Options/Other/Custom. Works well with excellent results. Can we have a "Colorspace Conversion" option added to the VPP GUI section of the NvEnc Encoder Options? Should include Reinhard, Hable and Mobius conversion options as a selection box. NvEnc supports a heap of manual options for Colorspace Conversions as well. These could be added too, perhaps at a later stage as I think most people will be happy with the 3 Auto selections listed above. Will be great to have the Colorspace options available in their proper easy place in the GUI so they can be used by the many and not have to be manually added by the very determined few.
1.0
Please add NvEnc --vpp-colorspace options. - HDR looks terrible on my TVs with washed out blacks and faded color. Both TVs support HDR and show it as enabled when HDR content is present. However, the colors and blacks still look terrible compared to normal SDR which looks great. This HDR issue is well confirmed on the internet and Tone mapping HDR to SDR has become the current best solution. I use Staxrip with the Nvidia H265 encoder as this is much faster than x265 and retains excellent quality. Nvidia encoder supports fast hardware tone mapping using "--vpp-colorspace hdr2sdr=reinhard" command line option. Currently we have to add this option manually in Encoder Options/Other/Custom. Works well with excellent results. Can we have a "Colorspace Conversion" option added to the VPP GUI section of the NvEnc Encoder Options? Should include Reinhard, Hable and Mobius conversion options as a selection box. NvEnc supports a heap of manual options for Colorspace Conversions as well. These could be added too, perhaps at a later stage as I think most people will be happy with the 3 Auto selections listed above. Will be great to have the Colorspace options available in their proper easy place in the GUI so they can be used by the many and not have to be manually added by the very determined few.
priority
please add nvenc vpp colorspace options hdr looks terrible on my tvs with washed out blacks and faded color both tvs support hdr and show it as enabled when hdr content is present however the colors and blacks still look terrible compared to normal sdr which looks great this hdr issue is well confirmed on the internet and tone mapping hdr to sdr has become the current best solution i use staxrip with the nvidia encoder as this is much faster than and retains excellent quality nvidia encoder supports fast hardware tone mapping using vpp colorspace reinhard command line option currently we have to add this option manually in encoder options other custom works well with excellent results can we have a colorspace conversion option added to the vpp gui section of the nvenc encoder options should include reinhard hable and mobius conversion options as a selection box nvenc supports a heap of manual options for colorspace conversions as well these could be added too perhaps at a later stage as i think most people will be happy with the auto selections listed above will be great to have the colorspace options available in their proper easy place in the gui so they can be used by the many and not have to be manually added by the very determined few
1
394,043
11,628,622,961
IssuesEvent
2020-02-27 18:42:21
wynn-rj/sword-and-bored-game
https://api.github.com/repos/wynn-rj/sword-and-bored-game
opened
Functioning Unit and Squad System
enhancement medium-priority strategy view
- Add "Available Unit" pool into base - Upon creating a unit you can place it in any City under your control. Creation is instant, sending it to another city takes 1 turn. If you leave it in the base you can use it on the same turn it is created. Add Squads that you can put units together in: - Between 1-6 units per squad - Squads have single turn on World Map, act together
1.0
Functioning Unit and Squad System - - Add "Available Unit" pool into base - Upon creating a unit you can place it in any City under your control. Creation is instant, sending it to another city takes 1 turn. If you leave it in the base you can use it on the same turn it is created. Add Squads that you can put units together in: - Between 1-6 units per squad - Squads have single turn on World Map, act together
priority
functioning unit and squad system add available unit pool into base upon creating a unit you can place it in any city under your control creation is instant sending it to another city takes turn if you leave it in the base you can use it on the same turn it is created add squads that you can put units together in between units per squad squads have single turn on world map act together
1
769,678
27,016,296,170
IssuesEvent
2023-02-10 19:43:57
CDCgov/prime-reportstream
https://api.github.com/repos/CDCgov/prime-reportstream
closed
CA - LA County Test Date Error
onboarding-ops Medium Priority California
LA County California is reporting issues with test dates being wrong. ##Background Information At LA county, we have received some reports with date issue. See summaries below and examples attached. Please let us know if you’re able to help resolve these issues. ![image.png](https://images.zenhubusercontent.com/62828ac9597e3bae41382df7/75da72a0-80eb-4c6f-9b26-049fbe9216f4)
1.0
CA - LA County Test Date Error - LA County California is reporting issues with test dates being wrong. ##Background Information At LA county, we have received some reports with date issue. See summaries below and examples attached. Please let us know if you’re able to help resolve these issues. ![image.png](https://images.zenhubusercontent.com/62828ac9597e3bae41382df7/75da72a0-80eb-4c6f-9b26-049fbe9216f4)
priority
ca la county test date error la county california is reporting issues with test dates being wrong background information at la county we have received some reports with date issue see summaries below and examples attached please let us know if you’re able to help resolve these issues
1
454,460
13,101,476,922
IssuesEvent
2020-08-04 03:52:52
ankidroid/Anki-Android
https://api.github.com/repos/ankidroid/Anki-Android
closed
Indicator for unsynced local changes
Accepted Anki Ecosystem Compatibility Enhancement Keep Open Priority-Medium Sync
Originally reported on Google Code with ID 2638 ``` I would like an indicator that shows that there are unsynced local changes (learning progress). I learn on many devices (staniory (PC) and mobile) in parallel and commonly sync to many times because i am unsure about the state of the devices. “Did i learn anything on the tablet today? Don’t know, better sync once more to be safe.” It could be something unobtrusive. Maybe an * added to the “You learned X cards in Y” line at the bottom of the deck picker. Is this feature available on similar apps (…)? Not really in learning apps afaik. The syncing is one of the great pluses of AnkiDroid compared to other apps. But Emacs uses a * to indicate unsaved (rather than unsynced) changes, which i find quite useful. ``` Reported by `ospalh` on 2015-06-10 09:46:18
1.0
Indicator for unsynced local changes - Originally reported on Google Code with ID 2638 ``` I would like an indicator that shows that there are unsynced local changes (learning progress). I learn on many devices (staniory (PC) and mobile) in parallel and commonly sync to many times because i am unsure about the state of the devices. “Did i learn anything on the tablet today? Don’t know, better sync once more to be safe.” It could be something unobtrusive. Maybe an * added to the “You learned X cards in Y” line at the bottom of the deck picker. Is this feature available on similar apps (…)? Not really in learning apps afaik. The syncing is one of the great pluses of AnkiDroid compared to other apps. But Emacs uses a * to indicate unsaved (rather than unsynced) changes, which i find quite useful. ``` Reported by `ospalh` on 2015-06-10 09:46:18
priority
indicator for unsynced local changes originally reported on google code with id i would like an indicator that shows that there are unsynced local changes learning progress i learn on many devices staniory pc and mobile in parallel and commonly sync to many times because i am unsure about the state of the devices “did i learn anything on the tablet today don’t know better sync once more to be safe ” it could be something unobtrusive maybe an added to the “you learned x cards in y” line at the bottom of the deck picker is this feature available on similar apps … not really in learning apps afaik the syncing is one of the great pluses of ankidroid compared to other apps but emacs uses a to indicate unsaved rather than unsynced changes which i find quite useful reported by ospalh on
1
778,471
27,318,042,161
IssuesEvent
2023-02-24 17:14:32
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] Enable single-row optimizations for valid RETURNING expressions
kind/enhancement area/ysql priority/medium
Jira Link: [DB-2310](https://yugabyte.atlassian.net/browse/DB-2310) In the fix for #1028 we are adding support for single-row optimization on UPDATE/DELETEs, however currently it only allows for simple RETURNING clauses that just specify columns. However this is too conservative, as there are valid single-row returning expressions (i.e. those that only use column values included in the query itself) that are more complex than just a standalone column. Example of query that is currently (once diff lands) optimized: `DELETE FROM t WHERE a = 1 RETURNING a;` Example of query that is valid but is not optimized: `DELETE FROM t WHERE a = 1 RETURNING a + 1` This issue tracks adding optimization support for the latter case and all general expressions that only involve column values that are included in the query.
1.0
[YSQL] Enable single-row optimizations for valid RETURNING expressions - Jira Link: [DB-2310](https://yugabyte.atlassian.net/browse/DB-2310) In the fix for #1028 we are adding support for single-row optimization on UPDATE/DELETEs, however currently it only allows for simple RETURNING clauses that just specify columns. However this is too conservative, as there are valid single-row returning expressions (i.e. those that only use column values included in the query itself) that are more complex than just a standalone column. Example of query that is currently (once diff lands) optimized: `DELETE FROM t WHERE a = 1 RETURNING a;` Example of query that is valid but is not optimized: `DELETE FROM t WHERE a = 1 RETURNING a + 1` This issue tracks adding optimization support for the latter case and all general expressions that only involve column values that are included in the query.
priority
enable single row optimizations for valid returning expressions jira link in the fix for we are adding support for single row optimization on update deletes however currently it only allows for simple returning clauses that just specify columns however this is too conservative as there are valid single row returning expressions i e those that only use column values included in the query itself that are more complex than just a standalone column example of query that is currently once diff lands optimized delete from t where a returning a example of query that is valid but is not optimized delete from t where a returning a this issue tracks adding optimization support for the latter case and all general expressions that only involve column values that are included in the query
1
829,504
31,881,440,433
IssuesEvent
2023-09-16 12:28:24
containrrr/watchtower
https://api.github.com/repos/containrrr/watchtower
closed
Notification Delay not working
Type: Bug Priority: Medium Status: Available
### Describe the bug WATCHTOWER_NOTIFICATION_DELAY / WATCHTOWER_NOTIFICATIONS_DELAY environment variables seems to have no effect. Notifications are sent directly Please note https://containrrr.dev/watchtower/notifications/ says it's WATCHTOWER_NOTIFICATION_DELAY but source code seems to indicate WATCHTOWER_NOTIFICATIONS_DELAY (notice the "S") https://github.com/containrrr/watchtower/blob/a56b9bdb7c7c8de96fc4e5f5fe5b5f4a952447be/internal/flags/flags.go#L208 To be sure I set both in my container ### Steps to reproduce 1. Set WATCHTOWER_NOTIFICATION_DELAY=120 and WATCHTOWER_NOTIFICATION_URL 2. Trigger a notification (by starting up an old image or setting WATCHTOWER_NO_STARTUP_MESSAGE to false and restart container) ### Expected behavior Notification should be triggered later ### Screenshots _No response_ ### Environment - Debian Bookworm - x86_64 - Docker Version : 24.0.5 "Env": [ "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "TZ=Europe/Paris", "WATCHTOWER_LABEL_ENABLE=true", "WATCHTOWER_MONITOR_ONLY=False", "WATCHTOWER_NOTIFICATION_URL=generic://xxxxxxxxxxxx/webhook/xxxxxxxxxx?template=json", "WATCHTOWER_NOTIFICATIONS=shoutrrr", "WATCHTOWER_SCHEDULE=0 0 21 * * *", "WATCHTOWER_NO_STARTUP_MESSAGE=False", "WATCHTOWER_NOTIFICATIONS_DELAY=120", "WATCHTOWER_NOTIFICATION_DELAY=120", "WATCHTOWER_DEBUG=True" ], ### Your logs ```text 2023-09-13T18:46:34.665775703Z time="2023-09-13T20:46:34+02:00" level=debug msg="Sleeping for a second to ensure the docker api client has been properly initialized." 2023-09-13T18:46:35.665954195Z time="2023-09-13T20:46:35+02:00" level=debug msg="Making sure everything is sane before starting" 2023-09-13T18:46:35.665993015Z time="2023-09-13T20:46:35+02:00" level=debug msg="Retrieving running containers" 2023-09-13T18:46:35.752821755Z time="2023-09-13T20:46:35+02:00" level=debug msg="There are no additional watchtower containers" 2023-09-13T18:46:35.752845425Z time="2023-09-13T20:46:35+02:00" level=debug msg="Watchtower HTTP API skipped." 2023-09-13T18:46:35.752851085Z time="2023-09-13T20:46:35+02:00" level=info msg="Watchtower 1.5.3" 2023-09-13T18:46:35.752855385Z time="2023-09-13T20:46:35+02:00" level=info msg="Using notifications: generic" 2023-09-13T18:46:35.752859425Z time="2023-09-13T20:46:35+02:00" level=info msg="Only checking containers using enable label" 2023-09-13T18:46:35.752867245Z time="2023-09-13T20:46:35+02:00" level=info msg="Scheduling first run: 2023-09-13 21:00:00 +0200 CEST" 2023-09-13T18:46:35.752871475Z time="2023-09-13T20:46:35+02:00" level=info msg="Note that the first check will be performed in 13 minutes, 24 seconds" 2023-09-13T19:00:00.000458126Z time="2023-09-13T21:00:00+02:00" level=debug msg="Checking containers for updated images" 2023-09-13T19:00:00.000508366Z time="2023-09-13T21:00:00+02:00" level=debug msg="Retrieving running containers" 2023-09-13T19:00:01.164754872Z time="2023-09-13T21:00:01+02:00" level=debug msg="Trying to load authentication credentials." container=/homeassistant image="docker.io/homeassistant/home-assistant:stable" 2023-09-13T19:00:01.164842423Z time="2023-09-13T21:00:01+02:00" level=debug msg="No credentials for docker.io found" config_file=/config.json 2023-09-13T19:00:01.164865053Z time="2023-09-13T21:00:01+02:00" level=debug msg="Got image name: docker.io/homeassistant/home-assistant:stable" 2023-09-13T19:00:01.164881323Z time="2023-09-13T21:00:01+02:00" level=debug msg="Checking if pull is needed" container=/homeassistant image="docker.io/homeassistant/home-assistant:stable" 2023-09-13T19:00:01.164909713Z time="2023-09-13T21:00:01+02:00" level=debug msg="Building challenge URL" URL="https://index.docker.io/v2/" 2023-09-13T19:00:01.488038988Z time="2023-09-13T21:00:01+02:00" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://auth.docker.io/token\",service=\"registry.docker.io\"" status="401 Unauthorized" 2023-09-13T19:00:01.488077238Z time="2023-09-13T21:00:01+02:00" level=debug msg="Checking challenge header content" realm="https://auth.docker.io/token" service=registry.docker.io 2023-09-13T19:00:01.488091068Z time="2023-09-13T21:00:01+02:00" level=debug msg="Setting scope for auth token" image=docker.io/homeassistant/home-assistant scope="repository:homeassistant/home-assistant:pull" 2023-09-13T19:00:01.488106329Z time="2023-09-13T21:00:01+02:00" level=debug msg="No credentials found." 2023-09-13T19:00:01.839030275Z time="2023-09-13T21:00:01+02:00" level=debug msg="Parsing image ref" host=index.docker.io image=docker.io/homeassistant/home-assistant normalized="docker.io/homeassistant/home-assistant:stable" tag=stable 2023-09-13T19:00:01.839075615Z time="2023-09-13T21:00:01+02:00" level=debug msg="Doing a HEAD request to fetch a digest" url="https://index.docker.io/v2/homeassistant/home-assistant/manifests/stable" 2023-09-13T19:00:02.228619850Z time="2023-09-13T21:00:02+02:00" level=debug msg="Found a remote digest to compare with" remote="sha256:223c9986ff318eaf35fac70351900cf766c15faf82fbe54150b73b7cfc3c933a" 2023-09-13T19:00:02.228666641Z time="2023-09-13T21:00:02+02:00" level=debug msg=Comparing local="sha256:183513a13fbcb758eb525f857cf3ec69a619a5ec27c117a2077ab2f4e7a3c052" remote="sha256:223c9986ff318eaf35fac70351900cf766c15faf82fbe54150b73b7cfc3c933a" 2023-09-13T19:00:02.228699001Z time="2023-09-13T21:00:02+02:00" level=debug msg="Digests did not match, doing a pull." 2023-09-13T19:00:02.228707471Z time="2023-09-13T21:00:02+02:00" level=debug msg="Pulling image" container=/homeassistant image="docker.io/homeassistant/home-assistant:stable" 2023-09-13T19:00:54.987751992Z time="2023-09-13T21:00:54+02:00" level=info msg="Found new docker.io/homeassistant/home-assistant:stable image (1bcf95d7b112)" 2023-09-13T19:01:14.350469904Z time="2023-09-13T21:01:14+02:00" level=debug msg="Removing container 10736233407b" 2023-09-13T19:01:14.393279697Z time="2023-09-13T21:01:14+02:00" level=info msg="Stopping /homeassistant (6da3dfd8d3a6) with SIGTERM" 2023-09-13T19:01:21.908834811Z time="2023-09-13T21:01:21+02:00" level=info msg="Session done" Failed=0 Scanned=11 Updated=2 notify=no 2023-09-13T19:01:21.910918050Z time="2023-09-13T21:01:21+02:00" level=debug msg="Scheduled next run: 2023-09-14 21:00:00 +0200 CEST" 2023-09-13T19:01:21.962319729Z time="2023-09-13T21:01:21+02:00" level=error msg="Failed to send shoutrrr notification" error="an error occurred while sending notification to generic webhook: server returned response status code 502 Bad Gateway" index=0 notify=no service=generic ``` ### Additional context _No response_
1.0
Notification Delay not working - ### Describe the bug WATCHTOWER_NOTIFICATION_DELAY / WATCHTOWER_NOTIFICATIONS_DELAY environment variables seems to have no effect. Notifications are sent directly Please note https://containrrr.dev/watchtower/notifications/ says it's WATCHTOWER_NOTIFICATION_DELAY but source code seems to indicate WATCHTOWER_NOTIFICATIONS_DELAY (notice the "S") https://github.com/containrrr/watchtower/blob/a56b9bdb7c7c8de96fc4e5f5fe5b5f4a952447be/internal/flags/flags.go#L208 To be sure I set both in my container ### Steps to reproduce 1. Set WATCHTOWER_NOTIFICATION_DELAY=120 and WATCHTOWER_NOTIFICATION_URL 2. Trigger a notification (by starting up an old image or setting WATCHTOWER_NO_STARTUP_MESSAGE to false and restart container) ### Expected behavior Notification should be triggered later ### Screenshots _No response_ ### Environment - Debian Bookworm - x86_64 - Docker Version : 24.0.5 "Env": [ "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "TZ=Europe/Paris", "WATCHTOWER_LABEL_ENABLE=true", "WATCHTOWER_MONITOR_ONLY=False", "WATCHTOWER_NOTIFICATION_URL=generic://xxxxxxxxxxxx/webhook/xxxxxxxxxx?template=json", "WATCHTOWER_NOTIFICATIONS=shoutrrr", "WATCHTOWER_SCHEDULE=0 0 21 * * *", "WATCHTOWER_NO_STARTUP_MESSAGE=False", "WATCHTOWER_NOTIFICATIONS_DELAY=120", "WATCHTOWER_NOTIFICATION_DELAY=120", "WATCHTOWER_DEBUG=True" ], ### Your logs ```text 2023-09-13T18:46:34.665775703Z time="2023-09-13T20:46:34+02:00" level=debug msg="Sleeping for a second to ensure the docker api client has been properly initialized." 2023-09-13T18:46:35.665954195Z time="2023-09-13T20:46:35+02:00" level=debug msg="Making sure everything is sane before starting" 2023-09-13T18:46:35.665993015Z time="2023-09-13T20:46:35+02:00" level=debug msg="Retrieving running containers" 2023-09-13T18:46:35.752821755Z time="2023-09-13T20:46:35+02:00" level=debug msg="There are no additional watchtower containers" 2023-09-13T18:46:35.752845425Z time="2023-09-13T20:46:35+02:00" level=debug msg="Watchtower HTTP API skipped." 2023-09-13T18:46:35.752851085Z time="2023-09-13T20:46:35+02:00" level=info msg="Watchtower 1.5.3" 2023-09-13T18:46:35.752855385Z time="2023-09-13T20:46:35+02:00" level=info msg="Using notifications: generic" 2023-09-13T18:46:35.752859425Z time="2023-09-13T20:46:35+02:00" level=info msg="Only checking containers using enable label" 2023-09-13T18:46:35.752867245Z time="2023-09-13T20:46:35+02:00" level=info msg="Scheduling first run: 2023-09-13 21:00:00 +0200 CEST" 2023-09-13T18:46:35.752871475Z time="2023-09-13T20:46:35+02:00" level=info msg="Note that the first check will be performed in 13 minutes, 24 seconds" 2023-09-13T19:00:00.000458126Z time="2023-09-13T21:00:00+02:00" level=debug msg="Checking containers for updated images" 2023-09-13T19:00:00.000508366Z time="2023-09-13T21:00:00+02:00" level=debug msg="Retrieving running containers" 2023-09-13T19:00:01.164754872Z time="2023-09-13T21:00:01+02:00" level=debug msg="Trying to load authentication credentials." container=/homeassistant image="docker.io/homeassistant/home-assistant:stable" 2023-09-13T19:00:01.164842423Z time="2023-09-13T21:00:01+02:00" level=debug msg="No credentials for docker.io found" config_file=/config.json 2023-09-13T19:00:01.164865053Z time="2023-09-13T21:00:01+02:00" level=debug msg="Got image name: docker.io/homeassistant/home-assistant:stable" 2023-09-13T19:00:01.164881323Z time="2023-09-13T21:00:01+02:00" level=debug msg="Checking if pull is needed" container=/homeassistant image="docker.io/homeassistant/home-assistant:stable" 2023-09-13T19:00:01.164909713Z time="2023-09-13T21:00:01+02:00" level=debug msg="Building challenge URL" URL="https://index.docker.io/v2/" 2023-09-13T19:00:01.488038988Z time="2023-09-13T21:00:01+02:00" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://auth.docker.io/token\",service=\"registry.docker.io\"" status="401 Unauthorized" 2023-09-13T19:00:01.488077238Z time="2023-09-13T21:00:01+02:00" level=debug msg="Checking challenge header content" realm="https://auth.docker.io/token" service=registry.docker.io 2023-09-13T19:00:01.488091068Z time="2023-09-13T21:00:01+02:00" level=debug msg="Setting scope for auth token" image=docker.io/homeassistant/home-assistant scope="repository:homeassistant/home-assistant:pull" 2023-09-13T19:00:01.488106329Z time="2023-09-13T21:00:01+02:00" level=debug msg="No credentials found." 2023-09-13T19:00:01.839030275Z time="2023-09-13T21:00:01+02:00" level=debug msg="Parsing image ref" host=index.docker.io image=docker.io/homeassistant/home-assistant normalized="docker.io/homeassistant/home-assistant:stable" tag=stable 2023-09-13T19:00:01.839075615Z time="2023-09-13T21:00:01+02:00" level=debug msg="Doing a HEAD request to fetch a digest" url="https://index.docker.io/v2/homeassistant/home-assistant/manifests/stable" 2023-09-13T19:00:02.228619850Z time="2023-09-13T21:00:02+02:00" level=debug msg="Found a remote digest to compare with" remote="sha256:223c9986ff318eaf35fac70351900cf766c15faf82fbe54150b73b7cfc3c933a" 2023-09-13T19:00:02.228666641Z time="2023-09-13T21:00:02+02:00" level=debug msg=Comparing local="sha256:183513a13fbcb758eb525f857cf3ec69a619a5ec27c117a2077ab2f4e7a3c052" remote="sha256:223c9986ff318eaf35fac70351900cf766c15faf82fbe54150b73b7cfc3c933a" 2023-09-13T19:00:02.228699001Z time="2023-09-13T21:00:02+02:00" level=debug msg="Digests did not match, doing a pull." 2023-09-13T19:00:02.228707471Z time="2023-09-13T21:00:02+02:00" level=debug msg="Pulling image" container=/homeassistant image="docker.io/homeassistant/home-assistant:stable" 2023-09-13T19:00:54.987751992Z time="2023-09-13T21:00:54+02:00" level=info msg="Found new docker.io/homeassistant/home-assistant:stable image (1bcf95d7b112)" 2023-09-13T19:01:14.350469904Z time="2023-09-13T21:01:14+02:00" level=debug msg="Removing container 10736233407b" 2023-09-13T19:01:14.393279697Z time="2023-09-13T21:01:14+02:00" level=info msg="Stopping /homeassistant (6da3dfd8d3a6) with SIGTERM" 2023-09-13T19:01:21.908834811Z time="2023-09-13T21:01:21+02:00" level=info msg="Session done" Failed=0 Scanned=11 Updated=2 notify=no 2023-09-13T19:01:21.910918050Z time="2023-09-13T21:01:21+02:00" level=debug msg="Scheduled next run: 2023-09-14 21:00:00 +0200 CEST" 2023-09-13T19:01:21.962319729Z time="2023-09-13T21:01:21+02:00" level=error msg="Failed to send shoutrrr notification" error="an error occurred while sending notification to generic webhook: server returned response status code 502 Bad Gateway" index=0 notify=no service=generic ``` ### Additional context _No response_
priority
notification delay not working describe the bug watchtower notification delay watchtower notifications delay environment variables seems to have no effect notifications are sent directly please note says it s watchtower notification delay but source code seems to indicate watchtower notifications delay notice the s to be sure i set both in my container steps to reproduce set watchtower notification delay and watchtower notification url trigger a notification by starting up an old image or setting watchtower no startup message to false and restart container expected behavior notification should be triggered later screenshots no response environment debian bookworm docker version env path usr local sbin usr local bin usr sbin usr bin sbin bin tz europe paris watchtower label enable true watchtower monitor only false watchtower notification url generic xxxxxxxxxxxx webhook xxxxxxxxxx template json watchtower notifications shoutrrr watchtower schedule watchtower no startup message false watchtower notifications delay watchtower notification delay watchtower debug true your logs text time level debug msg sleeping for a second to ensure the docker api client has been properly initialized time level debug msg making sure everything is sane before starting time level debug msg retrieving running containers time level debug msg there are no additional watchtower containers time level debug msg watchtower http api skipped time level info msg watchtower time level info msg using notifications generic time level info msg only checking containers using enable label time level info msg scheduling first run cest time level info msg note that the first check will be performed in minutes seconds time level debug msg checking containers for updated images time level debug msg retrieving running containers time level debug msg trying to load authentication credentials container homeassistant image docker io homeassistant home assistant stable time level debug msg no credentials for docker io found config file config json time level debug msg got image name docker io homeassistant home assistant stable time level debug msg checking if pull is needed container homeassistant image docker io homeassistant home assistant stable time level debug msg building challenge url url time level debug msg got response to challenge request header bearer realm status unauthorized time level debug msg checking challenge header content realm service registry docker io time level debug msg setting scope for auth token image docker io homeassistant home assistant scope repository homeassistant home assistant pull time level debug msg no credentials found time level debug msg parsing image ref host index docker io image docker io homeassistant home assistant normalized docker io homeassistant home assistant stable tag stable time level debug msg doing a head request to fetch a digest url time level debug msg found a remote digest to compare with remote time level debug msg comparing local remote time level debug msg digests did not match doing a pull time level debug msg pulling image container homeassistant image docker io homeassistant home assistant stable time level info msg found new docker io homeassistant home assistant stable image time level debug msg removing container time level info msg stopping homeassistant with sigterm time level info msg session done failed scanned updated notify no time level debug msg scheduled next run cest time level error msg failed to send shoutrrr notification error an error occurred while sending notification to generic webhook server returned response status code bad gateway index notify no service generic additional context no response
1
508,519
14,701,734,535
IssuesEvent
2021-01-04 12:24:47
seung-lab/neuroglancer
https://api.github.com/repos/seung-lab/neuroglancer
closed
meshes do not show up in Chrome
Priority: Medium Type: Bug
it works in Firefox. My chrome: Version 83.0.4103.116 (Official Build) (64-bit) OS: Ubuntu 18.04
1.0
meshes do not show up in Chrome - it works in Firefox. My chrome: Version 83.0.4103.116 (Official Build) (64-bit) OS: Ubuntu 18.04
priority
meshes do not show up in chrome it works in firefox my chrome version official build bit os ubuntu
1
41,972
2,869,091,092
IssuesEvent
2015-06-05 23:15:45
dart-lang/polymer-dart
https://api.github.com/repos/dart-lang/polymer-dart
opened
Cannot use iterables in bindings
bug PolymerMilestone-Next Priority-Medium
<a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)** _Originally opened as dart-lang/sdk#19945_ ---- They could generate infinite loops. Right now we are OK with them because we are not discarding changes in 'get value' in polymer-expressions, but you'd hit the issue if we fix that. We also hit this if you use path_observers. Here is a change that illustrates the problem: https://codereview.chromium.org/382683005 ObservableMap.keys would always be different, so we would think there was a change every time we do a read. The CL patches ObservableMap.keys to cache it's previous result to illustrate that the issue goes away if we know somehow that the value didn't change. I think we need to guarantee that with any Bindable, calling `get value` twice in a row should return either the same thing, or equal values. Haven't we solved this issue before already?
1.0
Cannot use iterables in bindings - <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)** _Originally opened as dart-lang/sdk#19945_ ---- They could generate infinite loops. Right now we are OK with them because we are not discarding changes in 'get value' in polymer-expressions, but you'd hit the issue if we fix that. We also hit this if you use path_observers. Here is a change that illustrates the problem: https://codereview.chromium.org/382683005 ObservableMap.keys would always be different, so we would think there was a change every time we do a read. The CL patches ObservableMap.keys to cache it's previous result to illustrate that the issue goes away if we know somehow that the value didn't change. I think we need to guarantee that with any Bindable, calling `get value` twice in a row should return either the same thing, or equal values. Haven't we solved this issue before already?
priority
cannot use iterables in bindings issue by originally opened as dart lang sdk they could generate infinite loops right now we are ok with them because we are not discarding changes in get value in polymer expressions but you d hit the issue if we fix that we also hit this if you use path observers here is a change that illustrates the problem observablemap keys would always be different so we would think there was a change every time we do a read the cl patches observablemap keys to cache it s previous result to illustrate that the issue goes away if we know somehow that the value didn t change i think we need to guarantee that with any bindable calling get value twice in a row should return either the same thing or equal values haven t we solved this issue before already
1
372,486
11,015,550,320
IssuesEvent
2019-12-05 02:01:53
NIFCLOUD-mbaas/ncmb_ios
https://api.github.com/repos/NIFCLOUD-mbaas/ncmb_ios
closed
UIWebview - Deprecated
Priority: Medium
つい最近からだと思いますが。アプリをApp Storeにアップロードすると、UIWebViewを使用しているアプリだと以下のようにワーニング出るようになってしまいました。 NCMBRichPushViewクラスにて使用しているようなので、UIWebViewからWKWebViewへの変更をお願いします。 We identified one or more issues with a recent delivery for your app, "xxxx". Your delivery was successful, but you may wish to correct the following issues in your next delivery: ITMS-90809: Deprecated API Usage - Apple will stop accepting submissions of apps that use UIWebView APIs .
1.0
UIWebview - Deprecated - つい最近からだと思いますが。アプリをApp Storeにアップロードすると、UIWebViewを使用しているアプリだと以下のようにワーニング出るようになってしまいました。 NCMBRichPushViewクラスにて使用しているようなので、UIWebViewからWKWebViewへの変更をお願いします。 We identified one or more issues with a recent delivery for your app, "xxxx". Your delivery was successful, but you may wish to correct the following issues in your next delivery: ITMS-90809: Deprecated API Usage - Apple will stop accepting submissions of apps that use UIWebView APIs .
priority
uiwebview deprecated つい最近からだと思いますが。アプリをapp storeにアップロードすると、uiwebviewを使用しているアプリだと以下のようにワーニング出るようになってしまいました。 ncmbrichpushviewクラスにて使用しているようなので、uiwebviewからwkwebviewへの変更をお願いします。 we identified one or more issues with a recent delivery for your app xxxx your delivery was successful but you may wish to correct the following issues in your next delivery itms deprecated api usage apple will stop accepting submissions of apps that use uiwebview apis
1
682,441
23,345,322,792
IssuesEvent
2022-08-09 17:23:09
Javacord/Javacord
https://api.github.com/repos/Javacord/Javacord
closed
MessageUpdateHandler is outdated
medium priority
The MessageUpdateHandler is outdated and for example does not update the message if components have been changed and also does not fire an event in this case
1.0
MessageUpdateHandler is outdated - The MessageUpdateHandler is outdated and for example does not update the message if components have been changed and also does not fire an event in this case
priority
messageupdatehandler is outdated the messageupdatehandler is outdated and for example does not update the message if components have been changed and also does not fire an event in this case
1
111,104
4,461,558,711
IssuesEvent
2016-08-24 06:20:27
bespokeinteractive/ipdapp
https://api.github.com/repos/bespokeinteractive/ipdapp
opened
IPD module: Search functionality within the IPD module not working as it should
bug Medium priority
PD module: Search functionality within the IPD module not working as it should Suggestion -Work on clearing on searched item to reveal all items in the list
1.0
IPD module: Search functionality within the IPD module not working as it should - PD module: Search functionality within the IPD module not working as it should Suggestion -Work on clearing on searched item to reveal all items in the list
priority
ipd module search functionality within the ipd module not working as it should pd module search functionality within the ipd module not working as it should suggestion work on clearing on searched item to reveal all items in the list
1
667,648
22,495,497,778
IssuesEvent
2022-06-23 07:11:03
OffprintStudios/dragonfish
https://api.github.com/repos/OffprintStudios/dragonfish
opened
Refreshing approval queue resulted in authentication issue where said work approver doesn't have permission to view
Bug Medium Priority
Reported by work approver: I just refreshed the queue page, and went from seeing the story that had been claimed by Techie, to getting the 'you do not have permission to view this page' message. Clicking on 'home' and then back to the queue via the dash worked as expected, though, now showing 'The queue is clear!' This likely ties into other general issues with authentication
1.0
Refreshing approval queue resulted in authentication issue where said work approver doesn't have permission to view - Reported by work approver: I just refreshed the queue page, and went from seeing the story that had been claimed by Techie, to getting the 'you do not have permission to view this page' message. Clicking on 'home' and then back to the queue via the dash worked as expected, though, now showing 'The queue is clear!' This likely ties into other general issues with authentication
priority
refreshing approval queue resulted in authentication issue where said work approver doesn t have permission to view reported by work approver i just refreshed the queue page and went from seeing the story that had been claimed by techie to getting the you do not have permission to view this page message clicking on home and then back to the queue via the dash worked as expected though now showing the queue is clear this likely ties into other general issues with authentication
1
13,785
2,610,300,572
IssuesEvent
2015-02-26 19:36:36
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Icons for hedgewars demo *.hwd and hedgewars save *.hws files
auto-migrated Priority-Medium Type-Enhancement
``` Hi. This isnt a real problem. its just cosmetic. On KDE Plasma, there are really talented icon designers. (oxygen icons) so maybe you could ask them if they could make this two icons für you or ill ask them if this is ok. Because this [?] symbol isnt so good :( ``` ----- Original issue reported on code.google.com by `apfelmau...@googlemail.com` on 7 Apr 2011 at 9:51
1.0
Icons for hedgewars demo *.hwd and hedgewars save *.hws files - ``` Hi. This isnt a real problem. its just cosmetic. On KDE Plasma, there are really talented icon designers. (oxygen icons) so maybe you could ask them if they could make this two icons für you or ill ask them if this is ok. Because this [?] symbol isnt so good :( ``` ----- Original issue reported on code.google.com by `apfelmau...@googlemail.com` on 7 Apr 2011 at 9:51
priority
icons for hedgewars demo hwd and hedgewars save hws files hi this isnt a real problem its just cosmetic on kde plasma there are really talented icon designers oxygen icons so maybe you could ask them if they could make this two icons für you or ill ask them if this is ok because this symbol isnt so good original issue reported on code google com by apfelmau googlemail com on apr at
1
798,399
28,262,986,501
IssuesEvent
2023-04-07 02:15:26
AY2223S2-CS2103-W17-3/tp
https://api.github.com/repos/AY2223S2-CS2103-W17-3/tp
closed
[PE-D][Tester D] Find functionality inconsistent with UserGuide (partial vs full word search)
bug docs priority.Medium
## Bug Report - ezSchedule App ### Issue Description There is an issue with find, where finding strings that are partial in a word will return the event. For example, `find e` will return any events which names' contain the letter e. This contradicts the userguide, which says that find finds based on **keyword**. ![Screenshot 2023-03-31 at 8.12.42 PM.png](https://raw.githubusercontent.com/wengkit1/CS2103-T-PE-Dry-run/main/files/de97d23a-7dda-44ad-b6d7-10e4245666ee.png) ### Expected Behavior I expected the app to find events based on word alone(non-case sensitive). As some words may be part of another word for example, event `eating fish` and event `fishing`, I may only want to find events that contain fish, and not care about events like fishing. ### Actual Behavior Currently, the app's find returns any string as long as they contain any letter passed as a parameter to find. ### Steps to Reproduce 1. Open the ezSchedule app. 2. Add a tennis event 3. Use the find command `find e` ### Environment - OS: Mac OS 13.3 Ventura - App Version: 1.3.1 ### Priority We consider this to be a very low priority bug, as it can cause confusion for users and make it slightly unintuitive to use the app, but not a huge problem. <!--session: 1680242278740-fb3aa958-ae37-4e91-b3e1-1f77cd8ac339--><!--Version: Web v3.4.7--> ------------- Labels: `type.FunctionalityBug` `severity.VeryLow` original: wengkit1/CS2103-T-PE-Dry-run#12
1.0
[PE-D][Tester D] Find functionality inconsistent with UserGuide (partial vs full word search) - ## Bug Report - ezSchedule App ### Issue Description There is an issue with find, where finding strings that are partial in a word will return the event. For example, `find e` will return any events which names' contain the letter e. This contradicts the userguide, which says that find finds based on **keyword**. ![Screenshot 2023-03-31 at 8.12.42 PM.png](https://raw.githubusercontent.com/wengkit1/CS2103-T-PE-Dry-run/main/files/de97d23a-7dda-44ad-b6d7-10e4245666ee.png) ### Expected Behavior I expected the app to find events based on word alone(non-case sensitive). As some words may be part of another word for example, event `eating fish` and event `fishing`, I may only want to find events that contain fish, and not care about events like fishing. ### Actual Behavior Currently, the app's find returns any string as long as they contain any letter passed as a parameter to find. ### Steps to Reproduce 1. Open the ezSchedule app. 2. Add a tennis event 3. Use the find command `find e` ### Environment - OS: Mac OS 13.3 Ventura - App Version: 1.3.1 ### Priority We consider this to be a very low priority bug, as it can cause confusion for users and make it slightly unintuitive to use the app, but not a huge problem. <!--session: 1680242278740-fb3aa958-ae37-4e91-b3e1-1f77cd8ac339--><!--Version: Web v3.4.7--> ------------- Labels: `type.FunctionalityBug` `severity.VeryLow` original: wengkit1/CS2103-T-PE-Dry-run#12
priority
find functionality inconsistent with userguide partial vs full word search bug report ezschedule app issue description there is an issue with find where finding strings that are partial in a word will return the event for example find e will return any events which names contain the letter e this contradicts the userguide which says that find finds based on keyword expected behavior i expected the app to find events based on word alone non case sensitive as some words may be part of another word for example event eating fish and event fishing i may only want to find events that contain fish and not care about events like fishing actual behavior currently the app s find returns any string as long as they contain any letter passed as a parameter to find steps to reproduce open the ezschedule app add a tennis event use the find command find e environment os mac os ventura app version priority we consider this to be a very low priority bug as it can cause confusion for users and make it slightly unintuitive to use the app but not a huge problem labels type functionalitybug severity verylow original t pe dry run
1
47,211
2,974,600,406
IssuesEvent
2015-07-15 02:15:09
Reimashi/jotai
https://api.github.com/repos/Reimashi/jotai
closed
No motherboard readings for MSI P67a-g43
auto-migrated Priority-Medium Type-Enhancement
``` What is the expected output? What do you see instead? details for the motherboard sesnsors What version of the product are you using? On what operating system? 0.3.2, windows 7 x64 Please provide any additional information below. MSI p67a-g43 motherboard Please attach a Report created with "File / Save Report...". ``` Original issue reported on code.google.com by `spudsd...@gmail.com` on 23 May 2011 at 2:41 Attachments: * [OpenHardwareMonitor.Report.txt](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-226/comment-0/OpenHardwareMonitor.Report.txt)
1.0
No motherboard readings for MSI P67a-g43 - ``` What is the expected output? What do you see instead? details for the motherboard sesnsors What version of the product are you using? On what operating system? 0.3.2, windows 7 x64 Please provide any additional information below. MSI p67a-g43 motherboard Please attach a Report created with "File / Save Report...". ``` Original issue reported on code.google.com by `spudsd...@gmail.com` on 23 May 2011 at 2:41 Attachments: * [OpenHardwareMonitor.Report.txt](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-226/comment-0/OpenHardwareMonitor.Report.txt)
priority
no motherboard readings for msi what is the expected output what do you see instead details for the motherboard sesnsors what version of the product are you using on what operating system windows please provide any additional information below msi motherboard please attach a report created with file save report original issue reported on code google com by spudsd gmail com on may at attachments
1
54,961
3,071,723,437
IssuesEvent
2015-08-19 13:44:35
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
opened
"Описание" в юникоде отабражается знаками вопроса в списке юзеров.
bug Component-Logic imported Priority-Medium Usability
_From [tret2...@gmail.com](https://code.google.com/u/116508191076211387118/) on November 17, 2010 12:32:47_ "Описание" в юникоде так и отабражается знаками вопроса в списке юзеров. В другом клиенте, на том же хабе все в норме. _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=222_
1.0
"Описание" в юникоде отабражается знаками вопроса в списке юзеров. - _From [tret2...@gmail.com](https://code.google.com/u/116508191076211387118/) on November 17, 2010 12:32:47_ "Описание" в юникоде так и отабражается знаками вопроса в списке юзеров. В другом клиенте, на том же хабе все в норме. _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=222_
priority
описание в юникоде отабражается знаками вопроса в списке юзеров from on november описание в юникоде так и отабражается знаками вопроса в списке юзеров в другом клиенте на том же хабе все в норме original issue
1
576,799
17,095,123,504
IssuesEvent
2021-07-09 00:28:38
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
opened
Need tangent matrix with respect to velocities for deformable dofs
priority: medium team: dynamics type: idea
### Background: In the current FEM formulation for deformable bodies, we are solving for the equation Ma − f_d(x,v) − fₑ(x) = fₑₓₜ, (1) which then gets discretized as F(a) = Ma − f_d(x₀,v(a, v₀)) − fₑ(x(a, v₀, x₀)) - fₑₓₜ = 0, (2) with the Newmark scheme. We then use Newton's method to solve the discretized equations with `a` as the unknown variables. In that process, we produce the tangent matrix ∇ₐF = dF/da, which is related to, but not the same as, the tangent matrix ∇ᵥF = dF/dv consumed by the contact solver. It would be nice to be able to reuse the tangent matrix from FEM solve when Newton's method converges instead of recalculating another tangent matrix for contact. ### Proposed solution: Change variable in the discretization of (1) and solve F(v) = Ma(v, a₀) − f_d(x₀, v) − fₑ(x(a₀, v, x₀)) - fₑₓₜ = 0, (3) instead with `v` as the unknown variables. The tangent matrix needed for solving (3) with Newton's method is ∇ᵥF = dF/dv, so we can reuse the tangent matrix in the contact solve when the Newton's iterations converge. To enable this alternative formulation, we need a new class similar to `NewmarkScheme` that captures the same time discretization scheme, but uses `v` instead of `a` as the "unknown variable". @amcastro-tri hinted at this idea in a PR review a while back.
1.0
Need tangent matrix with respect to velocities for deformable dofs - ### Background: In the current FEM formulation for deformable bodies, we are solving for the equation Ma − f_d(x,v) − fₑ(x) = fₑₓₜ, (1) which then gets discretized as F(a) = Ma − f_d(x₀,v(a, v₀)) − fₑ(x(a, v₀, x₀)) - fₑₓₜ = 0, (2) with the Newmark scheme. We then use Newton's method to solve the discretized equations with `a` as the unknown variables. In that process, we produce the tangent matrix ∇ₐF = dF/da, which is related to, but not the same as, the tangent matrix ∇ᵥF = dF/dv consumed by the contact solver. It would be nice to be able to reuse the tangent matrix from FEM solve when Newton's method converges instead of recalculating another tangent matrix for contact. ### Proposed solution: Change variable in the discretization of (1) and solve F(v) = Ma(v, a₀) − f_d(x₀, v) − fₑ(x(a₀, v, x₀)) - fₑₓₜ = 0, (3) instead with `v` as the unknown variables. The tangent matrix needed for solving (3) with Newton's method is ∇ᵥF = dF/dv, so we can reuse the tangent matrix in the contact solve when the Newton's iterations converge. To enable this alternative formulation, we need a new class similar to `NewmarkScheme` that captures the same time discretization scheme, but uses `v` instead of `a` as the "unknown variable". @amcastro-tri hinted at this idea in a PR review a while back.
priority
need tangent matrix with respect to velocities for deformable dofs background in the current fem formulation for deformable bodies we are solving for the equation ma − f d x v − fₑ x fₑₓₜ which then gets discretized as f a ma − f d x₀ v a v₀ − fₑ x a v₀ x₀ fₑₓₜ with the newmark scheme we then use newton s method to solve the discretized equations with a as the unknown variables in that process we produce the tangent matrix ∇ₐf df da which is related to but not the same as the tangent matrix ∇ᵥf df dv consumed by the contact solver it would be nice to be able to reuse the tangent matrix from fem solve when newton s method converges instead of recalculating another tangent matrix for contact proposed solution change variable in the discretization of and solve f v ma v a₀ − f d x₀ v − fₑ x a₀ v x₀ fₑₓₜ instead with v as the unknown variables the tangent matrix needed for solving with newton s method is ∇ᵥf df dv so we can reuse the tangent matrix in the contact solve when the newton s iterations converge to enable this alternative formulation we need a new class similar to newmarkscheme that captures the same time discretization scheme but uses v instead of a as the unknown variable amcastro tri hinted at this idea in a pr review a while back
1
498,450
14,407,243,543
IssuesEvent
2020-12-03 21:31:15
minio/docs
https://api.github.com/repos/minio/docs
closed
Initial Skeleton on MinIO + Private/Public Cloud
WIP priority: medium
# Summary Create the initial skeleton structure for MinIO + Kubernetes for Private/Public Cloud # Scope This issue covers creating skeleton structure for the following pages: - Getting Started (Quickstart w/ MinIO 4x4 + console, Self SIgned TLS, no KES) - Core Concepts - Architecture - Erasure Coding - Bitrot Protection - Deploy a MinIO Tenant - Manage a MinIO Tenant - Connect your Application - Security - User Management - Server Side Encryption - Transport Layer Security (TLS) - Production Recommendations
1.0
Initial Skeleton on MinIO + Private/Public Cloud - # Summary Create the initial skeleton structure for MinIO + Kubernetes for Private/Public Cloud # Scope This issue covers creating skeleton structure for the following pages: - Getting Started (Quickstart w/ MinIO 4x4 + console, Self SIgned TLS, no KES) - Core Concepts - Architecture - Erasure Coding - Bitrot Protection - Deploy a MinIO Tenant - Manage a MinIO Tenant - Connect your Application - Security - User Management - Server Side Encryption - Transport Layer Security (TLS) - Production Recommendations
priority
initial skeleton on minio private public cloud summary create the initial skeleton structure for minio kubernetes for private public cloud scope this issue covers creating skeleton structure for the following pages getting started quickstart w minio console self signed tls no kes core concepts architecture erasure coding bitrot protection deploy a minio tenant manage a minio tenant connect your application security user management server side encryption transport layer security tls production recommendations
1
78,028
3,508,788,391
IssuesEvent
2016-01-08 19:31:06
vigetlabs/trackomatic
https://api.github.com/repos/vigetlabs/trackomatic
closed
Set up docs and releases
medium priority
We should talk about how GitHub releases work, and how they interact with the Changelog. We should also probably tweak CONTRIBUTING.md so it reflects the project more specifically.
1.0
Set up docs and releases - We should talk about how GitHub releases work, and how they interact with the Changelog. We should also probably tweak CONTRIBUTING.md so it reflects the project more specifically.
priority
set up docs and releases we should talk about how github releases work and how they interact with the changelog we should also probably tweak contributing md so it reflects the project more specifically
1
30,054
2,722,146,282
IssuesEvent
2015-04-14 00:23:46
CruxFramework/crux-smart-faces
https://api.github.com/repos/CruxFramework/crux-smart-faces
closed
Centralize the Faces button
enhancement imported Milestone-M14-C4 Module-CruxSmartFaces Priority-Medium TargetVersion-5.3.0
_From [flavia.jesus@triggolabs.com](https://code.google.com/u/flavia.jesus@triggolabs.com/) on March 17, 2015 10:13:06_ The Faces button is decentralized. _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=637_
1.0
Centralize the Faces button - _From [flavia.jesus@triggolabs.com](https://code.google.com/u/flavia.jesus@triggolabs.com/) on March 17, 2015 10:13:06_ The Faces button is decentralized. _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=637_
priority
centralize the faces button from on march the faces button is decentralized original issue
1
799,426
28,306,669,892
IssuesEvent
2023-04-10 11:44:51
PrefectHQ/prefect
https://api.github.com/repos/PrefectHQ/prefect
closed
Subflow names not shown in timeline view
bug ui great writeup priority:medium status:in-progress from:sales team:observability
### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I refreshed the page and this issue still occurred. - [X] I checked if this issue was specific to the browser I was using by testing with a different browser. ### Bug summary Subflow names are not shown in the timeline view, only the randomly generated flow run name is shown. The radar chart was able to show sub flow names. ![image](https://user-images.githubusercontent.com/68969861/228359540-0ded0964-4c08-43e9-8405-4dafaabf09e8.png) ### Reproduction [Flow Code](https://github.com/taylor-curran/flow-patterns/blob/main/flows/subflows/task_wrapped_deployments.py) ``` from prefect import flow, task from prefect.deployments import run_deployment from prefect.task_runners import ConcurrentTaskRunner from pydantic import BaseModel @task() def upstream_task_h(): print("upstream task") return {"h": "upstream task"} @task() def upstream_task_i(): print("upstream task") return {"i": "upstream task"} @task() def wrapper_task_a(i, sim_failure_child_flow_a): print("wrapper task") a = run_deployment( "child-flow-a/dep-child-a", parameters={"i": i, "sim_failure_child_flow_a": sim_failure_child_flow_a}, ) return {"a": a.state.result()} @task() def wrapper_task_b(sim_failure_child_flow_b): print("wrapper task") b = run_deployment( name="child-flow-b/dep-child-b", parameters={"sim_failure_child_flow_b": sim_failure_child_flow_b}, ) # WARNING: We do not evaluate the result or state in this # wrapper task decoupling this wrapper task from its # subflow's state. return {"b": "not flow result"} @task() def wrapper_task_c(): print("wrapper task") c = run_deployment("child-flow-c/dep-child-c") return {"c": c.state.result()} @task() def downstream_task_j(a): print("downstream task") return {"j": "downstream task"} @task() async def downstream_task_j(a, c, sim_failure_downstream_task_j): if sim_failure_downstream_task_j: raise Exception("This is a test exception") else: print("downstream task") return {"j": "downstream task"} @task() def downstream_task_k(b="b"): print(b) print("downstream task") return {"k": "downstream task"} # --- class SimulatedFailure(BaseModel): child_flow_a: bool = False child_flow_b: bool = False downstream_task_j: bool = False default_simulated_failure = SimulatedFailure( child_flow_a=False, child_flow_b=False, downstream_task_j=False ) # prefect deployment build task_wrapped_deployments.py:task_wrapped_deployments -n dep_task_wrapped -t sub-flows -t task-wrapped -t parent -a @flow(task_runner=ConcurrentTaskRunner(), persist_result=True) def task_wrapped_deployments(sim_failure: SimulatedFailure = default_simulated_failure): h = upstream_task_h.submit() i = upstream_task_i.submit() a = wrapper_task_a.submit(i, sim_failure.child_flow_a) b = wrapper_task_b.submit( sim_failure_child_flow_b=sim_failure.child_flow_b, wait_for=[i] ) c = wrapper_task_c.submit() j = downstream_task_j.submit(a, c, sim_failure.downstream_task_j) k = downstream_task_k.submit(wait_for=[b]) return {"j": j, "k": k} # --- if __name__ == "__main__": task_wrapped_deployments( sim_failure=SimulatedFailure( child_flow_a=False, child_flow_b=True, downstream_task_j=False ) ) ``` ### Error ![image](https://user-images.githubusercontent.com/68969861/228359557-71dfe09d-0fa4-4df7-89e2-40cfb2d98f09.png) ### Browers - [X] Chrome - [ ] Firefox - [ ] Safari - [ ] Edge ### Prefect version ```Text Version: 2.8.6 API version: 0.8.4 Python version: 3.10.10 Git commit: 061d877b Built: Thu, Mar 16, 2023 2:58 PM OS/Arch: darwin/arm64 Profile: default Server type: cloud ``` ### Additional context _No response_
1.0
Subflow names not shown in timeline view - ### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I refreshed the page and this issue still occurred. - [X] I checked if this issue was specific to the browser I was using by testing with a different browser. ### Bug summary Subflow names are not shown in the timeline view, only the randomly generated flow run name is shown. The radar chart was able to show sub flow names. ![image](https://user-images.githubusercontent.com/68969861/228359540-0ded0964-4c08-43e9-8405-4dafaabf09e8.png) ### Reproduction [Flow Code](https://github.com/taylor-curran/flow-patterns/blob/main/flows/subflows/task_wrapped_deployments.py) ``` from prefect import flow, task from prefect.deployments import run_deployment from prefect.task_runners import ConcurrentTaskRunner from pydantic import BaseModel @task() def upstream_task_h(): print("upstream task") return {"h": "upstream task"} @task() def upstream_task_i(): print("upstream task") return {"i": "upstream task"} @task() def wrapper_task_a(i, sim_failure_child_flow_a): print("wrapper task") a = run_deployment( "child-flow-a/dep-child-a", parameters={"i": i, "sim_failure_child_flow_a": sim_failure_child_flow_a}, ) return {"a": a.state.result()} @task() def wrapper_task_b(sim_failure_child_flow_b): print("wrapper task") b = run_deployment( name="child-flow-b/dep-child-b", parameters={"sim_failure_child_flow_b": sim_failure_child_flow_b}, ) # WARNING: We do not evaluate the result or state in this # wrapper task decoupling this wrapper task from its # subflow's state. return {"b": "not flow result"} @task() def wrapper_task_c(): print("wrapper task") c = run_deployment("child-flow-c/dep-child-c") return {"c": c.state.result()} @task() def downstream_task_j(a): print("downstream task") return {"j": "downstream task"} @task() async def downstream_task_j(a, c, sim_failure_downstream_task_j): if sim_failure_downstream_task_j: raise Exception("This is a test exception") else: print("downstream task") return {"j": "downstream task"} @task() def downstream_task_k(b="b"): print(b) print("downstream task") return {"k": "downstream task"} # --- class SimulatedFailure(BaseModel): child_flow_a: bool = False child_flow_b: bool = False downstream_task_j: bool = False default_simulated_failure = SimulatedFailure( child_flow_a=False, child_flow_b=False, downstream_task_j=False ) # prefect deployment build task_wrapped_deployments.py:task_wrapped_deployments -n dep_task_wrapped -t sub-flows -t task-wrapped -t parent -a @flow(task_runner=ConcurrentTaskRunner(), persist_result=True) def task_wrapped_deployments(sim_failure: SimulatedFailure = default_simulated_failure): h = upstream_task_h.submit() i = upstream_task_i.submit() a = wrapper_task_a.submit(i, sim_failure.child_flow_a) b = wrapper_task_b.submit( sim_failure_child_flow_b=sim_failure.child_flow_b, wait_for=[i] ) c = wrapper_task_c.submit() j = downstream_task_j.submit(a, c, sim_failure.downstream_task_j) k = downstream_task_k.submit(wait_for=[b]) return {"j": j, "k": k} # --- if __name__ == "__main__": task_wrapped_deployments( sim_failure=SimulatedFailure( child_flow_a=False, child_flow_b=True, downstream_task_j=False ) ) ``` ### Error ![image](https://user-images.githubusercontent.com/68969861/228359557-71dfe09d-0fa4-4df7-89e2-40cfb2d98f09.png) ### Browers - [X] Chrome - [ ] Firefox - [ ] Safari - [ ] Edge ### Prefect version ```Text Version: 2.8.6 API version: 0.8.4 Python version: 3.10.10 Git commit: 061d877b Built: Thu, Mar 16, 2023 2:58 PM OS/Arch: darwin/arm64 Profile: default Server type: cloud ``` ### Additional context _No response_
priority
subflow names not shown in timeline view first check i added a descriptive title to this issue i used the github search to find a similar issue and didn t find it i refreshed the page and this issue still occurred i checked if this issue was specific to the browser i was using by testing with a different browser bug summary subflow names are not shown in the timeline view only the randomly generated flow run name is shown the radar chart was able to show sub flow names reproduction from prefect import flow task from prefect deployments import run deployment from prefect task runners import concurrenttaskrunner from pydantic import basemodel task def upstream task h print upstream task return h upstream task task def upstream task i print upstream task return i upstream task task def wrapper task a i sim failure child flow a print wrapper task a run deployment child flow a dep child a parameters i i sim failure child flow a sim failure child flow a return a a state result task def wrapper task b sim failure child flow b print wrapper task b run deployment name child flow b dep child b parameters sim failure child flow b sim failure child flow b warning we do not evaluate the result or state in this wrapper task decoupling this wrapper task from its subflow s state return b not flow result task def wrapper task c print wrapper task c run deployment child flow c dep child c return c c state result task def downstream task j a print downstream task return j downstream task task async def downstream task j a c sim failure downstream task j if sim failure downstream task j raise exception this is a test exception else print downstream task return j downstream task task def downstream task k b b print b print downstream task return k downstream task class simulatedfailure basemodel child flow a bool false child flow b bool false downstream task j bool false default simulated failure simulatedfailure child flow a false child flow b false downstream task j false prefect deployment build task wrapped deployments py task wrapped deployments n dep task wrapped t sub flows t task wrapped t parent a flow task runner concurrenttaskrunner persist result true def task wrapped deployments sim failure simulatedfailure default simulated failure h upstream task h submit i upstream task i submit a wrapper task a submit i sim failure child flow a b wrapper task b submit sim failure child flow b sim failure child flow b wait for c wrapper task c submit j downstream task j submit a c sim failure downstream task j k downstream task k submit wait for return j j k k if name main task wrapped deployments sim failure simulatedfailure child flow a false child flow b true downstream task j false error browers chrome firefox safari edge prefect version text version api version python version git commit built thu mar pm os arch darwin profile default server type cloud additional context no response
1
468,095
13,461,639,873
IssuesEvent
2020-09-09 15:03:14
hubmapconsortium/ccf-ui
https://api.github.com/repos/hubmapconsortium/ccf-ui
closed
As a researcher I need to in the Ingestion Portal modal that contains the app have a means to close the modal so that I can return to the Portal to add another registration or other task
Priority: Medium High SP:1 user-story
Ensure the researcher can get back to the Ingestion Portal by: * Providing a BACK button in the Logo Bar * Closing the modal when the user clicks the BACK button and revealing the Ingestion Portal that the modal overlays * Providing a chevron icon that points to the left and the text "BACK" Original ID: 3
1.0
As a researcher I need to in the Ingestion Portal modal that contains the app have a means to close the modal so that I can return to the Portal to add another registration or other task - Ensure the researcher can get back to the Ingestion Portal by: * Providing a BACK button in the Logo Bar * Closing the modal when the user clicks the BACK button and revealing the Ingestion Portal that the modal overlays * Providing a chevron icon that points to the left and the text "BACK" Original ID: 3
priority
as a researcher i need to in the ingestion portal modal that contains the app have a means to close the modal so that i can return to the portal to add another registration or other task ensure the researcher can get back to the ingestion portal by providing a back button in the logo bar closing the modal when the user clicks the back button and revealing the ingestion portal that the modal overlays providing a chevron icon that points to the left and the text back original id
1
238,206
7,776,362,661
IssuesEvent
2018-06-05 07:48:32
strapi/strapi
https://api.github.com/repos/strapi/strapi
closed
Relations with 1-N to same content-type becomes 1-1 relation
priority: medium status: confirmed 👍 type: bug 🐛
**Informations** - **Node.js version**: 9.9.0 - **npm version**: 5.6.0 - **Strapi version**: 3.0.0-alpha.12.1.3 - **Database**: MongoDB - **Operating system**: Windows 10 **What is the current behavior?** Adding a relation from a content-type to itself does not work for 1-N relations. After creation of the content-type, the server restarts and clicking on the content-type shows that the relation is indeed 1-N. But when creating a new object of that content-type, the content-type becomes a 1-1 relation for no apparant reason. N-N relations also work like they should. **Steps to reproduce the problem** 1) create a new content-type with a name and a 1-N relation with itself 2) navigate to the view of all content-types and add a new object of the content-type with a name, and save it 3) add a second object of that content-type in similar fashion 4) add a third object of that content-type and attempt to add the other 2 objects of that content-type to the object. The dropdown list now only allows to choose 1 of the 2 other objects, not both. 5) returning to the content-type settings, the content-type magically became a 1-1 relation **What is the expected behavior?** 1-N relations should get saved and not change to 1-1 when creating a new object **Suggested solutions** n.a. - [x] I'm sure that this feature hasn't already been referenced.
1.0
Relations with 1-N to same content-type becomes 1-1 relation - **Informations** - **Node.js version**: 9.9.0 - **npm version**: 5.6.0 - **Strapi version**: 3.0.0-alpha.12.1.3 - **Database**: MongoDB - **Operating system**: Windows 10 **What is the current behavior?** Adding a relation from a content-type to itself does not work for 1-N relations. After creation of the content-type, the server restarts and clicking on the content-type shows that the relation is indeed 1-N. But when creating a new object of that content-type, the content-type becomes a 1-1 relation for no apparant reason. N-N relations also work like they should. **Steps to reproduce the problem** 1) create a new content-type with a name and a 1-N relation with itself 2) navigate to the view of all content-types and add a new object of the content-type with a name, and save it 3) add a second object of that content-type in similar fashion 4) add a third object of that content-type and attempt to add the other 2 objects of that content-type to the object. The dropdown list now only allows to choose 1 of the 2 other objects, not both. 5) returning to the content-type settings, the content-type magically became a 1-1 relation **What is the expected behavior?** 1-N relations should get saved and not change to 1-1 when creating a new object **Suggested solutions** n.a. - [x] I'm sure that this feature hasn't already been referenced.
priority
relations with n to same content type becomes relation informations node js version npm version strapi version alpha database mongodb operating system windows what is the current behavior adding a relation from a content type to itself does not work for n relations after creation of the content type the server restarts and clicking on the content type shows that the relation is indeed n but when creating a new object of that content type the content type becomes a relation for no apparant reason n n relations also work like they should steps to reproduce the problem create a new content type with a name and a n relation with itself navigate to the view of all content types and add a new object of the content type with a name and save it add a second object of that content type in similar fashion add a third object of that content type and attempt to add the other objects of that content type to the object the dropdown list now only allows to choose of the other objects not both returning to the content type settings the content type magically became a relation what is the expected behavior n relations should get saved and not change to when creating a new object suggested solutions n a i m sure that this feature hasn t already been referenced
1
426,822
12,379,906,568
IssuesEvent
2020-05-19 13:15:00
ooni/probe-engine
https://api.github.com/repos/ooni/probe-engine
closed
jsonapi: disable logging body by default to avoid leaks
effort/S enhancement ooni/probe-engine priority/medium
When finalizing https://github.com/ooni/spec/pull/175, it occurred to me that probably want to guarantee that we aren't leaking information exchanged with orchestra, especially when this information is guaranteed to contain secrets. This is not an issue for now, because we are not using the jsonapi to exchange any secrets. But this may change as we evolve the `tor` experiment.
1.0
jsonapi: disable logging body by default to avoid leaks - When finalizing https://github.com/ooni/spec/pull/175, it occurred to me that probably want to guarantee that we aren't leaking information exchanged with orchestra, especially when this information is guaranteed to contain secrets. This is not an issue for now, because we are not using the jsonapi to exchange any secrets. But this may change as we evolve the `tor` experiment.
priority
jsonapi disable logging body by default to avoid leaks when finalizing it occurred to me that probably want to guarantee that we aren t leaking information exchanged with orchestra especially when this information is guaranteed to contain secrets this is not an issue for now because we are not using the jsonapi to exchange any secrets but this may change as we evolve the tor experiment
1
128,377
5,060,459,074
IssuesEvent
2016-12-22 12:03:00
HBHWoolacotts/RPii
https://api.github.com/repos/HBHWoolacotts/RPii
opened
Unable to remove email address from this customer's account - keeps coming back!
Label: General RP Bugs and Support Priority - Medium
Every time we remove it, it comes back! Also, the window doesn't scale correctly as the VAT flag field is overlapping the save & close buttons at the bottom. ![image](https://cloud.githubusercontent.com/assets/10868496/21425038/75ec6878-c83e-11e6-9195-30bce9445416.png)
1.0
Unable to remove email address from this customer's account - keeps coming back! - Every time we remove it, it comes back! Also, the window doesn't scale correctly as the VAT flag field is overlapping the save & close buttons at the bottom. ![image](https://cloud.githubusercontent.com/assets/10868496/21425038/75ec6878-c83e-11e6-9195-30bce9445416.png)
priority
unable to remove email address from this customer s account keeps coming back every time we remove it it comes back also the window doesn t scale correctly as the vat flag field is overlapping the save close buttons at the bottom
1
608,152
18,816,551,070
IssuesEvent
2021-11-10 00:22:54
ouch-org/ouch
https://api.github.com/repos/ouch-org/ouch
closed
Add nested folders to automated compression and decompression tests.
medium priority
Currently, we only test archive formats with multiple regular random files, there's no nested folders testing, so for that formats the tests need to be refactored.
1.0
Add nested folders to automated compression and decompression tests. - Currently, we only test archive formats with multiple regular random files, there's no nested folders testing, so for that formats the tests need to be refactored.
priority
add nested folders to automated compression and decompression tests currently we only test archive formats with multiple regular random files there s no nested folders testing so for that formats the tests need to be refactored
1
454,800
13,107,693,432
IssuesEvent
2020-08-04 15:38:05
department-of-veterans-affairs/caseflow
https://api.github.com/repos/department-of-veterans-affairs/caseflow
opened
Remove “Mark task complete” as an available action for BvaDispatchTasks
Feature: generic-queue Priority: Medium Product: caseflow-queue Stakeholder: BVA Team: Echo 🐬 Type: Enhancement
`BvaDispatchTask`s are automatically marked complete when an appeal is outcoded through IDT. The the only way a `BvaDispatchTask` can be considered "complete" is if the appeal is dispatched by outcoding the appeal through IDT. Currently, `MARK_COMPLETE` is an available action for the assignee of a `BvaDispatchTask`. This should not be provided as an option so that users cannot manually close their tasks without completing the work. ### Relevant [Batteam thread](https://dsva.slack.com/archives/CHX8FMP28/p1596549215130500) that spawned this Where we determine actions available for these tasks https://github.com/department-of-veterans-affairs/caseflow/blob/6256e1107c81c0df73ff7466bd0b5e533a797f63/app/models/tasks/bva_dispatch_task.rb#L8-L18 ### AC - [ ] `BvaDispatchTask`s cannot be marked complete by anyone
1.0
Remove “Mark task complete” as an available action for BvaDispatchTasks - `BvaDispatchTask`s are automatically marked complete when an appeal is outcoded through IDT. The the only way a `BvaDispatchTask` can be considered "complete" is if the appeal is dispatched by outcoding the appeal through IDT. Currently, `MARK_COMPLETE` is an available action for the assignee of a `BvaDispatchTask`. This should not be provided as an option so that users cannot manually close their tasks without completing the work. ### Relevant [Batteam thread](https://dsva.slack.com/archives/CHX8FMP28/p1596549215130500) that spawned this Where we determine actions available for these tasks https://github.com/department-of-veterans-affairs/caseflow/blob/6256e1107c81c0df73ff7466bd0b5e533a797f63/app/models/tasks/bva_dispatch_task.rb#L8-L18 ### AC - [ ] `BvaDispatchTask`s cannot be marked complete by anyone
priority
remove “mark task complete” as an available action for bvadispatchtasks bvadispatchtask s are automatically marked complete when an appeal is outcoded through idt the the only way a bvadispatchtask can be considered complete is if the appeal is dispatched by outcoding the appeal through idt currently mark complete is an available action for the assignee of a bvadispatchtask this should not be provided as an option so that users cannot manually close their tasks without completing the work relevant that spawned this where we determine actions available for these tasks ac bvadispatchtask s cannot be marked complete by anyone
1
78,297
3,509,559,163
IssuesEvent
2016-01-08 23:27:37
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Bugged sit (BB #1067)
migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** Ezopko **Original Date:** 18.07.2015 20:05:39 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1067 <hr> Bugged sit at all chair http://ctrlv.cz/CQoh
1.0
Bugged sit (BB #1067) - This issue was migrated from bitbucket. **Original Reporter:** Ezopko **Original Date:** 18.07.2015 20:05:39 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1067 <hr> Bugged sit at all chair http://ctrlv.cz/CQoh
priority
bugged sit bb this issue was migrated from bitbucket original reporter ezopko original date gmt original priority major original type bug original state resolved direct link bugged sit at all chair
1
603,715
18,670,634,871
IssuesEvent
2021-10-30 16:38:49
fslaborg/RProvider
https://api.github.com/repos/fslaborg/RProvider
closed
Project build when referencing RProvider hangs forever (ubuntu-latest - github actions)
Type: Bug Status: Available Priority: Medium
**Describe the bug** Projects that reference RProvider fail to build when using Ubuntu 20.04 as the OS within Github Actions using FAKE. When dotnet build is called for the specific project, it hangs forever. **To Reproduce** Steps to reproduce the behavior: 1. Enable ubuntu-latest in the github actions script for RProvider itself or Deedle (see fslaborg/deedle#538) 2. The build will hang forever on Ubuntu agent but complete on a macos or windows agent. For Deedle, this is the RProvider plugin project; for RProvider, this is the test project. **Expected behavior** No hangs and successful build. **Environment (please complete the following information):** - OS: Github Actions - ubuntu latest (20.04) - OS Version: 20.04 - Using in script or library: library - RProvider 2.0.1-beta2 - Installed R Version 4.0.2
1.0
Project build when referencing RProvider hangs forever (ubuntu-latest - github actions) - **Describe the bug** Projects that reference RProvider fail to build when using Ubuntu 20.04 as the OS within Github Actions using FAKE. When dotnet build is called for the specific project, it hangs forever. **To Reproduce** Steps to reproduce the behavior: 1. Enable ubuntu-latest in the github actions script for RProvider itself or Deedle (see fslaborg/deedle#538) 2. The build will hang forever on Ubuntu agent but complete on a macos or windows agent. For Deedle, this is the RProvider plugin project; for RProvider, this is the test project. **Expected behavior** No hangs and successful build. **Environment (please complete the following information):** - OS: Github Actions - ubuntu latest (20.04) - OS Version: 20.04 - Using in script or library: library - RProvider 2.0.1-beta2 - Installed R Version 4.0.2
priority
project build when referencing rprovider hangs forever ubuntu latest github actions describe the bug projects that reference rprovider fail to build when using ubuntu as the os within github actions using fake when dotnet build is called for the specific project it hangs forever to reproduce steps to reproduce the behavior enable ubuntu latest in the github actions script for rprovider itself or deedle see fslaborg deedle the build will hang forever on ubuntu agent but complete on a macos or windows agent for deedle this is the rprovider plugin project for rprovider this is the test project expected behavior no hangs and successful build environment please complete the following information os github actions ubuntu latest os version using in script or library library rprovider installed r version
1
563,578
16,701,118,520
IssuesEvent
2021-06-09 02:38:44
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
opened
Add PluralRules GetKeywords functionality
S-medium T-core discuss-priority
ICU4X currently has no functionality to return an enumeration of keywords that are supported by a PluralRules object for a given locale and `Cardinal`/`Ordinal` setting. The equivalent call that we currently use in SpiderMonkey is `uplrules_getKeywords()` https://unicode-org.github.io/icu-docs/apidoc/dev/icu4c/upluralrules_8h.html#a5f8a89ec2b58323e9f96ac0bf2c13ab5
1.0
Add PluralRules GetKeywords functionality - ICU4X currently has no functionality to return an enumeration of keywords that are supported by a PluralRules object for a given locale and `Cardinal`/`Ordinal` setting. The equivalent call that we currently use in SpiderMonkey is `uplrules_getKeywords()` https://unicode-org.github.io/icu-docs/apidoc/dev/icu4c/upluralrules_8h.html#a5f8a89ec2b58323e9f96ac0bf2c13ab5
priority
add pluralrules getkeywords functionality currently has no functionality to return an enumeration of keywords that are supported by a pluralrules object for a given locale and cardinal ordinal setting the equivalent call that we currently use in spidermonkey is uplrules getkeywords
1
623,745
19,677,461,295
IssuesEvent
2022-01-11 13:48:37
haydarsinan/SWE573-2021-Fall
https://api.github.com/repos/haydarsinan/SWE573-2021-Fall
closed
Geolocation Research for Services and Events
High Priority Medium Work Learning Research
Closer events and services should be shown to users firstly depending on their location. Therefore, geolocation feature should be implemented and researched.
1.0
Geolocation Research for Services and Events - Closer events and services should be shown to users firstly depending on their location. Therefore, geolocation feature should be implemented and researched.
priority
geolocation research for services and events closer events and services should be shown to users firstly depending on their location therefore geolocation feature should be implemented and researched
1
696,688
23,911,663,742
IssuesEvent
2022-09-09 08:47:42
Cheos137/ArmorpointsPlusplus
https://api.github.com/repos/Cheos137/ArmorpointsPlusplus
closed
[Feature Request]: option to disable compatibility warnings
feature request minecraft: all supported future update priority: medium
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description Add a new config option to hide messages warning about incompatibility with other mods. ### Solution Add a new config option to hide messages warning about incompatibility with other mods. ### Alternatives _No response_ ### Additional Context _No response_
1.0
[Feature Request]: option to disable compatibility warnings - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description Add a new config option to hide messages warning about incompatibility with other mods. ### Solution Add a new config option to hide messages warning about incompatibility with other mods. ### Alternatives _No response_ ### Additional Context _No response_
priority
option to disable compatibility warnings is there an existing issue for this i have searched the existing issues description add a new config option to hide messages warning about incompatibility with other mods solution add a new config option to hide messages warning about incompatibility with other mods alternatives no response additional context no response
1
351,344
10,516,076,161
IssuesEvent
2019-09-28 14:55:48
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
closed
Characters don't get dryad blood
:beetle: bug :beetle: :grey_exclamation: priority medium
**Mod Version** 6fc3865c **What expansions do you have installed?** All **Please explain your issue in as much detail as possible:** Characters don't get dryad blood **Steps to reproduce the issue:** Marry dryad to centaur or frost nymph **Upload an attachment below: .zip of your save, or screenshots:** <details> <summary>Click to expand</summary> ![image](https://user-images.githubusercontent.com/46576860/65746389-6cecf500-e107-11e9-877d-63d6b8b6d7fa.png) ![image](https://user-images.githubusercontent.com/46576860/65746607-f13f7800-e107-11e9-9b6e-af30b2552ded.png) </details>
1.0
Characters don't get dryad blood - **Mod Version** 6fc3865c **What expansions do you have installed?** All **Please explain your issue in as much detail as possible:** Characters don't get dryad blood **Steps to reproduce the issue:** Marry dryad to centaur or frost nymph **Upload an attachment below: .zip of your save, or screenshots:** <details> <summary>Click to expand</summary> ![image](https://user-images.githubusercontent.com/46576860/65746389-6cecf500-e107-11e9-877d-63d6b8b6d7fa.png) ![image](https://user-images.githubusercontent.com/46576860/65746607-f13f7800-e107-11e9-9b6e-af30b2552ded.png) </details>
priority
characters don t get dryad blood mod version what expansions do you have installed all please explain your issue in as much detail as possible characters don t get dryad blood steps to reproduce the issue marry dryad to centaur or frost nymph upload an attachment below zip of your save or screenshots click to expand
1
68,358
3,286,691,490
IssuesEvent
2015-10-29 05:03:41
cs2103aug2015-w10-4j/main
https://api.github.com/repos/cs2103aug2015-w10-4j/main
closed
As an advanced user, I can reorder the command given to create a task
priority.medium type.story
...so that I can focus more on the task rather than how to input the task
1.0
As an advanced user, I can reorder the command given to create a task - ...so that I can focus more on the task rather than how to input the task
priority
as an advanced user i can reorder the command given to create a task so that i can focus more on the task rather than how to input the task
1
208,752
7,158,110,450
IssuesEvent
2018-01-26 22:43:28
IfyAniefuna/experiment_metadata
https://api.github.com/repos/IfyAniefuna/experiment_metadata
opened
Include nutrient source units (g/L, etc) for within generated metadata
enhancement medium priority
The units may change over time, therefore the units being asked for by the current webapp needs to be automatically included in the generated metadata, example: user enters Glucose of 2 (g/L) in the webapp. The generated metadata form has the following: "carbon-source","Glucose(2g/L)"
1.0
Include nutrient source units (g/L, etc) for within generated metadata - The units may change over time, therefore the units being asked for by the current webapp needs to be automatically included in the generated metadata, example: user enters Glucose of 2 (g/L) in the webapp. The generated metadata form has the following: "carbon-source","Glucose(2g/L)"
priority
include nutrient source units g l etc for within generated metadata the units may change over time therefore the units being asked for by the current webapp needs to be automatically included in the generated metadata example user enters glucose of g l in the webapp the generated metadata form has the following carbon source glucose l
1
537,364
15,727,813,206
IssuesEvent
2021-03-29 13:07:47
SteamedHams722/automatic-octopus
https://api.github.com/repos/SteamedHams722/automatic-octopus
closed
Use Django instead of Heroku for Postgres
Medium Priority
Why It may make more sense to just use Django for hosting the Postgres db. The heroku hobby versions are very limiting and won't be able to handle the regular data loads. How - Investigate how Django will handle this data. - Set-up a postgres db that can run in Django - Get rid of the Heroku db.
1.0
Use Django instead of Heroku for Postgres - Why It may make more sense to just use Django for hosting the Postgres db. The heroku hobby versions are very limiting and won't be able to handle the regular data loads. How - Investigate how Django will handle this data. - Set-up a postgres db that can run in Django - Get rid of the Heroku db.
priority
use django instead of heroku for postgres why it may make more sense to just use django for hosting the postgres db the heroku hobby versions are very limiting and won t be able to handle the regular data loads how investigate how django will handle this data set up a postgres db that can run in django get rid of the heroku db
1
32,672
2,757,596,123
IssuesEvent
2015-04-27 15:40:15
pydata/numexpr
https://api.github.com/repos/pydata/numexpr
closed
Numexpr query with empty string does not work (discovered using pytables)
bug imported Priority-Medium
_From [daniel.h...@fmi.ch](https://code.google.com/u/113211097457152931757/) on December 03, 2013 12:37:10_ See https://github.com/PyTables/PyTables/issues/184 _Original issue: http://code.google.com/p/numexpr/issues/detail?id=121_
1.0
Numexpr query with empty string does not work (discovered using pytables) - _From [daniel.h...@fmi.ch](https://code.google.com/u/113211097457152931757/) on December 03, 2013 12:37:10_ See https://github.com/PyTables/PyTables/issues/184 _Original issue: http://code.google.com/p/numexpr/issues/detail?id=121_
priority
numexpr query with empty string does not work discovered using pytables from on december see original issue
1
765,922
26,865,725,822
IssuesEvent
2023-02-03 23:24:41
clt313/SuperballVR
https://api.github.com/repos/clt313/SuperballVR
closed
Update audio to use singleton pattern
priority: medium
Rather than juggling audio sources like we have been, use an AudioManager instead. Useful video: https://www.youtube.com/watch?v=6OT43pvUyfY&ab_channel=Brackeys
1.0
Update audio to use singleton pattern - Rather than juggling audio sources like we have been, use an AudioManager instead. Useful video: https://www.youtube.com/watch?v=6OT43pvUyfY&ab_channel=Brackeys
priority
update audio to use singleton pattern rather than juggling audio sources like we have been use an audiomanager instead useful video
1
321,342
9,797,389,890
IssuesEvent
2019-06-11 09:50:25
nhsconnect/gpconnect-provider-testing
https://api.github.com/repos/nhsconnect/gpconnect-provider-testing
closed
Default date range if only a start or end date parameter passed in getCareRecord
0.7.0 HTML Medium priority
the scenario where only a start or an end date need to be considered and the GP Connect program needs to update the specification to cater for the possibility of someone only sending a start date or an end date date banner section uplifted to add requirements on expected date range behaviours
1.0
Default date range if only a start or end date parameter passed in getCareRecord - the scenario where only a start or an end date need to be considered and the GP Connect program needs to update the specification to cater for the possibility of someone only sending a start date or an end date date banner section uplifted to add requirements on expected date range behaviours
priority
default date range if only a start or end date parameter passed in getcarerecord the scenario where only a start or an end date need to be considered and the gp connect program needs to update the specification to cater for the possibility of someone only sending a start date or an end date date banner section uplifted to add requirements on expected date range behaviours
1
428,536
12,412,966,962
IssuesEvent
2020-05-22 11:42:49
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Law spam
Priority: Medium Status: Fixed
2. Spam from law notification in chat. Step to reproduce: - create law like this: ![image](https://user-images.githubusercontent.com/45708377/82206814-92867300-9911-11ea-9b0f-8da1ef09304e.png) - start to place something 3 times. I have very strange notifications (I have civics debug off): ![image](https://user-images.githubusercontent.com/45708377/82206221-83eb8c00-9910-11ea-8a21-1407363fcb3a.png) - if you place it very fast it can appear here: ![image](https://user-images.githubusercontent.com/45708377/82207010-da0cff00-9911-11ea-8a2b-a333808a0faa.png) 3. Notification from executive office. Step to reproduce: - create executive action like: ![image](https://user-images.githubusercontent.com/45708377/82208825-e5155e80-9914-11ea-8cf4-3fd05dc2f3ec.png) - /civics winelection. This notification doesn't always appear: ![image](https://user-images.githubusercontent.com/45708377/82208797-d9c23300-9914-11ea-90ca-0e19f7bf298d.png) - I can't see it in chat only in tooltip: ![image](https://user-images.githubusercontent.com/45708377/82209121-6967e180-9915-11ea-94ae-ce36c29b5883.png)
1.0
Law spam - 2. Spam from law notification in chat. Step to reproduce: - create law like this: ![image](https://user-images.githubusercontent.com/45708377/82206814-92867300-9911-11ea-9b0f-8da1ef09304e.png) - start to place something 3 times. I have very strange notifications (I have civics debug off): ![image](https://user-images.githubusercontent.com/45708377/82206221-83eb8c00-9910-11ea-8a21-1407363fcb3a.png) - if you place it very fast it can appear here: ![image](https://user-images.githubusercontent.com/45708377/82207010-da0cff00-9911-11ea-8a2b-a333808a0faa.png) 3. Notification from executive office. Step to reproduce: - create executive action like: ![image](https://user-images.githubusercontent.com/45708377/82208825-e5155e80-9914-11ea-8cf4-3fd05dc2f3ec.png) - /civics winelection. This notification doesn't always appear: ![image](https://user-images.githubusercontent.com/45708377/82208797-d9c23300-9914-11ea-90ca-0e19f7bf298d.png) - I can't see it in chat only in tooltip: ![image](https://user-images.githubusercontent.com/45708377/82209121-6967e180-9915-11ea-94ae-ce36c29b5883.png)
priority
law spam spam from law notification in chat step to reproduce create law like this start to place something times i have very strange notifications i have civics debug off if you place it very fast it can appear here notification from executive office step to reproduce create executive action like civics winelection this notification doesn t always appear i can t see it in chat only in tooltip
1
645,145
20,996,211,346
IssuesEvent
2022-03-29 13:42:31
HabitRPG/habitica-android
https://api.github.com/repos/HabitRPG/habitica-android
closed
External links not working from Guild pages
Type: Bug Help wanted Priority: medium Status: pr: ready for review
User report: (also reproduced by @voidbytes ) "Device: LGE LM-V600 Android Version: 30 AppVersion: Version 3.4.1.1 (3090) User ID: d2656f22-44e5-4239-ad66-e0d40305090f Level: 5 Class: warrior Is in Inn: false Uses Costume: false Custom Day Start: 0 Timezone Offset: 300 Details: Unable to click external links in guild pages including wiki pages. Verified with second user that the issue exists. "
1.0
External links not working from Guild pages - User report: (also reproduced by @voidbytes ) "Device: LGE LM-V600 Android Version: 30 AppVersion: Version 3.4.1.1 (3090) User ID: d2656f22-44e5-4239-ad66-e0d40305090f Level: 5 Class: warrior Is in Inn: false Uses Costume: false Custom Day Start: 0 Timezone Offset: 300 Details: Unable to click external links in guild pages including wiki pages. Verified with second user that the issue exists. "
priority
external links not working from guild pages user report also reproduced by voidbytes device lge lm android version appversion version user id level class warrior is in inn false uses costume false custom day start timezone offset details unable to click external links in guild pages including wiki pages verified with second user that the issue exists
1
358,733
10,631,790,535
IssuesEvent
2019-10-15 09:04:56
ooni/probe-android
https://api.github.com/repos/ooni/probe-android
closed
Publish app apk to github release
enhancement priority/medium
We would like to add a direct link to download the OONI Probe app apk on Android to our website (see: https://github.com/TheTorProject/ooni-web/issues/277). It would be great if during the release process we add an extra step about uploading the built apk to the relevant github release.
1.0
Publish app apk to github release - We would like to add a direct link to download the OONI Probe app apk on Android to our website (see: https://github.com/TheTorProject/ooni-web/issues/277). It would be great if during the release process we add an extra step about uploading the built apk to the relevant github release.
priority
publish app apk to github release we would like to add a direct link to download the ooni probe app apk on android to our website see it would be great if during the release process we add an extra step about uploading the built apk to the relevant github release
1
643,909
20,961,610,727
IssuesEvent
2022-03-27 21:49:08
abedmaatalla/sipdroid
https://api.github.com/repos/abedmaatalla/sipdroid
closed
Allow customized "User-Agent" header
Priority-Medium auto-migrated Type-Enhancement
``` Some providers use "User-Agent" header to validate registration. Would it possible to allow customized user agent when setting up a line? ``` Original issue reported on code.google.com by `zhang...@gmail.com` on 26 Nov 2012 at 7:04
1.0
Allow customized "User-Agent" header - ``` Some providers use "User-Agent" header to validate registration. Would it possible to allow customized user agent when setting up a line? ``` Original issue reported on code.google.com by `zhang...@gmail.com` on 26 Nov 2012 at 7:04
priority
allow customized user agent header some providers use user agent header to validate registration would it possible to allow customized user agent when setting up a line original issue reported on code google com by zhang gmail com on nov at
1
496,241
14,343,581,275
IssuesEvent
2020-11-28 10:11:39
greasemonkey123/Redefine-A
https://api.github.com/repos/greasemonkey123/Redefine-A
opened
Better group admin settings
Medium priority Redefine:A enhancement
ability to add multiple groups to the admin list and a way to change the rank tolerance type from ">=" to "=", etc. Suggested by ShhhhhhhhhhhBro. - [ ] Loader edit. - [ ] Module edit. - [ ] Allow easily identifiable settings.
1.0
Better group admin settings - ability to add multiple groups to the admin list and a way to change the rank tolerance type from ">=" to "=", etc. Suggested by ShhhhhhhhhhhBro. - [ ] Loader edit. - [ ] Module edit. - [ ] Allow easily identifiable settings.
priority
better group admin settings ability to add multiple groups to the admin list and a way to change the rank tolerance type from to etc suggested by shhhhhhhhhhhbro loader edit module edit allow easily identifiable settings
1
597,445
18,163,665,677
IssuesEvent
2021-09-27 12:33:52
Undecided-Discord/Oshavery-Server
https://api.github.com/repos/Undecided-Discord/Oshavery-Server
opened
WebHook対応ほしい、ほしくない?
enhancement Priority: medium
## 概要 <!-- 概要を書いてください --> とりあえずGitHubのWebHookがほしいです ## なぜIssueを作ったか <!-- Issueを作ろうと思った理由 --> 開発の通知がほしいので... ## 参考 <!-- 参考になるものがあれば貼ってください --> ![image](https://user-images.githubusercontent.com/40442980/134908710-fd552206-3326-4841-a2ea-783bc63d4d12.png) ## 必要なこと <!-- わかるのであればタスク一覧を書いてください --> - [ ] embedの実装
1.0
WebHook対応ほしい、ほしくない? - ## 概要 <!-- 概要を書いてください --> とりあえずGitHubのWebHookがほしいです ## なぜIssueを作ったか <!-- Issueを作ろうと思った理由 --> 開発の通知がほしいので... ## 参考 <!-- 参考になるものがあれば貼ってください --> ![image](https://user-images.githubusercontent.com/40442980/134908710-fd552206-3326-4841-a2ea-783bc63d4d12.png) ## 必要なこと <!-- わかるのであればタスク一覧を書いてください --> - [ ] embedの実装
priority
webhook対応ほしい、ほしくない? 概要 とりあえずgithubのwebhookがほしいです なぜissueを作ったか 開発の通知がほしいので 参考 必要なこと embedの実装
1
24,385
2,667,388,995
IssuesEvent
2015-03-22 15:25:34
NewCreature/EOF
https://api.github.com/repos/NewCreature/EOF
opened
Allow MIDI playback
enhancement imported Priority-Medium
_From [raynebc](https://code.google.com/u/raynebc/) on June 23, 2010 12:22:14_ I was thinking about how work-intensive it is to prepare a new set of vocal tone samples. It seems to be that since Allegro has built-in support for MIDI, it may be worthwhile to allow it to be a user-option to try to initialize the system's default MIDI device so that EOF's output can optionally be sent to a software/hardware synthesizer. Besides just sending Note On and Off data, we'll need to consider how to allow the user to specify which synth tone to use. When I created the piano tones for EOF, I learned a bit about this process, but haven't mastered it fully. I'm looking into the creation of Cakewalk instrument definition files, which seem to be the most common format of instrument definitions. Anvil Studio will allow you to define information for its internal use if the MIDI package is purchased ($20 or so, but I don't know yet if it will allow an instrument definition file to be created). I'm willing to get this to test creating definition files, and with any luck could write functions to import such definitions into EOF. This would allow for a good deal of flexibility with customizing the sound/volume of vocal tones. This may also be a good way to begin MIDI integration, should EOF be given the ability to accept MIDI input in the future (ie. piano record mode, if FoF gains the ability to play keyboard tracks). Besides vocal tones, it could allow the user to supply their own clap, metronome, etc. without having to provide recorded audio samples to EOF. _Original issue: http://code.google.com/p/editor-on-fire/issues/detail?id=121_
1.0
Allow MIDI playback - _From [raynebc](https://code.google.com/u/raynebc/) on June 23, 2010 12:22:14_ I was thinking about how work-intensive it is to prepare a new set of vocal tone samples. It seems to be that since Allegro has built-in support for MIDI, it may be worthwhile to allow it to be a user-option to try to initialize the system's default MIDI device so that EOF's output can optionally be sent to a software/hardware synthesizer. Besides just sending Note On and Off data, we'll need to consider how to allow the user to specify which synth tone to use. When I created the piano tones for EOF, I learned a bit about this process, but haven't mastered it fully. I'm looking into the creation of Cakewalk instrument definition files, which seem to be the most common format of instrument definitions. Anvil Studio will allow you to define information for its internal use if the MIDI package is purchased ($20 or so, but I don't know yet if it will allow an instrument definition file to be created). I'm willing to get this to test creating definition files, and with any luck could write functions to import such definitions into EOF. This would allow for a good deal of flexibility with customizing the sound/volume of vocal tones. This may also be a good way to begin MIDI integration, should EOF be given the ability to accept MIDI input in the future (ie. piano record mode, if FoF gains the ability to play keyboard tracks). Besides vocal tones, it could allow the user to supply their own clap, metronome, etc. without having to provide recorded audio samples to EOF. _Original issue: http://code.google.com/p/editor-on-fire/issues/detail?id=121_
priority
allow midi playback from on june i was thinking about how work intensive it is to prepare a new set of vocal tone samples it seems to be that since allegro has built in support for midi it may be worthwhile to allow it to be a user option to try to initialize the system s default midi device so that eof s output can optionally be sent to a software hardware synthesizer besides just sending note on and off data we ll need to consider how to allow the user to specify which synth tone to use when i created the piano tones for eof i learned a bit about this process but haven t mastered it fully i m looking into the creation of cakewalk instrument definition files which seem to be the most common format of instrument definitions anvil studio will allow you to define information for its internal use if the midi package is purchased or so but i don t know yet if it will allow an instrument definition file to be created i m willing to get this to test creating definition files and with any luck could write functions to import such definitions into eof this would allow for a good deal of flexibility with customizing the sound volume of vocal tones this may also be a good way to begin midi integration should eof be given the ability to accept midi input in the future ie piano record mode if fof gains the ability to play keyboard tracks besides vocal tones it could allow the user to supply their own clap metronome etc without having to provide recorded audio samples to eof original issue
1
394,205
11,633,367,715
IssuesEvent
2020-02-28 08:04:40
Repair-DeskPOS/RepairDesk-BUGS-IMPROVEMENTS
https://api.github.com/repos/Repair-DeskPOS/RepairDesk-BUGS-IMPROVEMENTS
closed
Issue when adding inventory items to ticket
Medium Priority enhancement
Cannot add more than 1 item to an inventory item to a ticket after it has been created. We need a solution for if a tech comes across a defective part and how we can take multiple out of iinventory
1.0
Issue when adding inventory items to ticket - Cannot add more than 1 item to an inventory item to a ticket after it has been created. We need a solution for if a tech comes across a defective part and how we can take multiple out of iinventory
priority
issue when adding inventory items to ticket cannot add more than item to an inventory item to a ticket after it has been created we need a solution for if a tech comes across a defective part and how we can take multiple out of iinventory
1
40,942
2,868,956,368
IssuesEvent
2015-06-05 22:11:23
dart-lang/pub
https://api.github.com/repos/dart-lang/pub
closed
pub serve protocol enhancements
DevStory enhancement NotPlanned Priority-Medium
<a href="https://github.com/devoncarew"><img src="https://avatars.githubusercontent.com/u/1269969?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [devoncarew](https://github.com/devoncarew)** _Originally opened as dart-lang/sdk#13935_ ---- Here's a proposal here: https://gist.github.com/devoncarew/6897835 and a spec that uses a similar protocol: https://developers.google.com/chrome-developer-tools/docs/protocol/tot/page
1.0
pub serve protocol enhancements - <a href="https://github.com/devoncarew"><img src="https://avatars.githubusercontent.com/u/1269969?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [devoncarew](https://github.com/devoncarew)** _Originally opened as dart-lang/sdk#13935_ ---- Here's a proposal here: https://gist.github.com/devoncarew/6897835 and a spec that uses a similar protocol: https://developers.google.com/chrome-developer-tools/docs/protocol/tot/page
priority
pub serve protocol enhancements issue by originally opened as dart lang sdk here s a proposal here and a spec that uses a similar protocol
1
711,490
24,465,520,629
IssuesEvent
2022-10-07 14:42:14
canonical/ubuntu.com
https://api.github.com/repos/canonical/ubuntu.com
closed
Dead link and typographical error on ubuntu.com/about/packages
Priority: Medium
## Summary In the **Charms are packages for cloud software operations** section on ubuntu.com/about/packages, there appears to be a misplaced box at the end of the paragraph and the *Get started making a charm* link takes you to a non-existent Discourse page. ## Process 1. Navigate to https://ubuntu.com/about/packages 2. Scroll down to **Charms are packages for cloud software operations** section (see the misplaced box) 3. Click on *Get started making a charm* hyperlink. ## Current and expected result Noticed the box and did not expect it to be there. When I clicked *Get started making a charm* link I was taken to an "Oops not found" page rather than being taken to introductory documentation for making charms. ## Screenshot ### Misplaced box ![image](https://user-images.githubusercontent.com/40342202/176543593-e3b916e2-694f-4486-8e56-382888fd5077.png) ### Non-existent Discourse page ![image](https://user-images.githubusercontent.com/40342202/176543775-70c11c59-09d1-4527-a3ee-9bf354eb850b.png) ## Browser details Vivaldi | 5.3.2679.61 (Stable channel) stable (64-bit) -- | -- Revision | fcd44035934deba382cb933a4c62bd4fa7690078 OS | Linux JavaScript | V8 10.2.154.10 User Agent | Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/102.0.5005.136 Safari/537.36 Command Line | /usr/bin/vivaldi-stable --new-window --enable-crashpad --flag-switches-begin --flag-switches-end --save-page-as-mhtml Executable Path | /opt/vivaldi/vivaldi Profile Path | /home/nucci/.config/vivaldi/Default I also have the uBlock origin extension enabled.
1.0
Dead link and typographical error on ubuntu.com/about/packages - ## Summary In the **Charms are packages for cloud software operations** section on ubuntu.com/about/packages, there appears to be a misplaced box at the end of the paragraph and the *Get started making a charm* link takes you to a non-existent Discourse page. ## Process 1. Navigate to https://ubuntu.com/about/packages 2. Scroll down to **Charms are packages for cloud software operations** section (see the misplaced box) 3. Click on *Get started making a charm* hyperlink. ## Current and expected result Noticed the box and did not expect it to be there. When I clicked *Get started making a charm* link I was taken to an "Oops not found" page rather than being taken to introductory documentation for making charms. ## Screenshot ### Misplaced box ![image](https://user-images.githubusercontent.com/40342202/176543593-e3b916e2-694f-4486-8e56-382888fd5077.png) ### Non-existent Discourse page ![image](https://user-images.githubusercontent.com/40342202/176543775-70c11c59-09d1-4527-a3ee-9bf354eb850b.png) ## Browser details Vivaldi | 5.3.2679.61 (Stable channel) stable (64-bit) -- | -- Revision | fcd44035934deba382cb933a4c62bd4fa7690078 OS | Linux JavaScript | V8 10.2.154.10 User Agent | Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/102.0.5005.136 Safari/537.36 Command Line | /usr/bin/vivaldi-stable --new-window --enable-crashpad --flag-switches-begin --flag-switches-end --save-page-as-mhtml Executable Path | /opt/vivaldi/vivaldi Profile Path | /home/nucci/.config/vivaldi/Default I also have the uBlock origin extension enabled.
priority
dead link and typographical error on ubuntu com about packages summary in the charms are packages for cloud software operations section on ubuntu com about packages there appears to be a misplaced box at the end of the paragraph and the get started making a charm link takes you to a non existent discourse page process navigate to scroll down to charms are packages for cloud software operations section see the misplaced box click on get started making a charm hyperlink current and expected result noticed the box and did not expect it to be there when i clicked get started making a charm link i was taken to an oops not found page rather than being taken to introductory documentation for making charms screenshot misplaced box non existent discourse page browser details vivaldi   stable channel  stable  bit revision os linux javascript   user agent mozilla linux applewebkit khtml like gecko chrome safari command line usr bin vivaldi stable new window enable crashpad flag switches begin flag switches end save page as mhtml executable path opt vivaldi vivaldi profile path home nucci config vivaldi default i also have the ublock origin extension enabled
1
420,510
12,238,937,023
IssuesEvent
2020-05-04 20:43:04
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
opened
Allow "Likes" in Activity > Comments/Replies
bug component: activity priority: medium
**Describe the bug** Allow "Like" functionality in Activity > Replies. This is existing on BuddyBoss Wall with Boss Theme. **To Reproduce** Steps to reproduce the behavior: 1. Go to Activity, post something 2. Comment something to that post 3. You can only see Reply and Delete, but no Like option **Expected behavior** In BuddyBoss Wall, users have the option to Like the comments and replies on activity feed. This must be existing on BuddyBoss Platform as well. **Screenshots** ![oldtheme](https://user-images.githubusercontent.com/58522224/81011350-62a78c00-8e8a-11ea-9e93-61456af1e0e3.JPG) ![newtheme](https://user-images.githubusercontent.com/58522224/81011424-87036880-8e8a-11ea-94f9-16c05e53bfe7.JPG) **Support ticket links** https://secure.helpscout.net/conversation/1148788883/70864?folderId=3701241
1.0
Allow "Likes" in Activity > Comments/Replies - **Describe the bug** Allow "Like" functionality in Activity > Replies. This is existing on BuddyBoss Wall with Boss Theme. **To Reproduce** Steps to reproduce the behavior: 1. Go to Activity, post something 2. Comment something to that post 3. You can only see Reply and Delete, but no Like option **Expected behavior** In BuddyBoss Wall, users have the option to Like the comments and replies on activity feed. This must be existing on BuddyBoss Platform as well. **Screenshots** ![oldtheme](https://user-images.githubusercontent.com/58522224/81011350-62a78c00-8e8a-11ea-9e93-61456af1e0e3.JPG) ![newtheme](https://user-images.githubusercontent.com/58522224/81011424-87036880-8e8a-11ea-94f9-16c05e53bfe7.JPG) **Support ticket links** https://secure.helpscout.net/conversation/1148788883/70864?folderId=3701241
priority
allow likes in activity comments replies describe the bug allow like functionality in activity replies this is existing on buddyboss wall with boss theme to reproduce steps to reproduce the behavior go to activity post something comment something to that post you can only see reply and delete but no like option expected behavior in buddyboss wall users have the option to like the comments and replies on activity feed this must be existing on buddyboss platform as well screenshots support ticket links
1
365,414
10,781,161,450
IssuesEvent
2019-11-04 14:25:31
salesagility/SuiteCRM
https://api.github.com/repos/salesagility/SuiteCRM
closed
function isOwner() in SugarBean.php is not checking permissions ok
Bug Fix Proposed Medium Priority Resolved: Next Release
<!--- Provide a general summary of the issue in the **Title** above --> <!--- Before you open an issue, please check if a similar issue already exists or has been closed before. ---> <!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. ---> #### Issue <!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug --> Investigating why an user with "owner permissions" in 'Notes' module could not enter the DetailView of a record that was created by him, I found that we have a problem with the method [`isOwner()` in SugarBean.php](https://github.com/salesagility/SuiteCRM/blob/master/data/SugarBean.php#L1673) The `isOwner()` method is [checking: `isset($this->assigned_user_id)`](https://github.com/salesagility/SuiteCRM/blob/master/data/SugarBean.php#L1685) ```php } elseif (isset($this->assigned_user_id)) { if ($this->assigned_user_id == $user_id) { return true; } return false; } ``` But if `$this->assigned_user_id` is `''` the function [`isset`](https://www.php.net/isset) will return true, so the execution will enter into the `elseif` and will check `$this->assigned_user_id == $user_id` and in other words will check something like this: `'' == '7cb40869-3ca6-7c88-006c-5da61ac6dc6e'` which is `false`. This behavior prevents the execution from reaching the next check that will return `true` because the record was created by the user, so `$this->created_by == $user_id` is `true` ```php //other wise if there is a created_by that is the owner if (isset($this->created_by) && $this->created_by == $user_id) { return true; } ``` #### Expected Behavior <!--- Tell us what should happen --> `isOwner()` checks permissions in a good way ;-) #### Actual Behavior <!--- Tell us what happens instead --> <!--- Also please check relevant logs (suitecrm.log, php error.log etc.) --> A user that creates a 'Note' cannot go to the DetailView of that record. The `isOwner()` method is not checking permissions in a bad way. #### Possible Fix <!--- Not obligatory, but suggest a fix or reason for the bug --> We should use [`!empty()`](https://www.php.net/empty#refsect1-function.empty-description) instead of [`isset()`](https://www.php.net/isset) #### Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug include code to reproduce, if relevant --> 1. Create and Role and assign an user to this Role. 2. Set the followings permissions to the module Notes for this Role: ``` access: 89 delete: -99 edit: 75 export: -99 import: -99 list: 75 view: 75 ``` 3. Login with this user 4. Create a Note 5. Try to go to the DetailView. #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: 7.10.20 * Environment name and version (e.g. MySQL, PHP 7): MySQL 8.0.x / PHP 7.3.x
1.0
function isOwner() in SugarBean.php is not checking permissions ok - <!--- Provide a general summary of the issue in the **Title** above --> <!--- Before you open an issue, please check if a similar issue already exists or has been closed before. ---> <!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. ---> #### Issue <!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug --> Investigating why an user with "owner permissions" in 'Notes' module could not enter the DetailView of a record that was created by him, I found that we have a problem with the method [`isOwner()` in SugarBean.php](https://github.com/salesagility/SuiteCRM/blob/master/data/SugarBean.php#L1673) The `isOwner()` method is [checking: `isset($this->assigned_user_id)`](https://github.com/salesagility/SuiteCRM/blob/master/data/SugarBean.php#L1685) ```php } elseif (isset($this->assigned_user_id)) { if ($this->assigned_user_id == $user_id) { return true; } return false; } ``` But if `$this->assigned_user_id` is `''` the function [`isset`](https://www.php.net/isset) will return true, so the execution will enter into the `elseif` and will check `$this->assigned_user_id == $user_id` and in other words will check something like this: `'' == '7cb40869-3ca6-7c88-006c-5da61ac6dc6e'` which is `false`. This behavior prevents the execution from reaching the next check that will return `true` because the record was created by the user, so `$this->created_by == $user_id` is `true` ```php //other wise if there is a created_by that is the owner if (isset($this->created_by) && $this->created_by == $user_id) { return true; } ``` #### Expected Behavior <!--- Tell us what should happen --> `isOwner()` checks permissions in a good way ;-) #### Actual Behavior <!--- Tell us what happens instead --> <!--- Also please check relevant logs (suitecrm.log, php error.log etc.) --> A user that creates a 'Note' cannot go to the DetailView of that record. The `isOwner()` method is not checking permissions in a bad way. #### Possible Fix <!--- Not obligatory, but suggest a fix or reason for the bug --> We should use [`!empty()`](https://www.php.net/empty#refsect1-function.empty-description) instead of [`isset()`](https://www.php.net/isset) #### Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug include code to reproduce, if relevant --> 1. Create and Role and assign an user to this Role. 2. Set the followings permissions to the module Notes for this Role: ``` access: 89 delete: -99 edit: 75 export: -99 import: -99 list: 75 view: 75 ``` 3. Login with this user 4. Create a Note 5. Try to go to the DetailView. #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: 7.10.20 * Environment name and version (e.g. MySQL, PHP 7): MySQL 8.0.x / PHP 7.3.x
priority
function isowner in sugarbean php is not checking permissions ok issue investigating why an user with owner permissions in notes module could not enter the detailview of a record that was created by him i found that we have a problem with the method the isowner method is php elseif isset this assigned user id if this assigned user id user id return true return false but if this assigned user id is the function will return true so the execution will enter into the elseif and will check this assigned user id user id and in other words will check something like this which is false this behavior prevents the execution from reaching the next check that will return true because the record was created by the user so this created by user id is true php other wise if there is a created by that is the owner if isset this created by this created by user id return true expected behavior isowner checks permissions in a good way actual behavior a user that creates a note cannot go to the detailview of that record the isowner method is not checking permissions in a bad way possible fix we should use instead of steps to reproduce create and role and assign an user to this role set the followings permissions to the module notes for this role access delete edit export import list view login with this user create a note try to go to the detailview your environment suitecrm version used environment name and version e g mysql php mysql x php x
1
16,259
2,612,939,294
IssuesEvent
2015-02-27 17:37:55
skelterjohn/gomatrix
https://api.github.com/repos/skelterjohn/gomatrix
opened
support for native linear algebra libraries
auto-migrated Priority-Medium Type-Enhancement
``` Such as LAPACK, LINPACK or others. ``` Original issue reported on code.google.com by `jasm...@gmail.com` on 25 Nov 2009 at 10:04
1.0
support for native linear algebra libraries - ``` Such as LAPACK, LINPACK or others. ``` Original issue reported on code.google.com by `jasm...@gmail.com` on 25 Nov 2009 at 10:04
priority
support for native linear algebra libraries such as lapack linpack or others original issue reported on code google com by jasm gmail com on nov at
1
786,663
27,661,544,131
IssuesEvent
2023-03-12 15:27:37
AY2223S2-CS2113-T12-1/tp
https://api.github.com/repos/AY2223S2-CS2113-T12-1/tp
opened
[Task] Edit expenses and income
type.Task priority.Medium
Add functionality for user to edit existing expense and income. Edits the item at the specified INDEX, where INDEX must be accessible by the respective arrayList. Fiels provided will be updated to the new inputs.
1.0
[Task] Edit expenses and income - Add functionality for user to edit existing expense and income. Edits the item at the specified INDEX, where INDEX must be accessible by the respective arrayList. Fiels provided will be updated to the new inputs.
priority
edit expenses and income add functionality for user to edit existing expense and income edits the item at the specified index where index must be accessible by the respective arraylist fiels provided will be updated to the new inputs
1
356,035
10,587,744,270
IssuesEvent
2019-10-08 23:09:59
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] Create APIs to support "Publishing status dashboard"
enhancement priority: medium
**In relation to #3352...** We need new APIs to be able to support #3352's requests. **Describe the solution you'd like** Required APIs would be - Get packages - retrieves the list of packages on the publishing queue. Each package should have at a minimum scheduled date, user who submitted, status, environment, comment and package id. It should support several filters namely a path expression, environments, and states. - Get package details - retrieves the list of files associated with the package by package id. If this is not expensive to do at once on the `get packages` call, please do fit it there and we wouldn't need this second call. - Cancel packages - given a list of package ids, it'd _cancel_ all. ![Publishing queue design](https://user-images.githubusercontent.com/3928341/64194478-863fb000-ce7f-11e9-8421-a5426ff39f24.png)
1.0
[studio] Create APIs to support "Publishing status dashboard" - **In relation to #3352...** We need new APIs to be able to support #3352's requests. **Describe the solution you'd like** Required APIs would be - Get packages - retrieves the list of packages on the publishing queue. Each package should have at a minimum scheduled date, user who submitted, status, environment, comment and package id. It should support several filters namely a path expression, environments, and states. - Get package details - retrieves the list of files associated with the package by package id. If this is not expensive to do at once on the `get packages` call, please do fit it there and we wouldn't need this second call. - Cancel packages - given a list of package ids, it'd _cancel_ all. ![Publishing queue design](https://user-images.githubusercontent.com/3928341/64194478-863fb000-ce7f-11e9-8421-a5426ff39f24.png)
priority
create apis to support publishing status dashboard in relation to we need new apis to be able to support s requests describe the solution you d like required apis would be get packages retrieves the list of packages on the publishing queue each package should have at a minimum scheduled date user who submitted status environment comment and package id it should support several filters namely a path expression environments and states get package details retrieves the list of files associated with the package by package id if this is not expensive to do at once on the get packages call please do fit it there and we wouldn t need this second call cancel packages given a list of package ids it d cancel all
1
675,778
23,105,347,132
IssuesEvent
2022-07-27 08:19:53
owncloud/web
https://api.github.com/repos/owncloud/web
closed
Left Sidebar: Active navitem pretends to be not clickable
Type:Bug Priority:p3-medium
### Steps to reproduce 1. open ownCloud Web and hover over your active (blue) item in the left sidebar 2. Cursor-style is arrow (indicates "not clickable") ![Screenshot 000105](https://user-images.githubusercontent.com/26610733/181049112-d5b64af4-20c0-42cb-9f61-91b6e92d5a67.png) ### Expected behaviour Cursor style should indicate that its clickbale (Pointer) ![763px-Mouse-cursor-hand-pointer svg](https://user-images.githubusercontent.com/26610733/181050842-b54bf175-80b0-43ff-be6a-23ba35c6b4e2.png)
1.0
Left Sidebar: Active navitem pretends to be not clickable - ### Steps to reproduce 1. open ownCloud Web and hover over your active (blue) item in the left sidebar 2. Cursor-style is arrow (indicates "not clickable") ![Screenshot 000105](https://user-images.githubusercontent.com/26610733/181049112-d5b64af4-20c0-42cb-9f61-91b6e92d5a67.png) ### Expected behaviour Cursor style should indicate that its clickbale (Pointer) ![763px-Mouse-cursor-hand-pointer svg](https://user-images.githubusercontent.com/26610733/181050842-b54bf175-80b0-43ff-be6a-23ba35c6b4e2.png)
priority
left sidebar active navitem pretends to be not clickable steps to reproduce open owncloud web and hover over your active blue item in the left sidebar cursor style is arrow indicates not clickable expected behaviour cursor style should indicate that its clickbale pointer
1
346,969
10,422,298,123
IssuesEvent
2019-09-16 08:42:24
getkirby/kirby
https://api.github.com/repos/getkirby/kirby
closed
Switching pages causes select/checkbox/radio fields to briefly highlight as invalid
priority: medium 🔜 type: bug 🐛
### Describe the bug When switching pages in the panel (for example by using the left/right navigation to switch between sibling pages or by navigation to parent or child page) all fields of type radio/checkbox/select are highlighted as "invalid" just before the pages switches. So it's only there for a brief moment. Important to mention that this happens **although** the fields are not required and filled out. Also, when switching away and back to the page, the fields will still not show up as invalid – because they aren't. The "flashing" only appears in the split second before switching away. ### To Reproduce Steps to reproduce the behavior: 1. Add any field with the afflicted field types (select / checkbox / radio) to a page blueprint 2. Navigate away from the page, e.g. to a sibling page with same blueprint 3. See "flashing" of afflicted fields with "error" style (see screenshot above) ### Expected behavior Prior to Kirby 3.2.(3) switching pages did not result in the behavior described above. The fields should not validate as "invalid" if they aren't. ### Screenshots ![Bildschirmfoto 2019-08-16 um 13 48 27](https://user-images.githubusercontent.com/3918898/63165771-a91f3700-c02c-11e9-956d-644e02ccc7d7.png) ### Kirby Version 3.2.3
1.0
Switching pages causes select/checkbox/radio fields to briefly highlight as invalid - ### Describe the bug When switching pages in the panel (for example by using the left/right navigation to switch between sibling pages or by navigation to parent or child page) all fields of type radio/checkbox/select are highlighted as "invalid" just before the pages switches. So it's only there for a brief moment. Important to mention that this happens **although** the fields are not required and filled out. Also, when switching away and back to the page, the fields will still not show up as invalid – because they aren't. The "flashing" only appears in the split second before switching away. ### To Reproduce Steps to reproduce the behavior: 1. Add any field with the afflicted field types (select / checkbox / radio) to a page blueprint 2. Navigate away from the page, e.g. to a sibling page with same blueprint 3. See "flashing" of afflicted fields with "error" style (see screenshot above) ### Expected behavior Prior to Kirby 3.2.(3) switching pages did not result in the behavior described above. The fields should not validate as "invalid" if they aren't. ### Screenshots ![Bildschirmfoto 2019-08-16 um 13 48 27](https://user-images.githubusercontent.com/3918898/63165771-a91f3700-c02c-11e9-956d-644e02ccc7d7.png) ### Kirby Version 3.2.3
priority
switching pages causes select checkbox radio fields to briefly highlight as invalid describe the bug when switching pages in the panel for example by using the left right navigation to switch between sibling pages or by navigation to parent or child page all fields of type radio checkbox select are highlighted as invalid just before the pages switches so it s only there for a brief moment important to mention that this happens although the fields are not required and filled out also when switching away and back to the page the fields will still not show up as invalid – because they aren t the flashing only appears in the split second before switching away to reproduce steps to reproduce the behavior add any field with the afflicted field types select checkbox radio to a page blueprint navigate away from the page e g to a sibling page with same blueprint see flashing of afflicted fields with error style see screenshot above expected behavior prior to kirby switching pages did not result in the behavior described above the fields should not validate as invalid if they aren t screenshots kirby version
1
779,616
27,360,243,762
IssuesEvent
2023-02-27 15:25:02
orden-gg/fireball
https://api.github.com/repos/orden-gg/fireball
closed
Add prettier as dev dependency
enhancement priority: medium
It is dev only issue. We have missing `prettier` dependency in `package.json`.
1.0
Add prettier as dev dependency - It is dev only issue. We have missing `prettier` dependency in `package.json`.
priority
add prettier as dev dependency it is dev only issue we have missing prettier dependency in package json
1
117,022
4,710,115,384
IssuesEvent
2016-10-14 09:01:06
bespokeinteractive/mchapp
https://api.github.com/repos/bespokeinteractive/mchapp
closed
MCH Stores: User cannot clear search done for date i.e. from and to dates.
bug Medium Priority
User cannot clear search done for date -from and to dates. Suggestion: Add clear search button to allow clearing of search items within MCH store tabs.
1.0
MCH Stores: User cannot clear search done for date i.e. from and to dates. - User cannot clear search done for date -from and to dates. Suggestion: Add clear search button to allow clearing of search items within MCH store tabs.
priority
mch stores user cannot clear search done for date i e from and to dates user cannot clear search done for date from and to dates suggestion add clear search button to allow clearing of search items within mch store tabs
1
473,517
13,643,449,770
IssuesEvent
2020-09-25 17:08:38
LBL-EESA/TECA
https://api.github.com/repos/LBL-EESA/TECA
opened
app tests hard coded to use 2 mpi ranks
2_medium_priority
**Describe the bug** The app tests are hard coded to use 2 mpi ranks. For most of the test the mpi concurrency should be based on the cmake variable(s) TEST_CORES and/or HALF_TEST_CORES so that when there are more than 2 cores the tests run faster and MPI parallelism is stressed.
1.0
app tests hard coded to use 2 mpi ranks - **Describe the bug** The app tests are hard coded to use 2 mpi ranks. For most of the test the mpi concurrency should be based on the cmake variable(s) TEST_CORES and/or HALF_TEST_CORES so that when there are more than 2 cores the tests run faster and MPI parallelism is stressed.
priority
app tests hard coded to use mpi ranks describe the bug the app tests are hard coded to use mpi ranks for most of the test the mpi concurrency should be based on the cmake variable s test cores and or half test cores so that when there are more than cores the tests run faster and mpi parallelism is stressed
1
275,604
8,577,542,563
IssuesEvent
2018-11-13 00:30:09
minio/minio
https://api.github.com/repos/minio/minio
closed
Storage backend has reached its minimum free disk threshold
community priority: medium working as intended
<!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior I want to put object up to end of my free space. ## Current Behavior I try to put object with size about 10 Mib at the first time. But I cannot put the same object again. Minio throw an exception: ``` Error executing "PutObject" on "http://core.example.local:4567/prv0.example.local/mobile-app/packages/app_1.0.2.apk"; AWS HTTP error: Server error: `PUT http://core.example.local:4567/prv0.example.local/mobile-app/packages/app_1.0.2.apk` resulted in a `507 Insufficient Storage` response: <?xml version="1.0" encoding="UTF-8"?> <Error><Code>XMinioStorageFull</Code><Message>Storage backend has reached its min (truncated...) XMinioStorageFull (server): Storage backend has reached its minimum free disk threshold. Please delete a few objects to proceed. - <?xml version="1.0" encoding="UTF-8"?> <Error><Code>XMinioStorageFull</Code><Message>Storage backend has reached its minimum free disk threshold. Please delete a few objects to proceed.</Message><Key></Key><BucketName></BucketName><Resource>/prv0.example.local/mobile-app/packages/app_1.0.2.apk</Resource><RequestId>3L137</RequestId><HostId>3L137</HostId></Error> (File: /vagrant/vendor/aws/aws-sdk-php/src/WrappedHttpHandler.php, Line: 191): ``` It's only one file in only one bucket. I have only one bucket at all. Total size of all files is 10 Mib. ## Possible Solution I don't know. How to change threshold size of Minio? I have enough free space: ``` vagrant@example:/var/www/html$ df Filesystem 1K-blocks Used Available Use% Mounted on /dev/sda1 9620408 8127212 981452 90% / udev 10240 0 10240 0% /dev tmpfs 204876 4452 200424 3% /run tmpfs 512184 0 512184 0% /dev/shm tmpfs 5120 0 5120 0% /run/lock tmpfs 512184 0 512184 0% /sys/fs/cgroup 192.168.20.1:/Volumes/Project/example 28945984 7908088 21037896 28% /vagrant ``` ![image](https://user-images.githubusercontent.com/5145037/48294676-9416e580-e4a7-11e8-9b95-47cb4b7a2353.png) ## Steps to Reproduce (for bugs) ``` /** @var \Aws\S3\S3Client $s3 **/ $s3->putObject([ 'Key' => 'destination/file.apk', 'Bucket' => 'prv0.example.local', 'SourceFile' => '/path/to/file.apk', 'ContentType' => 'application/vnd.android.package-archive' ]); ``` ## Context Everything is general, nothing special. ## Regression I don't know, may be it was before. ## Your Environment * Version used (`minio version`): ``` Version: 2018-07-13T00:09:07Z Release-Tag: RELEASE.2018-07-13T00-09-07Z Commit-ID: 2aa18cafc674faf97201a973504cb4caa3a3e67c ``` * Environment name and version (e.g. nginx 1.9.1): ``` Vagrant 2.1.2 Nginx 1.6.2 PHP 7.1.10-1+0~20170929170631.9+jessie~1.gbp501135 (cli) (built: Sep 29 2017 17:33:58) ( NTS ) ``` * Server type and version: ``` Linux example 3.16.0-4-amd64 #1 SMP Debian 3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux Debian Jessie x64 ``` * Operating System and version (`uname -a`): ``` Darwin macbook.local 16.7.0 Darwin Kernel Version 16.7.0: Thu Jun 21 20:07:39 PDT 2018; root:xnu-3789.73.14~1/RELEASE_X86_64 x86_64 ``` * Link to your project: It's private project.
1.0
Storage backend has reached its minimum free disk threshold - <!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior I want to put object up to end of my free space. ## Current Behavior I try to put object with size about 10 Mib at the first time. But I cannot put the same object again. Minio throw an exception: ``` Error executing "PutObject" on "http://core.example.local:4567/prv0.example.local/mobile-app/packages/app_1.0.2.apk"; AWS HTTP error: Server error: `PUT http://core.example.local:4567/prv0.example.local/mobile-app/packages/app_1.0.2.apk` resulted in a `507 Insufficient Storage` response: <?xml version="1.0" encoding="UTF-8"?> <Error><Code>XMinioStorageFull</Code><Message>Storage backend has reached its min (truncated...) XMinioStorageFull (server): Storage backend has reached its minimum free disk threshold. Please delete a few objects to proceed. - <?xml version="1.0" encoding="UTF-8"?> <Error><Code>XMinioStorageFull</Code><Message>Storage backend has reached its minimum free disk threshold. Please delete a few objects to proceed.</Message><Key></Key><BucketName></BucketName><Resource>/prv0.example.local/mobile-app/packages/app_1.0.2.apk</Resource><RequestId>3L137</RequestId><HostId>3L137</HostId></Error> (File: /vagrant/vendor/aws/aws-sdk-php/src/WrappedHttpHandler.php, Line: 191): ``` It's only one file in only one bucket. I have only one bucket at all. Total size of all files is 10 Mib. ## Possible Solution I don't know. How to change threshold size of Minio? I have enough free space: ``` vagrant@example:/var/www/html$ df Filesystem 1K-blocks Used Available Use% Mounted on /dev/sda1 9620408 8127212 981452 90% / udev 10240 0 10240 0% /dev tmpfs 204876 4452 200424 3% /run tmpfs 512184 0 512184 0% /dev/shm tmpfs 5120 0 5120 0% /run/lock tmpfs 512184 0 512184 0% /sys/fs/cgroup 192.168.20.1:/Volumes/Project/example 28945984 7908088 21037896 28% /vagrant ``` ![image](https://user-images.githubusercontent.com/5145037/48294676-9416e580-e4a7-11e8-9b95-47cb4b7a2353.png) ## Steps to Reproduce (for bugs) ``` /** @var \Aws\S3\S3Client $s3 **/ $s3->putObject([ 'Key' => 'destination/file.apk', 'Bucket' => 'prv0.example.local', 'SourceFile' => '/path/to/file.apk', 'ContentType' => 'application/vnd.android.package-archive' ]); ``` ## Context Everything is general, nothing special. ## Regression I don't know, may be it was before. ## Your Environment * Version used (`minio version`): ``` Version: 2018-07-13T00:09:07Z Release-Tag: RELEASE.2018-07-13T00-09-07Z Commit-ID: 2aa18cafc674faf97201a973504cb4caa3a3e67c ``` * Environment name and version (e.g. nginx 1.9.1): ``` Vagrant 2.1.2 Nginx 1.6.2 PHP 7.1.10-1+0~20170929170631.9+jessie~1.gbp501135 (cli) (built: Sep 29 2017 17:33:58) ( NTS ) ``` * Server type and version: ``` Linux example 3.16.0-4-amd64 #1 SMP Debian 3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux Debian Jessie x64 ``` * Operating System and version (`uname -a`): ``` Darwin macbook.local 16.7.0 Darwin Kernel Version 16.7.0: Thu Jun 21 20:07:39 PDT 2018; root:xnu-3789.73.14~1/RELEASE_X86_64 x86_64 ``` * Link to your project: It's private project.
priority
storage backend has reached its minimum free disk threshold expected behavior i want to put object up to end of my free space current behavior i try to put object with size about mib at the first time but i cannot put the same object again minio throw an exception error executing putobject on aws http error server error put resulted in a insufficient storage response xminiostoragefull storage backend has reached its min truncated xminiostoragefull server storage backend has reached its minimum free disk threshold please delete a few objects to proceed xminiostoragefull storage backend has reached its minimum free disk threshold please delete a few objects to proceed example local mobile app packages app apk file vagrant vendor aws aws sdk php src wrappedhttphandler php line it s only one file in only one bucket i have only one bucket at all total size of all files is mib possible solution i don t know how to change threshold size of minio i have enough free space vagrant example var www html df filesystem blocks used available use mounted on dev udev dev tmpfs run tmpfs dev shm tmpfs run lock tmpfs sys fs cgroup volumes project example vagrant steps to reproduce for bugs var aws putobject key destination file apk bucket example local sourcefile path to file apk contenttype application vnd android package archive context everything is general nothing special regression i don t know may be it was before your environment version used minio version version release tag release commit id environment name and version e g nginx vagrant nginx php jessie cli built sep nts server type and version linux example smp debian gnu linux debian jessie operating system and version uname a darwin macbook local darwin kernel version thu jun pdt root xnu release link to your project it s private project
1
314,456
9,597,324,302
IssuesEvent
2019-05-09 20:58:44
minio/minio
https://api.github.com/repos/minio/minio
closed
[gateway] b2 - unable to create s3ql filesystem
community priority: medium working as intended
<!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> Expect mkfs.s3ql to create a filesystem. ## Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ``` 2019-05-06 11:49:34.210 14 INFO MainThread s3ql.mkfs.main: Generating random encryption key... 2019-05-06 11:49:34.215 14 DEBUG MainThread s3ql.backends.s3c.open_write: started with s3ql_passphrase 2019-05-06 11:49:34.217 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase 2019-05-06 11:49:34.218 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None 2019-05-06 11:49:34.219 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:36.330 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase 2019-05-06 11:49:36.330 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None 2019-05-06 11:49:36.335 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:37.484 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.02 Hz 2019-05-06 11:49:37.484 14 DEBUG MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f115226b22e7e6c2c_d20190506_m114935_c001_v0001039_t0027-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 1)... 2019-05-06 11:49:37.511 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase 2019-05-06 11:49:37.512 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None 2019-05-06 11:49:37.513 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:40.137 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase 2019-05-06 11:49:40.137 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None 2019-05-06 11:49:40.138 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:41.718 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.03 Hz 2019-05-06 11:49:41.718 14 DEBUG MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f1025162da09e4b6b_d20190506_m114938_c001_v0001117_t0052-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 2)... 2019-05-06 11:49:41.775 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase 2019-05-06 11:49:41.776 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None 2019-05-06 11:49:41.777 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:45.836 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase 2019-05-06 11:49:45.836 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None 2019-05-06 11:49:45.837 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:46.991 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.05 Hz 2019-05-06 11:49:46.992 14 INFO MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f116d93e2350333e3_d20190506_m114942_c001_v0001101_t0013-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 3)... ``` etc... ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> mkfs.s3ql works fine when minio is acting as a server. The B2 gateway seems to be broken. ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. minio gateway b2 2. mkfs.s3ql --backend-options no-ssl s3c://minio:9000/s3ql ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used (`minio version`): * Environment name and version (e.g. nginx 1.9.1): * Server type and version: * Operating System and version (`uname -a`): * Link to your project: minio/minio docker container: / # minio version Version: 2019-05-02T19:07:09Z Release-Tag: RELEASE.2019-05-02T19-07-09Z Commit-ID: 4b858b562a0887e10bfd0414dc87e68f1af31c3a s3ql docker container: kronostechnologies/s3ql ``` # dpkg -l s3ql Desired=Unknown/Install/Remove/Purge/Hold | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad) ||/ Name Version Architecture Description +++-==============-============-============-================================================= ii s3ql 3.0+dfsg-1 amd64 Full-featured file system for online data storage ```
1.0
[gateway] b2 - unable to create s3ql filesystem - <!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> Expect mkfs.s3ql to create a filesystem. ## Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ``` 2019-05-06 11:49:34.210 14 INFO MainThread s3ql.mkfs.main: Generating random encryption key... 2019-05-06 11:49:34.215 14 DEBUG MainThread s3ql.backends.s3c.open_write: started with s3ql_passphrase 2019-05-06 11:49:34.217 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase 2019-05-06 11:49:34.218 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None 2019-05-06 11:49:34.219 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:36.330 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase 2019-05-06 11:49:36.330 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None 2019-05-06 11:49:36.335 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:37.484 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.02 Hz 2019-05-06 11:49:37.484 14 DEBUG MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f115226b22e7e6c2c_d20190506_m114935_c001_v0001039_t0027-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 1)... 2019-05-06 11:49:37.511 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase 2019-05-06 11:49:37.512 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None 2019-05-06 11:49:37.513 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:40.137 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase 2019-05-06 11:49:40.137 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None 2019-05-06 11:49:40.138 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:41.718 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.03 Hz 2019-05-06 11:49:41.718 14 DEBUG MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f1025162da09e4b6b_d20190506_m114938_c001_v0001117_t0052-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 2)... 2019-05-06 11:49:41.775 14 DEBUG MainThread s3ql.backends.s3c.close: started with s3ql_passphrase 2019-05-06 11:49:41.776 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with PUT /s3ql_passphrase?None, qs=None 2019-05-06 11:49:41.777 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending PUT /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:45.836 14 DEBUG MainThread s3ql.backends.s3c.delete: started with s3ql_passphrase 2019-05-06 11:49:45.836 14 DEBUG MainThread s3ql.backends.s3c._do_request: started with DELETE /s3ql_passphrase?None, qs=None 2019-05-06 11:49:45.837 14 DEBUG MainThread s3ql.backends.s3c._send_request: sending DELETE /s3qlfs2/s3ql_passphrase 2019-05-06 11:49:46.991 14 DEBUG MainThread s3ql.backends.common.wrapped: Average retry rate: 0.05 Hz 2019-05-06 11:49:46.992 14 INFO MainThread s3ql.backends.common.wrapped: Encountered BadDigestError (BadDigest: MD5 mismatch for s3ql_passphrase (received: 4_z617e9c48e57980c364a50619_f116d93e2350333e3_d20190506_m114942_c001_v0001101_t0013-1, sent: 6381fec6215a58688e6be0259b9c1246)), retrying ObjectW.close (attempt 3)... ``` etc... ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> mkfs.s3ql works fine when minio is acting as a server. The B2 gateway seems to be broken. ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. minio gateway b2 2. mkfs.s3ql --backend-options no-ssl s3c://minio:9000/s3ql ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used (`minio version`): * Environment name and version (e.g. nginx 1.9.1): * Server type and version: * Operating System and version (`uname -a`): * Link to your project: minio/minio docker container: / # minio version Version: 2019-05-02T19:07:09Z Release-Tag: RELEASE.2019-05-02T19-07-09Z Commit-ID: 4b858b562a0887e10bfd0414dc87e68f1af31c3a s3ql docker container: kronostechnologies/s3ql ``` # dpkg -l s3ql Desired=Unknown/Install/Remove/Purge/Hold | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad) ||/ Name Version Architecture Description +++-==============-============-============-================================================= ii s3ql 3.0+dfsg-1 amd64 Full-featured file system for online data storage ```
priority
unable to create filesystem expected behavior expect mkfs to create a filesystem current behavior info mainthread mkfs main generating random encryption key debug mainthread backends open write started with passphrase debug mainthread backends close started with passphrase debug mainthread backends do request started with put passphrase none qs none debug mainthread backends send request sending put passphrase debug mainthread backends delete started with passphrase debug mainthread backends do request started with delete passphrase none qs none debug mainthread backends send request sending delete passphrase debug mainthread backends common wrapped average retry rate hz debug mainthread backends common wrapped encountered baddigesterror baddigest mismatch for passphrase received sent retrying objectw close attempt debug mainthread backends close started with passphrase debug mainthread backends do request started with put passphrase none qs none debug mainthread backends send request sending put passphrase debug mainthread backends delete started with passphrase debug mainthread backends do request started with delete passphrase none qs none debug mainthread backends send request sending delete passphrase debug mainthread backends common wrapped average retry rate hz debug mainthread backends common wrapped encountered baddigesterror baddigest mismatch for passphrase received sent retrying objectw close attempt debug mainthread backends close started with passphrase debug mainthread backends do request started with put passphrase none qs none debug mainthread backends send request sending put passphrase debug mainthread backends delete started with passphrase debug mainthread backends do request started with delete passphrase none qs none debug mainthread backends send request sending delete passphrase debug mainthread backends common wrapped average retry rate hz info mainthread backends common wrapped encountered baddigesterror baddigest mismatch for passphrase received sent retrying objectw close attempt etc possible solution mkfs works fine when minio is acting as a server the gateway seems to be broken steps to reproduce for bugs minio gateway mkfs backend options no ssl minio your environment version used minio version environment name and version e g nginx server type and version operating system and version uname a link to your project minio minio docker container minio version version release tag release commit id docker container kronostechnologies dpkg l desired unknown install remove purge hold status not inst conf files unpacked half conf half inst trig await trig pend err none reinst required status err uppercase bad name version architecture description ii dfsg full featured file system for online data storage
1
88,646
3,783,485,984
IssuesEvent
2016-03-19 05:27:56
parishod/viewer-api
https://api.github.com/repos/parishod/viewer-api
closed
Copy to clipboard feature for social sharing modal.
Priority: Medium Status: In Progress Status: Review Needed Type: Enhancement
Copy to clipboard button should be added along with the already existing social buttons. Clicking on copy to clipboard button should copy to clipboard if possible and should also display text box filled with url and pre selected.
1.0
Copy to clipboard feature for social sharing modal. - Copy to clipboard button should be added along with the already existing social buttons. Clicking on copy to clipboard button should copy to clipboard if possible and should also display text box filled with url and pre selected.
priority
copy to clipboard feature for social sharing modal copy to clipboard button should be added along with the already existing social buttons clicking on copy to clipboard button should copy to clipboard if possible and should also display text box filled with url and pre selected
1
709,597
24,384,255,451
IssuesEvent
2022-10-04 10:22:17
trustwallet/wallet-core
https://api.github.com/repos/trustwallet/wallet-core
opened
Support custom EVM chain without code change, but runtime data
enhancement priority:low size:medium improvements
**Is your feature request related to a problem? Please describe.** Support for a new EVM-clone chain has been simplified (see #2214), but still requires code change and new release in wallet-core. It would be nice if new custom chain can be added to a wallet dynamically at runtime, but still some support from wallet core. **Describe the solution you'd like** Two possible solutions in the wallet app: 1. A custom EVM-clone chain can be added by the user, and then app uses Ethereum coin for address derivation, and Ethereum coin and custom chainId for signing. No involvement from wallet core. 2. A custom EVM-clone chain can be added by the user. App 'registers' the new chain properties with wallet code. Wallet core extends its relevant coin-specific methods to also support the dynamically registered chains (additionally to the ones statically defined in `registry.json`). **Limitation** Since dynamically registered types are not known at build time, generated sources will not contain these coins. **Describe alternatives you've considered** See above number 1). **Checklist** <!--- Group checklist per issue needed, one specific feature of your goal --> <!--- Each big task can have subtask, doesn't hesitate to split into small pull request to simplify the review process --> - [ ] task number 1 - [ ] subtask number 1 - [ ] subtask number 2 - [ ] task number 2 - [ ] task number 3 **Resources** <!--- Link resources this way: [My Resource Title](link) -->
1.0
Support custom EVM chain without code change, but runtime data - **Is your feature request related to a problem? Please describe.** Support for a new EVM-clone chain has been simplified (see #2214), but still requires code change and new release in wallet-core. It would be nice if new custom chain can be added to a wallet dynamically at runtime, but still some support from wallet core. **Describe the solution you'd like** Two possible solutions in the wallet app: 1. A custom EVM-clone chain can be added by the user, and then app uses Ethereum coin for address derivation, and Ethereum coin and custom chainId for signing. No involvement from wallet core. 2. A custom EVM-clone chain can be added by the user. App 'registers' the new chain properties with wallet code. Wallet core extends its relevant coin-specific methods to also support the dynamically registered chains (additionally to the ones statically defined in `registry.json`). **Limitation** Since dynamically registered types are not known at build time, generated sources will not contain these coins. **Describe alternatives you've considered** See above number 1). **Checklist** <!--- Group checklist per issue needed, one specific feature of your goal --> <!--- Each big task can have subtask, doesn't hesitate to split into small pull request to simplify the review process --> - [ ] task number 1 - [ ] subtask number 1 - [ ] subtask number 2 - [ ] task number 2 - [ ] task number 3 **Resources** <!--- Link resources this way: [My Resource Title](link) -->
priority
support custom evm chain without code change but runtime data is your feature request related to a problem please describe support for a new evm clone chain has been simplified see but still requires code change and new release in wallet core it would be nice if new custom chain can be added to a wallet dynamically at runtime but still some support from wallet core describe the solution you d like two possible solutions in the wallet app a custom evm clone chain can be added by the user and then app uses ethereum coin for address derivation and ethereum coin and custom chainid for signing no involvement from wallet core a custom evm clone chain can be added by the user app registers the new chain properties with wallet code wallet core extends its relevant coin specific methods to also support the dynamically registered chains additionally to the ones statically defined in registry json limitation since dynamically registered types are not known at build time generated sources will not contain these coins describe alternatives you ve considered see above number checklist task number subtask number subtask number task number task number resources
1
568,990
16,992,239,909
IssuesEvent
2021-06-30 22:29:12
nexB/scancode.io
https://api.github.com/repos/nexB/scancode.io
closed
meta-issue: design-needed: Finding and acting on issues and exceptions to review in scan
enhancement medium priority
Once a scan is completed, I want to be able to efficiently review "issues and exception" in my scan such as: 1. files or packages under a certain license (e.g. possibly in relation with policies #90 ) 2. inconclusive detections (such as unknown licenses and similar) and possibly a few more. It could take the form of filtered lists or some dashboard that I could "zoom" to and needs design. The outcome is to identify a number of related files or packages that share some origin and license exception, issues or policy problem. Once I "zoom in" a problem, I would to be able to act on this such as: 1. changing or setting some status that would state that these resources or packages have been reviewed 2. changing or setting and updated license and or copyright or origin 3. mark a bunch files with some detected clues are having non-relevant clues (say for instance because of an http://example.com URLs) 4. tag files or packages for future extra review or action (something like a TODO list) and possibly a few more.
1.0
meta-issue: design-needed: Finding and acting on issues and exceptions to review in scan - Once a scan is completed, I want to be able to efficiently review "issues and exception" in my scan such as: 1. files or packages under a certain license (e.g. possibly in relation with policies #90 ) 2. inconclusive detections (such as unknown licenses and similar) and possibly a few more. It could take the form of filtered lists or some dashboard that I could "zoom" to and needs design. The outcome is to identify a number of related files or packages that share some origin and license exception, issues or policy problem. Once I "zoom in" a problem, I would to be able to act on this such as: 1. changing or setting some status that would state that these resources or packages have been reviewed 2. changing or setting and updated license and or copyright or origin 3. mark a bunch files with some detected clues are having non-relevant clues (say for instance because of an http://example.com URLs) 4. tag files or packages for future extra review or action (something like a TODO list) and possibly a few more.
priority
meta issue design needed finding and acting on issues and exceptions to review in scan once a scan is completed i want to be able to efficiently review issues and exception in my scan such as files or packages under a certain license e g possibly in relation with policies inconclusive detections such as unknown licenses and similar and possibly a few more it could take the form of filtered lists or some dashboard that i could zoom to and needs design the outcome is to identify a number of related files or packages that share some origin and license exception issues or policy problem once i zoom in a problem i would to be able to act on this such as changing or setting some status that would state that these resources or packages have been reviewed changing or setting and updated license and or copyright or origin mark a bunch files with some detected clues are having non relevant clues say for instance because of an urls tag files or packages for future extra review or action something like a todo list and possibly a few more
1
26,280
2,684,288,338
IssuesEvent
2015-03-28 20:51:56
ConEmu/old-issues
https://api.github.com/repos/ConEmu/old-issues
opened
Auto update nightmare
2–5 stars bug imported Priority-Medium
_From [sspirido...@gmail.com](https://code.google.com/u/106209959253029514009/) on February 08, 2013 15:39:50_ Required information! OS version: Win8 x64 ConEmu version: 2013.02.05 and all previous Far version (if you are using Far Manager): Far Manager, version 3.0 (build 3011) x64 *Bug description* Mess on auto update ConEmu setup with automatically update on dev. releases. 1. ComEmu (with FAR) ask to upload new release - Yes 2. ComEmu (with FAR) ask to update - Yes 3. New console has been opened 4. ConEmu (with new console) ask to upload new release - No (Grr..) 5. Installator reports about open FAR and console (Grrr...) 6. If select close it can't close new console (Grrrr....) 7. If select do not close and reboot after the installator do something 8. Installator ask to reboot - No, I d'like to do it later. 9. Installator crash with some error (Grrrr!!!). 10. If I run the FAR (with conEmu) after that quest WITHOUT reboot it will updated and works correctly on the first look. I am lost. *Steps to reproduction* 1. See above _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=930_
1.0
Auto update nightmare - _From [sspirido...@gmail.com](https://code.google.com/u/106209959253029514009/) on February 08, 2013 15:39:50_ Required information! OS version: Win8 x64 ConEmu version: 2013.02.05 and all previous Far version (if you are using Far Manager): Far Manager, version 3.0 (build 3011) x64 *Bug description* Mess on auto update ConEmu setup with automatically update on dev. releases. 1. ComEmu (with FAR) ask to upload new release - Yes 2. ComEmu (with FAR) ask to update - Yes 3. New console has been opened 4. ConEmu (with new console) ask to upload new release - No (Grr..) 5. Installator reports about open FAR and console (Grrr...) 6. If select close it can't close new console (Grrrr....) 7. If select do not close and reboot after the installator do something 8. Installator ask to reboot - No, I d'like to do it later. 9. Installator crash with some error (Grrrr!!!). 10. If I run the FAR (with conEmu) after that quest WITHOUT reboot it will updated and works correctly on the first look. I am lost. *Steps to reproduction* 1. See above _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=930_
priority
auto update nightmare from on february required information os version conemu version and all previous far version if you are using far manager far manager version build bug description mess on auto update conemu setup with automatically update on dev releases comemu with far ask to upload new release yes comemu with far ask to update yes new console has been opened conemu with new console ask to upload new release no grr installator reports about open far and console grrr if select close it can t close new console grrrr if select do not close and reboot after the installator do something installator ask to reboot no i d like to do it later installator crash with some error grrrr if i run the far with conemu after that quest without reboot it will updated and works correctly on the first look i am lost steps to reproduction see above original issue
1
208,927
7,163,083,995
IssuesEvent
2018-01-29 05:08:47
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[6.4] Console is not opening until server comes online
Medium Priority
John asked me to post this as apparently it is not the behavior wanted: The Console does not open up on desktop until after the world generation/server load is compete and the port opens to make game available. Once the game is online the console will also load but it sort of removed a major tool for system admins to know what the server is doing/what state it is in other then just by watching CPU behavior in the task manager, especially at this time when we are having many different issues trying to get the servers to load.
1.0
[6.4] Console is not opening until server comes online - John asked me to post this as apparently it is not the behavior wanted: The Console does not open up on desktop until after the world generation/server load is compete and the port opens to make game available. Once the game is online the console will also load but it sort of removed a major tool for system admins to know what the server is doing/what state it is in other then just by watching CPU behavior in the task manager, especially at this time when we are having many different issues trying to get the servers to load.
priority
console is not opening until server comes online john asked me to post this as apparently it is not the behavior wanted the console does not open up on desktop until after the world generation server load is compete and the port opens to make game available once the game is online the console will also load but it sort of removed a major tool for system admins to know what the server is doing what state it is in other then just by watching cpu behavior in the task manager especially at this time when we are having many different issues trying to get the servers to load
1
609,691
18,884,930,333
IssuesEvent
2021-11-15 06:18:22
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
closed
Message Improvement
feature: enhancement priority: medium
Some code refactoring required to fix some performance-related core logic - With the Message thread ajax, We sending all recipients lists for every thread. it'll be a little heavy with a large network We should send recipients with some limit like 10 and we also need to send total recipients with thread objects. - We need to provide separate ajax to get message thread recipients so we can support pagination. - One major refactoring we can do is, we can move our message logic based on Rest endpoint instead of ajax, by this change we no need to maintain two chunks of code. ----- Customer Query for the Reference Slow SQL queries from group message causing the private messages and message notification to load so slow and sometimes causing 500 internal error, it happens when there are thousands of group members in a group message thread. This is the slow SQL query reported by client: https://prnt.sc/wiqzai -------- https://secure.helpscout.net/conversation/1470954591/134010 Test Group and with only one member the Group message had no delay and acted as one would expect. some other Group - with over 49 members it has this massive delay. Related HS Tickets: https://secure.helpscout.net/conversation/1371027857/115494/ https://secure.helpscout.net/conversation/1388206152/118467/ https://secure.helpscout.net/conversation/1470954591/134010/ https://secure.helpscout.net/conversation/1567963949/152658/
1.0
Message Improvement - Some code refactoring required to fix some performance-related core logic - With the Message thread ajax, We sending all recipients lists for every thread. it'll be a little heavy with a large network We should send recipients with some limit like 10 and we also need to send total recipients with thread objects. - We need to provide separate ajax to get message thread recipients so we can support pagination. - One major refactoring we can do is, we can move our message logic based on Rest endpoint instead of ajax, by this change we no need to maintain two chunks of code. ----- Customer Query for the Reference Slow SQL queries from group message causing the private messages and message notification to load so slow and sometimes causing 500 internal error, it happens when there are thousands of group members in a group message thread. This is the slow SQL query reported by client: https://prnt.sc/wiqzai -------- https://secure.helpscout.net/conversation/1470954591/134010 Test Group and with only one member the Group message had no delay and acted as one would expect. some other Group - with over 49 members it has this massive delay. Related HS Tickets: https://secure.helpscout.net/conversation/1371027857/115494/ https://secure.helpscout.net/conversation/1388206152/118467/ https://secure.helpscout.net/conversation/1470954591/134010/ https://secure.helpscout.net/conversation/1567963949/152658/
priority
message improvement some code refactoring required to fix some performance related core logic with the message thread ajax we sending all recipients lists for every thread it ll be a little heavy with a large network we should send recipients with some limit like and we also need to send total recipients with thread objects we need to provide separate ajax to get message thread recipients so we can support pagination one major refactoring we can do is we can move our message logic based on rest endpoint instead of ajax by this change we no need to maintain two chunks of code customer query for the reference slow sql queries from group message causing the private messages and message notification to load so slow and sometimes causing internal error it happens when there are thousands of group members in a group message thread this is the slow sql query reported by client test group and with only one member the group message had no delay and acted as one would expect some other group with over members it has this massive delay related hs tickets
1