Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
785,248 | 27,606,014,823 | IssuesEvent | 2023-03-09 13:03:20 | tahowallet/extension | https://api.github.com/repos/tahowallet/extension | closed | Insufficient funds warning is not disappearing | Type: Bug Priority: Medium Status: Available | ### Discord Discussion Link
_No response_
### What browsers are you seeing the problem on?
Chrome
### What were you trying to do?
https://user-images.githubusercontent.com/20949277/214549920-df406d09-2cf4-4225-b591-d18d6cffa095.mov
### What did not work?
Insufficient funds warning is not disappearing even after gas settings are changed and there is enough base asset to pay for gas.
### Version
v0.18.9
### Relevant log output
_No response_ | 1.0 | Insufficient funds warning is not disappearing - ### Discord Discussion Link
_No response_
### What browsers are you seeing the problem on?
Chrome
### What were you trying to do?
https://user-images.githubusercontent.com/20949277/214549920-df406d09-2cf4-4225-b591-d18d6cffa095.mov
### What did not work?
Insufficient funds warning is not disappearing even after gas settings are changed and there is enough base asset to pay for gas.
### Version
v0.18.9
### Relevant log output
_No response_ | priority | insufficient funds warning is not disappearing discord discussion link no response what browsers are you seeing the problem on chrome what were you trying to do what did not work insufficient funds warning is not disappearing even after gas settings are changed and there is enough base asset to pay for gas version relevant log output no response | 1 |
16,586 | 2,615,119,397 | IssuesEvent | 2015-03-01 05:45:20 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | HttpRequest.executeAsync | auto-migrated Component-HTTP Milestone-Version1.5.0 Priority-Medium Type-Enhancement | ```
External references, such as a standards document, or specification?
N/A
Java environments (e.g. Java 6, Android 2.3, App Engine 1.4.2, or All)?
All
Please describe the feature requested.
It would be nice to provide an executeAsync method in HttpRequest that started
a new thread to execute the request and when it got the final response would
call an implementation of an HttpAsyncCallback interface similar to the one in
GWT:
http://google-web-toolkit.googlecode.com/svn/javadoc/2.2/com/google/gwt/user/cli
ent/rpc/AsyncCallback.html
Sample usage:
public static void run(HttpRequest request) {
request.executeAsync(new HttpAsyncCallback() {
public void onSuccess(HttpResponse response) {
if (response.isSuccessStatusCode) {
// server sent a success response
} else {
// server sent an error response
}
}
public void onFailure(Throwable caught) {
// Convenient way to find out which exception was thrown.
try {
throw caught;
} catch (IOException e) {
// some networking problem?
} catch (Throwable e) {
// last resort -- a very unexpected exception
}
}
});
}
```
Original issue reported on code.google.com by `yan...@google.com` on 19 Mar 2011 at 9:07 | 1.0 | HttpRequest.executeAsync - ```
External references, such as a standards document, or specification?
N/A
Java environments (e.g. Java 6, Android 2.3, App Engine 1.4.2, or All)?
All
Please describe the feature requested.
It would be nice to provide an executeAsync method in HttpRequest that started
a new thread to execute the request and when it got the final response would
call an implementation of an HttpAsyncCallback interface similar to the one in
GWT:
http://google-web-toolkit.googlecode.com/svn/javadoc/2.2/com/google/gwt/user/cli
ent/rpc/AsyncCallback.html
Sample usage:
public static void run(HttpRequest request) {
request.executeAsync(new HttpAsyncCallback() {
public void onSuccess(HttpResponse response) {
if (response.isSuccessStatusCode) {
// server sent a success response
} else {
// server sent an error response
}
}
public void onFailure(Throwable caught) {
// Convenient way to find out which exception was thrown.
try {
throw caught;
} catch (IOException e) {
// some networking problem?
} catch (Throwable e) {
// last resort -- a very unexpected exception
}
}
});
}
```
Original issue reported on code.google.com by `yan...@google.com` on 19 Mar 2011 at 9:07 | priority | httprequest executeasync external references such as a standards document or specification n a java environments e g java android app engine or all all please describe the feature requested it would be nice to provide an executeasync method in httprequest that started a new thread to execute the request and when it got the final response would call an implementation of an httpasynccallback interface similar to the one in gwt ent rpc asynccallback html sample usage public static void run httprequest request request executeasync new httpasynccallback public void onsuccess httpresponse response if response issuccessstatuscode server sent a success response else server sent an error response public void onfailure throwable caught convenient way to find out which exception was thrown try throw caught catch ioexception e some networking problem catch throwable e last resort a very unexpected exception original issue reported on code google com by yan google com on mar at | 1 |
283,504 | 8,719,731,675 | IssuesEvent | 2018-12-08 03:45:21 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | build_visit2_7_0 fails building python with icc. | bug likelihood medium priority reviewed severity high | Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line:
env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=I/usr/local/tools/mvapichgnu/include ./build_visit2_7_0 \
console nothirdparty novisit \
mesa cmake python vtk qt icet pyside \
hdf5 silo szip \
makeflags -j4
Here is the Python failure:
icc pthread fnostrictaliasing g O2 DNDEBUG m64 fPIC O2 -fPIC
Ibuild/temp.linuxx86_64-2.7/libffi/include
Ibuild/temp.linuxx86_64-2.7/libffi
I/nfs/tmp2/brugger/build_python/Python2.7.5/Modules/_ctypes/libffi/src
I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linuxx86_64_icc/include
I. IInclude I./Include I/usr/local/include
I/nfs/tmp2/brugger/build_python/Python2.7.5/Include
I/nfs/tmp2/brugger/build_python/Python2.7.5 -c
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
-o
build/temp.linuxx86_642.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o
Wall fexceptions
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56):
error: identifier "__m128" is undefined
UINT128 i128;
^
compilation aborted for
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
(code 2)
Python build finished, but the necessary bits to build these modules were not
found:
bsddb185 dl imageop
sunaudiodev
To find the necessary bits, look in setup.py in detect_modules() for the
module's name.
Failed to build these modules:
_ctypes
running build_scripts
running install_lib
Here is the Qt failure:
icpc c m64 fPIC O2 O2 falignfunctions=16 ansialias fstrict-aliasing
fPIC fnojumptables w1 Wcheck -wd654,1572,411,873,1125,2259
DQT_BOOTSTRAPPED DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII
DQT_NO_CAST_TO_ASCII DQT_NO_CODECS DQT_NO_DATASTREAM DQT_NO_GEOM_VARIANT
DQT_NO_LIBRARY DQT_NO_QOBJECT DQT_NO_STL DQT_NO_SYSTEMLOCALE
DQT_NO_TEXTSTREAM DQT_NO_THREAD -DQT_NO_UNICODETABLES
DQT_NO_USING_NAMESPACE DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE
D_LARGEFILE_SOURCE I../../../mkspecs/linuxicc64 I. I../../../include
I../../../include/QtCore I../../../include/QtXml -o
.obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp
rm -f libbootstrap.a
xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o
.obj/releasestatic/qlatincodec.o .obj/releasestatic/qsimplecodec.o
.obj/releasestatic/qtextcodec.o .obj/releasestatic/qtsciicodec.o
.obj/releasestatic/qutfcodec.o .obj/releasestatic/qglobal.o
.obj/releasestatic/qmalloc.o .obj/releasestatic/qnumeric.o
.obj/releasestatic/qabstractfileengine.o .obj/releasestatic/qbuffer.o
.obj/releasestatic/qdatastream.o .obj/releasestatic/qdir.o
.obj/releasestatic/qdiriterator.o .obj/releasestatic/qfile.o
.obj/releasestatic/qfileinfo.o .obj/releasestatic/qfilesystementry.o
.obj/releasestatic/qfilesystemengine.o .obj/releasestatic/qfsfileengine.o
.obj/releasestatic/qfsfileengine_iterator.o .obj/releasestatic/qiodevice.o
.obj/releasestatic/qtemporaryfile.o .obj/releasestatic/qtextstream.o
.obj/releasestatic/qmetatype.o .obj/releasestatic/qvariant.o
.obj/releasestatic/qsystemerror.o .obj/releasestatic/qbitarray.o
.obj/releasestatic/qbytearray.o .obj/releasestatic/qbytearraymatcher.o
.obj/releasestatic/qdatetime.o .obj/releasestatic/qhash.o
.obj/releasestatic/qlist.o .obj/releasestatic/qlocale.o
.obj/releasestatic/qlocale_tools.o .obj/releasestatic/qmap.o
.obj/releasestatic/qregexp.o .obj/releasestatic/qstring.o
.obj/releasestatic/qstringlist.o .obj/releasestatic/qvector.o
.obj/releasestatic/qvsnprintf.o .obj/releasestatic/qxmlutils.o
.obj/releasestatic/qxmlstream.o .obj/releasestatic/qdom.o
.obj/releasestatic/qxml.o .obj/releasestatic/qfilesystemengine_unix.o
.obj/release-static/qfilesystemiterator_unix.o
.obj/releasestatic/qfsfileengine_unix.o .obj/releasestatic/qlocale_unix.o
make[1]: xiar: Command not found
make[1]: ***** [libbootstrap.a] Error 127
make[1]: Leaving directory
`/nfs/tmp2/brugger/build_python/qteverywhereopensourcesrc4.8.3/src/tools/bootstrap'
make: ***** [subtoolsbootstrapmake_defaultordered] Error 2
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1651
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: build_visit2_7_0 fails building python with icc.
Assigned to: Kathleen Biagas
Category:
Target version: 2.7.1
Author: Eric Brugger
Start: 11/08/2013
Due date:
% Done: 0
Estimated time:
Created: 11/08/2013 05:32 pm
Updated: 12/11/2013 05:33 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: 2.7.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line:
env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=-I/usr/local/tools/mvapich-gnu/include ./build_visit2_7_0 \
--console --no-thirdparty --no-visit \
--mesa --cmake --python --vtk --qt --icet --pyside \
--hdf5 --silo --szip \
--makeflags -j4
Here is the Python failure:
icc -pthread -fno-strict-aliasing -g -O2 -DNDEBUG -m64 -fPIC -O2 -fPIC
-Ibuild/temp.linux-x86_64-2.7/libffi/include
-Ibuild/temp.linux-x86_64-2.7/libffi
-I/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src
-I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linux-x86_64_icc/include
-I. -IInclude -I./Include -I/usr/local/include
-I/nfs/tmp2/brugger/build_python/Python-2.7.5/Include
-I/nfs/tmp2/brugger/build_python/Python-2.7.5 -c
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
-o
build/temp.linux-x86_64-2.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o
-Wall -fexceptions
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56):
error: identifier "__m128" is undefined
UINT128 i128;
^
compilation aborted for
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
(code 2)
Python build finished, but the necessary bits to build these modules were not
found:
bsddb185 dl imageop
sunaudiodev
To find the necessary bits, look in setup.py in detect_modules() for the
module's name.
Failed to build these modules:
_ctypes
running build_scripts
running install_lib
Here is the Qt failure:
icpc -c -m64 -fPIC -O2 -O2 -falign-functions=16 -ansi-alias -fstrict-aliasing
-fPIC -fno-jump-tables -w1 -Wcheck -wd654,1572,411,873,1125,2259
-DQT_BOOTSTRAPPED -DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII
-DQT_NO_CAST_TO_ASCII -DQT_NO_CODECS -DQT_NO_DATASTREAM -DQT_NO_GEOM_VARIANT
-DQT_NO_LIBRARY -DQT_NO_QOBJECT -DQT_NO_STL -DQT_NO_SYSTEMLOCALE
-DQT_NO_TEXTSTREAM -DQT_NO_THREAD -DQT_NO_UNICODETABLES
-DQT_NO_USING_NAMESPACE -DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE
-D_LARGEFILE_SOURCE -I../../../mkspecs/linux-icc-64 -I. -I../../../include
-I../../../include/QtCore -I../../../include/QtXml -o
.obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp
rm -f libbootstrap.a
xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o
.obj/release-static/qlatincodec.o .obj/release-static/qsimplecodec.o
.obj/release-static/qtextcodec.o .obj/release-static/qtsciicodec.o
.obj/release-static/qutfcodec.o .obj/release-static/qglobal.o
.obj/release-static/qmalloc.o .obj/release-static/qnumeric.o
.obj/release-static/qabstractfileengine.o .obj/release-static/qbuffer.o
.obj/release-static/qdatastream.o .obj/release-static/qdir.o
.obj/release-static/qdiriterator.o .obj/release-static/qfile.o
.obj/release-static/qfileinfo.o .obj/release-static/qfilesystementry.o
.obj/release-static/qfilesystemengine.o .obj/release-static/qfsfileengine.o
.obj/release-static/qfsfileengine_iterator.o .obj/release-static/qiodevice.o
.obj/release-static/qtemporaryfile.o .obj/release-static/qtextstream.o
.obj/release-static/qmetatype.o .obj/release-static/qvariant.o
.obj/release-static/qsystemerror.o .obj/release-static/qbitarray.o
.obj/release-static/qbytearray.o .obj/release-static/qbytearraymatcher.o
.obj/release-static/qdatetime.o .obj/release-static/qhash.o
.obj/release-static/qlist.o .obj/release-static/qlocale.o
.obj/release-static/qlocale_tools.o .obj/release-static/qmap.o
.obj/release-static/qregexp.o .obj/release-static/qstring.o
.obj/release-static/qstringlist.o .obj/release-static/qvector.o
.obj/release-static/qvsnprintf.o .obj/release-static/qxmlutils.o
.obj/release-static/qxmlstream.o .obj/release-static/qdom.o
.obj/release-static/qxml.o .obj/release-static/qfilesystemengine_unix.o
.obj/release-static/qfilesystemiterator_unix.o
.obj/release-static/qfsfileengine_unix.o .obj/release-static/qlocale_unix.o
make[1]: xiar: Command not found
make[1]: *** [libbootstrap.a] Error 127
make[1]: Leaving directory
`/nfs/tmp2/brugger/build_python/qt-everywhere-opensource-src-4.8.3/src/tools/bootstrap'
make: *** [sub-tools-bootstrap-make_default-ordered] Error 2
Comments:
Hi Eric,The Python error may not be critical (at least for now)..Looks like Qt is configuring the AR command to be "xiar"mkspecs/linux-icc/qmake.conf:QMAKE_AR = xiar cqsA quick fix might be to set a simple "alias xiar=ar" or modify the qmake.conf file to what ever might be appropriate call to create archives for your system.Best regards,Hari
The QT build issue Eric encountered is due to the LC setup. We need to issue a 'use icc' before attempting to build, as doing so makes available the full suite of intel compiler tools (including xiar used by QT).
2.7RC update 22380, merge-to-trunk 22382Added minor python patch to bv_python, fixing ffi64 issue when compiling with icc. The patch is from a newer version of ffi64 library.
| 1.0 | build_visit2_7_0 fails building python with icc. - Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line:
env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=I/usr/local/tools/mvapichgnu/include ./build_visit2_7_0 \
console nothirdparty novisit \
mesa cmake python vtk qt icet pyside \
hdf5 silo szip \
makeflags -j4
Here is the Python failure:
icc pthread fnostrictaliasing g O2 DNDEBUG m64 fPIC O2 -fPIC
Ibuild/temp.linuxx86_64-2.7/libffi/include
Ibuild/temp.linuxx86_64-2.7/libffi
I/nfs/tmp2/brugger/build_python/Python2.7.5/Modules/_ctypes/libffi/src
I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linuxx86_64_icc/include
I. IInclude I./Include I/usr/local/include
I/nfs/tmp2/brugger/build_python/Python2.7.5/Include
I/nfs/tmp2/brugger/build_python/Python2.7.5 -c
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
-o
build/temp.linuxx86_642.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o
Wall fexceptions
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56):
error: identifier "__m128" is undefined
UINT128 i128;
^
compilation aborted for
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
(code 2)
Python build finished, but the necessary bits to build these modules were not
found:
bsddb185 dl imageop
sunaudiodev
To find the necessary bits, look in setup.py in detect_modules() for the
module's name.
Failed to build these modules:
_ctypes
running build_scripts
running install_lib
Here is the Qt failure:
icpc c m64 fPIC O2 O2 falignfunctions=16 ansialias fstrict-aliasing
fPIC fnojumptables w1 Wcheck -wd654,1572,411,873,1125,2259
DQT_BOOTSTRAPPED DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII
DQT_NO_CAST_TO_ASCII DQT_NO_CODECS DQT_NO_DATASTREAM DQT_NO_GEOM_VARIANT
DQT_NO_LIBRARY DQT_NO_QOBJECT DQT_NO_STL DQT_NO_SYSTEMLOCALE
DQT_NO_TEXTSTREAM DQT_NO_THREAD -DQT_NO_UNICODETABLES
DQT_NO_USING_NAMESPACE DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE
D_LARGEFILE_SOURCE I../../../mkspecs/linuxicc64 I. I../../../include
I../../../include/QtCore I../../../include/QtXml -o
.obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp
rm -f libbootstrap.a
xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o
.obj/releasestatic/qlatincodec.o .obj/releasestatic/qsimplecodec.o
.obj/releasestatic/qtextcodec.o .obj/releasestatic/qtsciicodec.o
.obj/releasestatic/qutfcodec.o .obj/releasestatic/qglobal.o
.obj/releasestatic/qmalloc.o .obj/releasestatic/qnumeric.o
.obj/releasestatic/qabstractfileengine.o .obj/releasestatic/qbuffer.o
.obj/releasestatic/qdatastream.o .obj/releasestatic/qdir.o
.obj/releasestatic/qdiriterator.o .obj/releasestatic/qfile.o
.obj/releasestatic/qfileinfo.o .obj/releasestatic/qfilesystementry.o
.obj/releasestatic/qfilesystemengine.o .obj/releasestatic/qfsfileengine.o
.obj/releasestatic/qfsfileengine_iterator.o .obj/releasestatic/qiodevice.o
.obj/releasestatic/qtemporaryfile.o .obj/releasestatic/qtextstream.o
.obj/releasestatic/qmetatype.o .obj/releasestatic/qvariant.o
.obj/releasestatic/qsystemerror.o .obj/releasestatic/qbitarray.o
.obj/releasestatic/qbytearray.o .obj/releasestatic/qbytearraymatcher.o
.obj/releasestatic/qdatetime.o .obj/releasestatic/qhash.o
.obj/releasestatic/qlist.o .obj/releasestatic/qlocale.o
.obj/releasestatic/qlocale_tools.o .obj/releasestatic/qmap.o
.obj/releasestatic/qregexp.o .obj/releasestatic/qstring.o
.obj/releasestatic/qstringlist.o .obj/releasestatic/qvector.o
.obj/releasestatic/qvsnprintf.o .obj/releasestatic/qxmlutils.o
.obj/releasestatic/qxmlstream.o .obj/releasestatic/qdom.o
.obj/releasestatic/qxml.o .obj/releasestatic/qfilesystemengine_unix.o
.obj/release-static/qfilesystemiterator_unix.o
.obj/releasestatic/qfsfileengine_unix.o .obj/releasestatic/qlocale_unix.o
make[1]: xiar: Command not found
make[1]: ***** [libbootstrap.a] Error 127
make[1]: Leaving directory
`/nfs/tmp2/brugger/build_python/qteverywhereopensourcesrc4.8.3/src/tools/bootstrap'
make: ***** [subtoolsbootstrapmake_defaultordered] Error 2
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1651
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: build_visit2_7_0 fails building python with icc.
Assigned to: Kathleen Biagas
Category:
Target version: 2.7.1
Author: Eric Brugger
Start: 11/08/2013
Due date:
% Done: 0
Estimated time:
Created: 11/08/2013 05:32 pm
Updated: 12/11/2013 05:33 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: 2.7.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Rick Angelini reported that he was unable to build visit on his SGI/ICE system using icc 12.1.0. I was able to reproduce the problem on edge with icc 12.1.5. In my case build_visit continued past the python failure and eventually had a build failure in Qt. Here is the build_visit line:
env C_COMPILER=icc CXX_COMPILER=icpc PAR_INCLUDE=-I/usr/local/tools/mvapich-gnu/include ./build_visit2_7_0 \
--console --no-thirdparty --no-visit \
--mesa --cmake --python --vtk --qt --icet --pyside \
--hdf5 --silo --szip \
--makeflags -j4
Here is the Python failure:
icc -pthread -fno-strict-aliasing -g -O2 -DNDEBUG -m64 -fPIC -O2 -fPIC
-Ibuild/temp.linux-x86_64-2.7/libffi/include
-Ibuild/temp.linux-x86_64-2.7/libffi
-I/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src
-I/nfs/tmp2/brugger/build_python/visit/python/2.7.5/linux-x86_64_icc/include
-I. -IInclude -I./Include -I/usr/local/include
-I/nfs/tmp2/brugger/build_python/Python-2.7.5/Include
-I/nfs/tmp2/brugger/build_python/Python-2.7.5 -c
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
-o
build/temp.linux-x86_64-2.7/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.o
-Wall -fexceptions
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c(56):
error: identifier "__m128" is undefined
UINT128 i128;
^
compilation aborted for
/nfs/tmp2/brugger/build_python/Python-2.7.5/Modules/_ctypes/libffi/src/x86/ffi64.c
(code 2)
Python build finished, but the necessary bits to build these modules were not
found:
bsddb185 dl imageop
sunaudiodev
To find the necessary bits, look in setup.py in detect_modules() for the
module's name.
Failed to build these modules:
_ctypes
running build_scripts
running install_lib
Here is the Qt failure:
icpc -c -m64 -fPIC -O2 -O2 -falign-functions=16 -ansi-alias -fstrict-aliasing
-fPIC -fno-jump-tables -w1 -Wcheck -wd654,1572,411,873,1125,2259
-DQT_BOOTSTRAPPED -DQT_LITE_UNICODE -DQT_NO_CAST_FROM_ASCII
-DQT_NO_CAST_TO_ASCII -DQT_NO_CODECS -DQT_NO_DATASTREAM -DQT_NO_GEOM_VARIANT
-DQT_NO_LIBRARY -DQT_NO_QOBJECT -DQT_NO_STL -DQT_NO_SYSTEMLOCALE
-DQT_NO_TEXTSTREAM -DQT_NO_THREAD -DQT_NO_UNICODETABLES
-DQT_NO_USING_NAMESPACE -DQT_NO_DEPRECATED -D_LARGEFILE64_SOURCE
-D_LARGEFILE_SOURCE -I../../../mkspecs/linux-icc-64 -I. -I../../../include
-I../../../include/QtCore -I../../../include/QtXml -o
.obj/release-static/qlocale_unix.o ../../corelib/tools/qlocale_unix.cpp
rm -f libbootstrap.a
xiar cqs libbootstrap.a .obj/release-static/qisciicodec.o
.obj/release-static/qlatincodec.o .obj/release-static/qsimplecodec.o
.obj/release-static/qtextcodec.o .obj/release-static/qtsciicodec.o
.obj/release-static/qutfcodec.o .obj/release-static/qglobal.o
.obj/release-static/qmalloc.o .obj/release-static/qnumeric.o
.obj/release-static/qabstractfileengine.o .obj/release-static/qbuffer.o
.obj/release-static/qdatastream.o .obj/release-static/qdir.o
.obj/release-static/qdiriterator.o .obj/release-static/qfile.o
.obj/release-static/qfileinfo.o .obj/release-static/qfilesystementry.o
.obj/release-static/qfilesystemengine.o .obj/release-static/qfsfileengine.o
.obj/release-static/qfsfileengine_iterator.o .obj/release-static/qiodevice.o
.obj/release-static/qtemporaryfile.o .obj/release-static/qtextstream.o
.obj/release-static/qmetatype.o .obj/release-static/qvariant.o
.obj/release-static/qsystemerror.o .obj/release-static/qbitarray.o
.obj/release-static/qbytearray.o .obj/release-static/qbytearraymatcher.o
.obj/release-static/qdatetime.o .obj/release-static/qhash.o
.obj/release-static/qlist.o .obj/release-static/qlocale.o
.obj/release-static/qlocale_tools.o .obj/release-static/qmap.o
.obj/release-static/qregexp.o .obj/release-static/qstring.o
.obj/release-static/qstringlist.o .obj/release-static/qvector.o
.obj/release-static/qvsnprintf.o .obj/release-static/qxmlutils.o
.obj/release-static/qxmlstream.o .obj/release-static/qdom.o
.obj/release-static/qxml.o .obj/release-static/qfilesystemengine_unix.o
.obj/release-static/qfilesystemiterator_unix.o
.obj/release-static/qfsfileengine_unix.o .obj/release-static/qlocale_unix.o
make[1]: xiar: Command not found
make[1]: *** [libbootstrap.a] Error 127
make[1]: Leaving directory
`/nfs/tmp2/brugger/build_python/qt-everywhere-opensource-src-4.8.3/src/tools/bootstrap'
make: *** [sub-tools-bootstrap-make_default-ordered] Error 2
Comments:
Hi Eric,The Python error may not be critical (at least for now)..Looks like Qt is configuring the AR command to be "xiar"mkspecs/linux-icc/qmake.conf:QMAKE_AR = xiar cqsA quick fix might be to set a simple "alias xiar=ar" or modify the qmake.conf file to what ever might be appropriate call to create archives for your system.Best regards,Hari
The QT build issue Eric encountered is due to the LC setup. We need to issue a 'use icc' before attempting to build, as doing so makes available the full suite of intel compiler tools (including xiar used by QT).
2.7RC update 22380, merge-to-trunk 22382Added minor python patch to bv_python, fixing ffi64 issue when compiling with icc. The patch is from a newer version of ffi64 library.
| priority | build fails building python with icc rick angelini reported that he was unable to build visit on his sgi ice system using icc i was able to reproduce the problem on edge with icc in my case build visit continued past the python failure and eventually had a build failure in qt here is the build visit line env c compiler icc cxx compiler icpc par include i usr local tools mvapichgnu include build console nothirdparty novisit mesa cmake python vtk qt icet pyside silo szip makeflags here is the python failure icc pthread fnostrictaliasing g dndebug fpic fpic ibuild temp libffi include ibuild temp libffi i nfs brugger build python modules ctypes libffi src i nfs brugger build python visit python icc include i iinclude i include i usr local include i nfs brugger build python include i nfs brugger build python c nfs brugger build python python modules ctypes libffi src c o build temp nfs brugger build python python modules ctypes libffi src o wall fexceptions nfs brugger build python python modules ctypes libffi src c error identifier is undefined compilation aborted for nfs brugger build python python modules ctypes libffi src c code python build finished but the necessary bits to build these modules were not found dl imageop sunaudiodev to find the necessary bits look in setup py in detect modules for the module s name failed to build these modules ctypes running build scripts running install lib here is the qt failure icpc c fpic falignfunctions ansialias fstrict aliasing fpic fnojumptables wcheck dqt bootstrapped dqt lite unicode dqt no cast from ascii dqt no cast to ascii dqt no codecs dqt no datastream dqt no geom variant dqt no library dqt no qobject dqt no stl dqt no systemlocale dqt no textstream dqt no thread dqt no unicodetables dqt no using namespace dqt no deprecated d source d largefile source i mkspecs i i include i include qtcore i include qtxml o obj release static qlocale unix o corelib tools qlocale unix cpp rm f libbootstrap a xiar cqs libbootstrap a obj release static qisciicodec o obj releasestatic qlatincodec o obj releasestatic qsimplecodec o obj releasestatic qtextcodec o obj releasestatic qtsciicodec o obj releasestatic qutfcodec o obj releasestatic qglobal o obj releasestatic qmalloc o obj releasestatic qnumeric o obj releasestatic qabstractfileengine o obj releasestatic qbuffer o obj releasestatic qdatastream o obj releasestatic qdir o obj releasestatic qdiriterator o obj releasestatic qfile o obj releasestatic qfileinfo o obj releasestatic qfilesystementry o obj releasestatic qfilesystemengine o obj releasestatic qfsfileengine o obj releasestatic qfsfileengine iterator o obj releasestatic qiodevice o obj releasestatic qtemporaryfile o obj releasestatic qtextstream o obj releasestatic qmetatype o obj releasestatic qvariant o obj releasestatic qsystemerror o obj releasestatic qbitarray o obj releasestatic qbytearray o obj releasestatic qbytearraymatcher o obj releasestatic qdatetime o obj releasestatic qhash o obj releasestatic qlist o obj releasestatic qlocale o obj releasestatic qlocale tools o obj releasestatic qmap o obj releasestatic qregexp o obj releasestatic qstring o obj releasestatic qstringlist o obj releasestatic qvector o obj releasestatic qvsnprintf o obj releasestatic qxmlutils o obj releasestatic qxmlstream o obj releasestatic qdom o obj releasestatic qxml o obj releasestatic qfilesystemengine unix o obj release static qfilesystemiterator unix o obj releasestatic qfsfileengine unix o obj releasestatic qlocale unix o make xiar command not found make error make leaving directory nfs brugger build python src tools bootstrap make error redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject build fails building python with icc assigned to kathleen biagas category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity very serious found in version impact expected use os all support group any description rick angelini reported that he was unable to build visit on his sgi ice system using icc i was able to reproduce the problem on edge with icc in my case build visit continued past the python failure and eventually had a build failure in qt here is the build visit line env c compiler icc cxx compiler icpc par include i usr local tools mvapich gnu include build console no thirdparty no visit mesa cmake python vtk qt icet pyside silo szip makeflags here is the python failure icc pthread fno strict aliasing g dndebug fpic fpic ibuild temp linux libffi include ibuild temp linux libffi i nfs brugger build python python modules ctypes libffi src i nfs brugger build python visit python linux icc include i iinclude i include i usr local include i nfs brugger build python python include i nfs brugger build python python c nfs brugger build python python modules ctypes libffi src c o build temp linux nfs brugger build python python modules ctypes libffi src o wall fexceptions nfs brugger build python python modules ctypes libffi src c error identifier is undefined compilation aborted for nfs brugger build python python modules ctypes libffi src c code python build finished but the necessary bits to build these modules were not found dl imageop sunaudiodev to find the necessary bits look in setup py in detect modules for the module s name failed to build these modules ctypes running build scripts running install lib here is the qt failure icpc c fpic falign functions ansi alias fstrict aliasing fpic fno jump tables wcheck dqt bootstrapped dqt lite unicode dqt no cast from ascii dqt no cast to ascii dqt no codecs dqt no datastream dqt no geom variant dqt no library dqt no qobject dqt no stl dqt no systemlocale dqt no textstream dqt no thread dqt no unicodetables dqt no using namespace dqt no deprecated d source d largefile source i mkspecs linux icc i i include i include qtcore i include qtxml o obj release static qlocale unix o corelib tools qlocale unix cpp rm f libbootstrap a xiar cqs libbootstrap a obj release static qisciicodec o obj release static qlatincodec o obj release static qsimplecodec o obj release static qtextcodec o obj release static qtsciicodec o obj release static qutfcodec o obj release static qglobal o obj release static qmalloc o obj release static qnumeric o obj release static qabstractfileengine o obj release static qbuffer o obj release static qdatastream o obj release static qdir o obj release static qdiriterator o obj release static qfile o obj release static qfileinfo o obj release static qfilesystementry o obj release static qfilesystemengine o obj release static qfsfileengine o obj release static qfsfileengine iterator o obj release static qiodevice o obj release static qtemporaryfile o obj release static qtextstream o obj release static qmetatype o obj release static qvariant o obj release static qsystemerror o obj release static qbitarray o obj release static qbytearray o obj release static qbytearraymatcher o obj release static qdatetime o obj release static qhash o obj release static qlist o obj release static qlocale o obj release static qlocale tools o obj release static qmap o obj release static qregexp o obj release static qstring o obj release static qstringlist o obj release static qvector o obj release static qvsnprintf o obj release static qxmlutils o obj release static qxmlstream o obj release static qdom o obj release static qxml o obj release static qfilesystemengine unix o obj release static qfilesystemiterator unix o obj release static qfsfileengine unix o obj release static qlocale unix o make xiar command not found make error make leaving directory nfs brugger build python qt everywhere opensource src src tools bootstrap make error comments hi eric the python error may not be critical at least for now looks like qt is configuring the ar command to be xiar mkspecs linux icc qmake conf qmake ar xiar cqsa quick fix might be to set a simple alias xiar ar or modify the qmake conf file to what ever might be appropriate call to create archives for your system best regards hari the qt build issue eric encountered is due to the lc setup we need to issue a use icc before attempting to build as doing so makes available the full suite of intel compiler tools including xiar used by qt update merge to trunk minor python patch to bv python fixing issue when compiling with icc the patch is from a newer version of library | 1 |
232,761 | 7,674,770,808 | IssuesEvent | 2018-05-15 06:08:02 | NIFCloud-mbaas/ncmb_js | https://api.github.com/repos/NIFCloud-mbaas/ncmb_js | closed | ライセンス表記修正 | Priority: Medium in progress | * サイト、ドキュメントと合わせて、以下のように修正
```
Copyright 2017 FUJITSU CLOUD TECHNOLOGIES LIMITED All Rights Reserved.
```
↓
```
Copyright 2017-2018 FUJITSU CLOUD TECHNOLOGIES LIMITED All Rights Reserved.
```
| 1.0 | ライセンス表記修正 - * サイト、ドキュメントと合わせて、以下のように修正
```
Copyright 2017 FUJITSU CLOUD TECHNOLOGIES LIMITED All Rights Reserved.
```
↓
```
Copyright 2017-2018 FUJITSU CLOUD TECHNOLOGIES LIMITED All Rights Reserved.
```
| priority | ライセンス表記修正 サイト、ドキュメントと合わせて、以下のように修正 copyright fujitsu cloud technologies limited all rights reserved ↓ copyright fujitsu cloud technologies limited all rights reserved | 1 |
580,002 | 17,202,842,546 | IssuesEvent | 2021-07-17 16:09:58 | michaelrsweet/rasterview | https://api.github.com/repos/michaelrsweet/rasterview | closed | Mac OS 11.4 (Big Sur) support | bug platform issue priority-medium | Rasterview currently does not show anything when opening a PWG file on Big Sur. Is this a known issue?
| 1.0 | Mac OS 11.4 (Big Sur) support - Rasterview currently does not show anything when opening a PWG file on Big Sur. Is this a known issue?
| priority | mac os big sur support rasterview currently does not show anything when opening a pwg file on big sur is this a known issue | 1 |
341,541 | 10,296,046,006 | IssuesEvent | 2019-08-28 11:10:39 | medic/medic | https://api.github.com/repos/medic/medic | opened | Refactor TasksContent controller to use unique URLs for actions | Priority: 2 - Medium Type: Technical issue | Currently tasks have one or more actions associated with them. Having multiple possible actions is an underused feature of tasks but it would be really useful for example if an ANC visit task could lead to a Delivery or a Visit report being filed to resolve it.
If you select a task with exactly one action then you go straight to completing that action. If there are multiple possible actions, _or_ the task has a detailed description (another underused feature) then you see a summary page and can select one of the actions.
Currently these both invoke the `$scope.performAction` function, but this requires a redux listener to detect when a task is selected. This is bad form.
Instead each action should have a unique URL (eg: `/tasks/:id/actions/:actionId`) which is handled by a new child controller. Put the `$scope.performAction` logic in the new controller. When selecting a task with only one action we can go straight to `/tasks/:id/actions/0`. When selecting a task with multiple actions use the existing TasksContent controller to render the detail and have each action button link to the new action URL.
| 1.0 | Refactor TasksContent controller to use unique URLs for actions - Currently tasks have one or more actions associated with them. Having multiple possible actions is an underused feature of tasks but it would be really useful for example if an ANC visit task could lead to a Delivery or a Visit report being filed to resolve it.
If you select a task with exactly one action then you go straight to completing that action. If there are multiple possible actions, _or_ the task has a detailed description (another underused feature) then you see a summary page and can select one of the actions.
Currently these both invoke the `$scope.performAction` function, but this requires a redux listener to detect when a task is selected. This is bad form.
Instead each action should have a unique URL (eg: `/tasks/:id/actions/:actionId`) which is handled by a new child controller. Put the `$scope.performAction` logic in the new controller. When selecting a task with only one action we can go straight to `/tasks/:id/actions/0`. When selecting a task with multiple actions use the existing TasksContent controller to render the detail and have each action button link to the new action URL.
| priority | refactor taskscontent controller to use unique urls for actions currently tasks have one or more actions associated with them having multiple possible actions is an underused feature of tasks but it would be really useful for example if an anc visit task could lead to a delivery or a visit report being filed to resolve it if you select a task with exactly one action then you go straight to completing that action if there are multiple possible actions or the task has a detailed description another underused feature then you see a summary page and can select one of the actions currently these both invoke the scope performaction function but this requires a redux listener to detect when a task is selected this is bad form instead each action should have a unique url eg tasks id actions actionid which is handled by a new child controller put the scope performaction logic in the new controller when selecting a task with only one action we can go straight to tasks id actions when selecting a task with multiple actions use the existing taskscontent controller to render the detail and have each action button link to the new action url | 1 |
260,219 | 8,205,124,892 | IssuesEvent | 2018-09-03 09:07:59 | medialab/fonio | https://api.github.com/repos/medialab/fonio | closed | Library panel bug | bug priority:medium user interface | Bug leaving a blank space in the middle column of the editor layout and outputting console errors.
1. Create a story.
2. Go to the library page and add any asset.
3. Go back to the summary.
4. Click on a section
The editor's layout is now messed up. | 1.0 | Library panel bug - Bug leaving a blank space in the middle column of the editor layout and outputting console errors.
1. Create a story.
2. Go to the library page and add any asset.
3. Go back to the summary.
4. Click on a section
The editor's layout is now messed up. | priority | library panel bug bug leaving a blank space in the middle column of the editor layout and outputting console errors create a story go to the library page and add any asset go back to the summary click on a section the editor s layout is now messed up | 1 |
640,591 | 20,793,681,311 | IssuesEvent | 2022-03-17 06:44:08 | kubeapps/kubeapps | https://api.github.com/repos/kubeapps/kubeapps | closed | [fluxv2] Add `defaultUpgradePolicy` in the Fluxv2 | kind/feature priority/medium size/S component/plugin-flux | ### Description:
As per the offline discussion we had, a short-term solution for allowing passing the version selection in Flux v2 would be just allowing the Fluxv2 to be configured with a plugin-wide config as we were doing in Carvel (see https://github.com/kubeapps/kubeapps/issues/4346#issuecomment-1054819538)
Later on, we can start discussing long-term solutions like https://github.com/kubeapps/kubeapps/issues/4365, but it is not a priority right now, as the community's feedback is that this version selection is not extensively being used out there, so just adding a plugin-wide option would be enough for now.
| 1.0 | [fluxv2] Add `defaultUpgradePolicy` in the Fluxv2 - ### Description:
As per the offline discussion we had, a short-term solution for allowing passing the version selection in Flux v2 would be just allowing the Fluxv2 to be configured with a plugin-wide config as we were doing in Carvel (see https://github.com/kubeapps/kubeapps/issues/4346#issuecomment-1054819538)
Later on, we can start discussing long-term solutions like https://github.com/kubeapps/kubeapps/issues/4365, but it is not a priority right now, as the community's feedback is that this version selection is not extensively being used out there, so just adding a plugin-wide option would be enough for now.
| priority | add defaultupgradepolicy in the description as per the offline discussion we had a short term solution for allowing passing the version selection in flux would be just allowing the to be configured with a plugin wide config as we were doing in carvel see later on we can start discussing long term solutions like but it is not a priority right now as the community s feedback is that this version selection is not extensively being used out there so just adding a plugin wide option would be enough for now | 1 |
25,575 | 2,683,848,445 | IssuesEvent | 2015-03-28 11:35:35 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | PictureView2 mod18 - нужна переделка! | 1 star bug imported invalid Priority-Medium | _From [victo...@mail333.com](https://code.google.com/u/114732384912597087095/) on November 30, 2009 03:44:46_
Что имеем: сбои в настройках если их удалить из Реестра. Все строки и имена
параметров слетели раз, gfl.pvd не находит субплугины, два, кстати это же
явление относится и к другим субплугинам. Скриншоты привожу. Реестровые
блоки настроек так же. Ещё одно замечание какое стоит устранить: в строке
string:GFL subplugins folder\ стоит на уровне исходников убрать "\" в конце
данной строки в меню.
В общем по моему, данную версию надо доделать - ошибок многовато вылезло.
**Attachment:** [log.png PictureView2_mod17.reg PictureView2_mod18.reg gfl.png](http://code.google.com/p/conemu-maximus5/issues/detail?id=134)
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=134_ | 1.0 | PictureView2 mod18 - нужна переделка! - _From [victo...@mail333.com](https://code.google.com/u/114732384912597087095/) on November 30, 2009 03:44:46_
Что имеем: сбои в настройках если их удалить из Реестра. Все строки и имена
параметров слетели раз, gfl.pvd не находит субплугины, два, кстати это же
явление относится и к другим субплугинам. Скриншоты привожу. Реестровые
блоки настроек так же. Ещё одно замечание какое стоит устранить: в строке
string:GFL subplugins folder\ стоит на уровне исходников убрать "\" в конце
данной строки в меню.
В общем по моему, данную версию надо доделать - ошибок многовато вылезло.
**Attachment:** [log.png PictureView2_mod17.reg PictureView2_mod18.reg gfl.png](http://code.google.com/p/conemu-maximus5/issues/detail?id=134)
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=134_ | priority | нужна переделка from on november что имеем сбои в настройках если их удалить из реестра все строки и имена параметров слетели раз gfl pvd не находит субплугины два кстати это же явление относится и к другим субплугинам скриншоты привожу реестровые блоки настроек так же ещё одно замечание какое стоит устранить в строке string gfl subplugins folder стоит на уровне исходников убрать в конце данной строки в меню в общем по моему данную версию надо доделать ошибок многовато вылезло attachment original issue | 1 |
57,290 | 3,081,254,477 | IssuesEvent | 2015-08-22 14:46:19 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Level must be saved before uploading to level database or you lose all changes | 019x 020 bug imported Priority-Medium | _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on April 29, 2015 12:11:34_
Level must be saved before uploading to level database or you lose all changes
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=504_ | 1.0 | Level must be saved before uploading to level database or you lose all changes - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on April 29, 2015 12:11:34_
Level must be saved before uploading to level database or you lose all changes
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=504_ | priority | level must be saved before uploading to level database or you lose all changes from on april level must be saved before uploading to level database or you lose all changes original issue | 1 |
124,135 | 4,892,394,784 | IssuesEvent | 2016-11-18 19:35:31 | SELinuxProject/selinux-kernel | https://api.github.com/repos/SELinuxProject/selinux-kernel | opened | RFE: add LSM/SELinux hooks for bus1 | enhancement priority/medium | The bus1 effort seems to be on a path to upstreaming. The bus1 developers are including a set of LSM hooks based on the binder hooks and discussions with them; we will ultimately need to validate those hooks and implement them for SELinux, along with corresponding policy changes.
* http://www.bus1.org | 1.0 | RFE: add LSM/SELinux hooks for bus1 - The bus1 effort seems to be on a path to upstreaming. The bus1 developers are including a set of LSM hooks based on the binder hooks and discussions with them; we will ultimately need to validate those hooks and implement them for SELinux, along with corresponding policy changes.
* http://www.bus1.org | priority | rfe add lsm selinux hooks for the effort seems to be on a path to upstreaming the developers are including a set of lsm hooks based on the binder hooks and discussions with them we will ultimately need to validate those hooks and implement them for selinux along with corresponding policy changes | 1 |
630,604 | 20,114,516,247 | IssuesEvent | 2022-02-07 18:04:10 | vmware/singleton | https://api.github.com/repos/vmware/singleton | opened | [BUG] [Ruby Client] Client library has hardcoded "default" locale. | kind/bug priority/medium area/ruby-client | **Describe the bug**
The client is not fetching a list of fallback locales and source localle from client configuration file. Instead, a "default" locale is hardcoded.
**To Reproduce**
Steps to reproduce the behavior:
1. Use the ruby client in a sample application.
2. Error: Notice that in the client configuration file, there is no config to set the following:
- list of fallback locales (ordered in priority)
- sourece locale
**Expected behavior**
The above two configurations must be in the client config file. These are to be used for the client library workflow logic below.
When target locale is not supported, invalid or nil:
a. If list of fallback locales is NOT configured in the client config file, getString must return the source string.
b. If list of fallback locales is configured in the client config file, getString must use the next fallback locale from the list as target locale, and return the translation. It must go through the list until it is able to return a translation in a fallback locale. If at the end of the list and all fallback locales failed, it must return the source string.
| 1.0 | [BUG] [Ruby Client] Client library has hardcoded "default" locale. - **Describe the bug**
The client is not fetching a list of fallback locales and source localle from client configuration file. Instead, a "default" locale is hardcoded.
**To Reproduce**
Steps to reproduce the behavior:
1. Use the ruby client in a sample application.
2. Error: Notice that in the client configuration file, there is no config to set the following:
- list of fallback locales (ordered in priority)
- sourece locale
**Expected behavior**
The above two configurations must be in the client config file. These are to be used for the client library workflow logic below.
When target locale is not supported, invalid or nil:
a. If list of fallback locales is NOT configured in the client config file, getString must return the source string.
b. If list of fallback locales is configured in the client config file, getString must use the next fallback locale from the list as target locale, and return the translation. It must go through the list until it is able to return a translation in a fallback locale. If at the end of the list and all fallback locales failed, it must return the source string.
| priority | client library has hardcoded default locale describe the bug the client is not fetching a list of fallback locales and source localle from client configuration file instead a default locale is hardcoded to reproduce steps to reproduce the behavior use the ruby client in a sample application error notice that in the client configuration file there is no config to set the following list of fallback locales ordered in priority sourece locale expected behavior the above two configurations must be in the client config file these are to be used for the client library workflow logic below when target locale is not supported invalid or nil a if list of fallback locales is not configured in the client config file getstring must return the source string b if list of fallback locales is configured in the client config file getstring must use the next fallback locale from the list as target locale and return the translation it must go through the list until it is able to return a translation in a fallback locale if at the end of the list and all fallback locales failed it must return the source string | 1 |
399,761 | 11,760,122,010 | IssuesEvent | 2020-03-13 18:44:34 | ayumi-cloud/oc-security-module | https://api.github.com/repos/ayumi-cloud/oc-security-module | opened | Expand device labels in firewall dashboard | Dashboard Firewall New UI Priority: Medium enhancement in-progress | ### Enhancement idea
- [ ] Expand device labels in firewall dashboard.
| 1.0 | Expand device labels in firewall dashboard - ### Enhancement idea
- [ ] Expand device labels in firewall dashboard.
| priority | expand device labels in firewall dashboard enhancement idea expand device labels in firewall dashboard | 1 |
127,013 | 5,011,564,715 | IssuesEvent | 2016-12-13 08:23:25 | mrb1218/e-Sports | https://api.github.com/repos/mrb1218/e-Sports | closed | Improve how a user creates a favorite | medium priority | Two things.
1) Currently, the user must enter a Team ID or a Player ID. The user should be able to enter a team name or player ign instead. OR we could add an "Add as Favorite" button to the individual team/player pages. This button would only show when you are logged in and would create a favorite using the current_user.id and the team.id or player.id
2) The controller is setup to not allow the user to enter both and will not commit the favorite to the database but we need to alert the user that there was an error "Cannot enter both a teamID and a playerID" | 1.0 | Improve how a user creates a favorite - Two things.
1) Currently, the user must enter a Team ID or a Player ID. The user should be able to enter a team name or player ign instead. OR we could add an "Add as Favorite" button to the individual team/player pages. This button would only show when you are logged in and would create a favorite using the current_user.id and the team.id or player.id
2) The controller is setup to not allow the user to enter both and will not commit the favorite to the database but we need to alert the user that there was an error "Cannot enter both a teamID and a playerID" | priority | improve how a user creates a favorite two things currently the user must enter a team id or a player id the user should be able to enter a team name or player ign instead or we could add an add as favorite button to the individual team player pages this button would only show when you are logged in and would create a favorite using the current user id and the team id or player id the controller is setup to not allow the user to enter both and will not commit the favorite to the database but we need to alert the user that there was an error cannot enter both a teamid and a playerid | 1 |
725,444 | 24,961,990,327 | IssuesEvent | 2022-11-01 16:13:31 | gitblit/gitblit | https://api.github.com/repos/gitblit/gitblit | closed | Add copy repository url button or single click to selected url? | Catg-Enhancement Priority-Medium Status-Accepted | #### Same as title and three click will selected extraneous characters | 1.0 | Add copy repository url button or single click to selected url? - #### Same as title and three click will selected extraneous characters | priority | add copy repository url button or single click to selected url same as title and three click will selected extraneous characters | 1 |
377,202 | 11,165,898,188 | IssuesEvent | 2019-12-27 11:05:10 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | USER ISSUE: Menu Sound Volume not controlled by SFX Volume slider | Medium Priority | I play with the SFX Volume set to 50%, but the click when I open the ESC menu while in-game doesn't seem to be affected by this setting, so it's really loud. | 1.0 | USER ISSUE: Menu Sound Volume not controlled by SFX Volume slider - I play with the SFX Volume set to 50%, but the click when I open the ESC menu while in-game doesn't seem to be affected by this setting, so it's really loud. | priority | user issue menu sound volume not controlled by sfx volume slider i play with the sfx volume set to but the click when i open the esc menu while in game doesn t seem to be affected by this setting so it s really loud | 1 |
60,622 | 3,130,902,568 | IssuesEvent | 2015-09-09 12:13:26 | dhowe/RiTa | https://api.github.com/repos/dhowe/RiTa | opened | Optimize tests/RiLexicon-tests.js | PRIORITY: Medium RiTaJS | This test is currently taking ~40 seconds when running outside of the browser (headlessly). Please profile and propose some steps for reducing the time. | 1.0 | Optimize tests/RiLexicon-tests.js - This test is currently taking ~40 seconds when running outside of the browser (headlessly). Please profile and propose some steps for reducing the time. | priority | optimize tests rilexicon tests js this test is currently taking seconds when running outside of the browser headlessly please profile and propose some steps for reducing the time | 1 |
823,511 | 31,022,295,279 | IssuesEvent | 2023-08-10 06:37:09 | queuing4oranges/lesbotoc-website | https://api.github.com/repos/queuing4oranges/lesbotoc-website | opened | Admin/events - refresh problem when editing events | bug medium priority lesbotoc-admin | - check also other requests for bugs with refreshing | 1.0 | Admin/events - refresh problem when editing events - - check also other requests for bugs with refreshing | priority | admin events refresh problem when editing events check also other requests for bugs with refreshing | 1 |
23,449 | 2,659,624,900 | IssuesEvent | 2015-03-18 22:09:17 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | File upload: Page is broken after file ingest completes, can't publish or download tabular files. | Priority: Medium Status: QA Type: Bug |
I created a dataset, saved. Uploaded akiotestfiles.zip, waited for it to complete ingest and then clicked publish, nothing happened. Left page, came back and it works. | 1.0 | File upload: Page is broken after file ingest completes, can't publish or download tabular files. -
I created a dataset, saved. Uploaded akiotestfiles.zip, waited for it to complete ingest and then clicked publish, nothing happened. Left page, came back and it works. | priority | file upload page is broken after file ingest completes can t publish or download tabular files i created a dataset saved uploaded akiotestfiles zip waited for it to complete ingest and then clicked publish nothing happened left page came back and it works | 1 |
79,545 | 3,536,217,338 | IssuesEvent | 2016-01-17 03:35:58 | ESAPI/esapi-java-legacy | https://api.github.com/repos/ESAPI/esapi-java-legacy | closed | Incorrect Equality test on floating point values | bug imported Priority-Medium | _From [eamonn.w...@gmail.com](https://code.google.com/u/104114803397285871233/) on November 26, 2013 14:23:09_
I have V2.1 sources.
2 places in RuleUtil.java have an invalid floating point comparison logic.
The corrected block(for both cases) is:
if ( Double.compare(Double.parseDouble(s), ((Double)o).doubleValue()) == 0 ) {
return true;
}
_Original issue: http://code.google.com/p/owasp-esapi-java/issues/detail?id=313_ | 1.0 | Incorrect Equality test on floating point values - _From [eamonn.w...@gmail.com](https://code.google.com/u/104114803397285871233/) on November 26, 2013 14:23:09_
I have V2.1 sources.
2 places in RuleUtil.java have an invalid floating point comparison logic.
The corrected block(for both cases) is:
if ( Double.compare(Double.parseDouble(s), ((Double)o).doubleValue()) == 0 ) {
return true;
}
_Original issue: http://code.google.com/p/owasp-esapi-java/issues/detail?id=313_ | priority | incorrect equality test on floating point values from on november i have sources places in ruleutil java have an invalid floating point comparison logic the corrected block for both cases is if double compare double parsedouble s double o doublevalue return true original issue | 1 |
416,704 | 12,150,597,324 | IssuesEvent | 2020-04-24 18:16:33 | TykTechnologies/tyk | https://api.github.com/repos/TykTechnologies/tyk | closed | Extend NewRelic support | Priority: Medium enhancement wontfix | **Do you want to request a *feature* or report a *bug*?**
feature
**What is the current behavior?**
Right now NewRelic tracks integration all incoming requests, even including Admin API. However, it will be nice to track Tyk->Upstream calls as well, to see Upstream latency without added Tyk one (or if Tyk configured as cache). Additionally, it will be good to add tracking of Redis and RCP calls.
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
2.5 | 1.0 | Extend NewRelic support - **Do you want to request a *feature* or report a *bug*?**
feature
**What is the current behavior?**
Right now NewRelic tracks integration all incoming requests, even including Admin API. However, it will be nice to track Tyk->Upstream calls as well, to see Upstream latency without added Tyk one (or if Tyk configured as cache). Additionally, it will be good to add tracking of Redis and RCP calls.
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
2.5 | priority | extend newrelic support do you want to request a feature or report a bug feature what is the current behavior right now newrelic tracks integration all incoming requests even including admin api however it will be nice to track tyk upstream calls as well to see upstream latency without added tyk one or if tyk configured as cache additionally it will be good to add tracking of redis and rcp calls which versions of tyk affected by this issue did this work in previous versions of tyk | 1 |
9,233 | 2,607,932,414 | IssuesEvent | 2015-02-26 00:27:19 | chrsmithdemos/minify | https://api.github.com/repos/chrsmithdemos/minify | opened | Allow "min" to operate in subdirectory of doc_root | auto-migrated Priority-Medium Type-Enhancement | ```
Since many developers test sites in a subdirectory on their localhost
server, it makes sense to at least try to formally support this, along
with user accounts like http://testserver/~acct/
Settings for different servers also makes sense, or at least documentation
on how to differentiate based on $_SERVER['SERVER_NAME'].
```
-----
Original issue reported on code.google.com by `mrclay....@gmail.com` on 19 Aug 2009 at 8:07 | 1.0 | Allow "min" to operate in subdirectory of doc_root - ```
Since many developers test sites in a subdirectory on their localhost
server, it makes sense to at least try to formally support this, along
with user accounts like http://testserver/~acct/
Settings for different servers also makes sense, or at least documentation
on how to differentiate based on $_SERVER['SERVER_NAME'].
```
-----
Original issue reported on code.google.com by `mrclay....@gmail.com` on 19 Aug 2009 at 8:07 | priority | allow min to operate in subdirectory of doc root since many developers test sites in a subdirectory on their localhost server it makes sense to at least try to formally support this along with user accounts like settings for different servers also makes sense or at least documentation on how to differentiate based on server original issue reported on code google com by mrclay gmail com on aug at | 1 |
609,292 | 18,870,252,750 | IssuesEvent | 2021-11-13 03:21:03 | scilus/fibernavigator | https://api.github.com/repos/scilus/fibernavigator | closed | Chosen color in colorpicker is not used when drawing on anatomy | bug imported Priority-Medium OpSys-OSX | _Original author: caroline...@gmail.com (January 31, 2012 19:43:32)_
<b>What steps will reproduce the problem?</b>
1. Create a new anatomy RGB or use a RGB anatomy
2. Put it on the top of the list to see it
3. In the drawing toolbar, choose a color for the brush with the color picker
4. Draw on the anatomy. It will draw in white instead of the chosen color.
<b>What is the expected output? What do you see instead?</b>
It should draw with the right color instead of white.
_Original issue: http://code.google.com/p/fibernavigator/issues/detail?id=53_
| 1.0 | Chosen color in colorpicker is not used when drawing on anatomy - _Original author: caroline...@gmail.com (January 31, 2012 19:43:32)_
<b>What steps will reproduce the problem?</b>
1. Create a new anatomy RGB or use a RGB anatomy
2. Put it on the top of the list to see it
3. In the drawing toolbar, choose a color for the brush with the color picker
4. Draw on the anatomy. It will draw in white instead of the chosen color.
<b>What is the expected output? What do you see instead?</b>
It should draw with the right color instead of white.
_Original issue: http://code.google.com/p/fibernavigator/issues/detail?id=53_
| priority | chosen color in colorpicker is not used when drawing on anatomy original author caroline gmail com january what steps will reproduce the problem create a new anatomy rgb or use a rgb anatomy put it on the top of the list to see it in the drawing toolbar choose a color for the brush with the color picker draw on the anatomy it will draw in white instead of the chosen color what is the expected output what do you see instead it should draw with the right color instead of white original issue | 1 |
52,120 | 3,021,647,971 | IssuesEvent | 2015-07-31 15:51:49 | ghaering/pysqlite | https://api.github.com/repos/ghaering/pysqlite | closed | [Sugestion] Unify execute() and executescript() | bug imported Priority-Medium | _From [piranna](https://code.google.com/u/piranna/) on August 14, 2011 15:33:00_
execute() method (both in cursor and connection) accept as a second parameter a dict or a tuple to use as parameters for the SQL query being executed, but only accept one query at a time, so when you need to exec several queries is ineficient and, as far as i know, insecure in multithreaded applications because if you are not in a transaction one is created automatically for every statement, and another connection would be able to access to the database between two queries-in-a-row (because they are different calls).
On the other hand, executescript() can be able to exec several SQL queries in a row (and AFAIK, the other threads are not able to access to the database until executescript() has finished to exec all its queries), but doesn't accept any parameter for the query so you have to compose it by hand, what it's somewhat insecure and in any case, ugly.
What i proposs is to allow to execute() to support the execution of several queries in a row as executescript() does, and if there's no other technicall difference or performance is not affected, i would deprecate executescript(), having this way only two functions to exec queries, execute() and executemany() (that doesn't have almost anything in common with the others and i don't know if it would be possible or desirable also to unify...).
_Original issue: http://code.google.com/p/pysqlite/issues/detail?id=41_ | 1.0 | [Sugestion] Unify execute() and executescript() - _From [piranna](https://code.google.com/u/piranna/) on August 14, 2011 15:33:00_
execute() method (both in cursor and connection) accept as a second parameter a dict or a tuple to use as parameters for the SQL query being executed, but only accept one query at a time, so when you need to exec several queries is ineficient and, as far as i know, insecure in multithreaded applications because if you are not in a transaction one is created automatically for every statement, and another connection would be able to access to the database between two queries-in-a-row (because they are different calls).
On the other hand, executescript() can be able to exec several SQL queries in a row (and AFAIK, the other threads are not able to access to the database until executescript() has finished to exec all its queries), but doesn't accept any parameter for the query so you have to compose it by hand, what it's somewhat insecure and in any case, ugly.
What i proposs is to allow to execute() to support the execution of several queries in a row as executescript() does, and if there's no other technicall difference or performance is not affected, i would deprecate executescript(), having this way only two functions to exec queries, execute() and executemany() (that doesn't have almost anything in common with the others and i don't know if it would be possible or desirable also to unify...).
_Original issue: http://code.google.com/p/pysqlite/issues/detail?id=41_ | priority | unify execute and executescript from on august execute method both in cursor and connection accept as a second parameter a dict or a tuple to use as parameters for the sql query being executed but only accept one query at a time so when you need to exec several queries is ineficient and as far as i know insecure in multithreaded applications because if you are not in a transaction one is created automatically for every statement and another connection would be able to access to the database between two queries in a row because they are different calls on the other hand executescript can be able to exec several sql queries in a row and afaik the other threads are not able to access to the database until executescript has finished to exec all its queries but doesn t accept any parameter for the query so you have to compose it by hand what it s somewhat insecure and in any case ugly what i proposs is to allow to execute to support the execution of several queries in a row as executescript does and if there s no other technicall difference or performance is not affected i would deprecate executescript having this way only two functions to exec queries execute and executemany that doesn t have almost anything in common with the others and i don t know if it would be possible or desirable also to unify original issue | 1 |
459,441 | 13,193,126,901 | IssuesEvent | 2020-08-13 14:46:58 | pesos/grofer | https://api.github.com/repos/pesos/grofer | opened | [FEATURE REQ] add functionality for getting help for a particular command | enhancement priority: medium | **Is your feature request related to a problem? Please describe.**
When running a particular interactive command like `grofer` or `grofer proc` it would be nice to have a way to get something like a table of available key bindings and what they do for that command.
**Describe the solution you'd like**
Pressing a key like `h` should pause the UI rendering and display a table of the available key bindings for that command and what they do like `↑` to scroll, `k` to kill a process in the `grofer proc` command once #16 is done etc. | 1.0 | [FEATURE REQ] add functionality for getting help for a particular command - **Is your feature request related to a problem? Please describe.**
When running a particular interactive command like `grofer` or `grofer proc` it would be nice to have a way to get something like a table of available key bindings and what they do for that command.
**Describe the solution you'd like**
Pressing a key like `h` should pause the UI rendering and display a table of the available key bindings for that command and what they do like `↑` to scroll, `k` to kill a process in the `grofer proc` command once #16 is done etc. | priority | add functionality for getting help for a particular command is your feature request related to a problem please describe when running a particular interactive command like grofer or grofer proc it would be nice to have a way to get something like a table of available key bindings and what they do for that command describe the solution you d like pressing a key like h should pause the ui rendering and display a table of the available key bindings for that command and what they do like ↑ to scroll k to kill a process in the grofer proc command once is done etc | 1 |
71,703 | 3,367,617,883 | IssuesEvent | 2015-11-22 10:19:03 | music-encoding/music-encoding | https://api.github.com/repos/music-encoding/music-encoding | closed | Simplify relationship between textphraseLike.limited and textphraseLike models | Priority: Medium Type: Enhancement | _From [pd...@virginia.edu](https://code.google.com/u/103686026181985548448/) on January 27, 2015 17:37:58_
Since textphraseLike.limited is a subset of textphraseLike, it should be a member of textphraseLike. milestoneLike.txt should *not* be a member of either textphraseLike.limited or textphraseLike.
memberOf textphraseLike.limited:
addressLike, bibLike, dateLike, editorialLike, figureLike, identifierLike, annotLike, locrefLike, measurementLike, nameLike*, rendLike, symbol, titleLike, lbLike
memberOf textphraseLike:
textphraseLike.limited, pbLike
memberOf milestoneLike.text:
lbLike, pbLike
_Original issue: http://code.google.com/p/music-encoding/issues/detail?id=210_ | 1.0 | Simplify relationship between textphraseLike.limited and textphraseLike models - _From [pd...@virginia.edu](https://code.google.com/u/103686026181985548448/) on January 27, 2015 17:37:58_
Since textphraseLike.limited is a subset of textphraseLike, it should be a member of textphraseLike. milestoneLike.txt should *not* be a member of either textphraseLike.limited or textphraseLike.
memberOf textphraseLike.limited:
addressLike, bibLike, dateLike, editorialLike, figureLike, identifierLike, annotLike, locrefLike, measurementLike, nameLike*, rendLike, symbol, titleLike, lbLike
memberOf textphraseLike:
textphraseLike.limited, pbLike
memberOf milestoneLike.text:
lbLike, pbLike
_Original issue: http://code.google.com/p/music-encoding/issues/detail?id=210_ | priority | simplify relationship between textphraselike limited and textphraselike models from on january since textphraselike limited is a subset of textphraselike it should be a member of textphraselike milestonelike txt should not be a member of either textphraselike limited or textphraselike memberof textphraselike limited addresslike biblike datelike editoriallike figurelike identifierlike annotlike locreflike measurementlike namelike rendlike symbol titlelike lblike memberof textphraselike textphraselike limited pblike memberof milestonelike text lblike pblike original issue | 1 |
232,674 | 7,673,900,133 | IssuesEvent | 2018-05-15 00:37:25 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Networking Issue - Non Default Ports | Medium Priority | When setting up my server, I used ports 4000 and 4001. Before 7.4, this worked as expected, people could see the server, see the status of it, connect, etc. Since 7.4, the server perpetually shows as offline in the server browser and does not appear in the public server tab.
Expected Behavior:
Server shows as online and joinable without IP address
Actual Behavior:
Only shows online and joinable when using default ports | 1.0 | Networking Issue - Non Default Ports - When setting up my server, I used ports 4000 and 4001. Before 7.4, this worked as expected, people could see the server, see the status of it, connect, etc. Since 7.4, the server perpetually shows as offline in the server browser and does not appear in the public server tab.
Expected Behavior:
Server shows as online and joinable without IP address
Actual Behavior:
Only shows online and joinable when using default ports | priority | networking issue non default ports when setting up my server i used ports and before this worked as expected people could see the server see the status of it connect etc since the server perpetually shows as offline in the server browser and does not appear in the public server tab expected behavior server shows as online and joinable without ip address actual behavior only shows online and joinable when using default ports | 1 |
175,518 | 6,551,696,172 | IssuesEvent | 2017-09-05 15:33:42 | ColinW520/hydra | https://api.github.com/repos/ColinW520/hydra | opened | Redirect after changing payment method (credit card) | Lift: LIGHT Priority: MEDIUM Type: BUG | After a user changes/updates their credit card (with an existing account), redirect them to the settings page. | 1.0 | Redirect after changing payment method (credit card) - After a user changes/updates their credit card (with an existing account), redirect them to the settings page. | priority | redirect after changing payment method credit card after a user changes updates their credit card with an existing account redirect them to the settings page | 1 |
582,661 | 17,367,176,547 | IssuesEvent | 2021-07-30 08:56:42 | gambitph/Stackable | https://api.github.com/repos/gambitph/Stackable | closed | Use wp.blockEditor.ColorPaletteControl instead of our own Component | enhancement medium priority | We implemented our own ColorPaletteControl since the one in Gutenberg wasn't exposed in the API. Now that https://github.com/WordPress/gutenberg/issues/13018 is resolved, we can now do:
`import { ColorPalleteControl } from '@wordpress/block-editor'`
This only works in WP 5.4+, so we'll need to do backward compatibility for lower versions. | 1.0 | Use wp.blockEditor.ColorPaletteControl instead of our own Component - We implemented our own ColorPaletteControl since the one in Gutenberg wasn't exposed in the API. Now that https://github.com/WordPress/gutenberg/issues/13018 is resolved, we can now do:
`import { ColorPalleteControl } from '@wordpress/block-editor'`
This only works in WP 5.4+, so we'll need to do backward compatibility for lower versions. | priority | use wp blockeditor colorpalettecontrol instead of our own component we implemented our own colorpalettecontrol since the one in gutenberg wasn t exposed in the api now that is resolved we can now do import colorpalletecontrol from wordpress block editor this only works in wp so we ll need to do backward compatibility for lower versions | 1 |
489,106 | 14,101,266,429 | IssuesEvent | 2020-11-06 06:27:24 | MarketSquare/robotframework-browser | https://api.github.com/repos/MarketSquare/robotframework-browser | reopened | `Fill secret` leaks the secret into log.html | bug good first issue priority: medium | **Describe the bug**
According to [documentation](https://marketsquare.github.io/robotframework-browser/Browser.html#Fill%20Secret):
> The difference to Fill Text is that this keyword does not log the text to be written into the text field.
However, one can get the content of `secret` to log.html
**To Reproduce**
Steps to reproduce the behavior:
1. Create a test case that uses `Fill Secret`
eg.
```
*** Settings ***
Library Browser retry_assertions_for=5s
*** Variables ***
${HEADLESS}= ${True}
${BROWSER}= chromium
*** Test Cases ***
TC
Open browser http://google.com browser=${BROWSER} headless=${HEADLESS}
Click css=iframe >>> id=introAgreeButton
Fill secret xpath=//input[@title="Search"] | //input[@title="Haku"] I'm a secret!
```
1. Run the test with `--loglevel TRACE`
**Expected behavior**
"I'm a secret" shouldn't appear in log.html
**Screenshots**
<img width="811" alt="Screenshot 2020-10-15 at 21 11 20" src="https://user-images.githubusercontent.com/280218/96169578-03ff3780-0f2b-11eb-85ea-a7611cd4269d.png">
| 1.0 | `Fill secret` leaks the secret into log.html - **Describe the bug**
According to [documentation](https://marketsquare.github.io/robotframework-browser/Browser.html#Fill%20Secret):
> The difference to Fill Text is that this keyword does not log the text to be written into the text field.
However, one can get the content of `secret` to log.html
**To Reproduce**
Steps to reproduce the behavior:
1. Create a test case that uses `Fill Secret`
eg.
```
*** Settings ***
Library Browser retry_assertions_for=5s
*** Variables ***
${HEADLESS}= ${True}
${BROWSER}= chromium
*** Test Cases ***
TC
Open browser http://google.com browser=${BROWSER} headless=${HEADLESS}
Click css=iframe >>> id=introAgreeButton
Fill secret xpath=//input[@title="Search"] | //input[@title="Haku"] I'm a secret!
```
1. Run the test with `--loglevel TRACE`
**Expected behavior**
"I'm a secret" shouldn't appear in log.html
**Screenshots**
<img width="811" alt="Screenshot 2020-10-15 at 21 11 20" src="https://user-images.githubusercontent.com/280218/96169578-03ff3780-0f2b-11eb-85ea-a7611cd4269d.png">
| priority | fill secret leaks the secret into log html describe the bug according to the difference to fill text is that this keyword does not log the text to be written into the text field however one can get the content of secret to log html to reproduce steps to reproduce the behavior create a test case that uses fill secret eg settings library browser retry assertions for variables headless true browser chromium test cases tc open browser browser browser headless headless click css iframe id introagreebutton fill secret xpath input input i m a secret run the test with loglevel trace expected behavior i m a secret shouldn t appear in log html screenshots img width alt screenshot at src | 1 |
554,474 | 16,421,301,215 | IssuesEvent | 2021-05-19 12:52:27 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | net.socket.tls.tls_ext: frdm_k64f test failure | area: Networking bug platform: NXP priority: medium | **Describe the bug**
all testing fails
**To Reproduce**
Steps to reproduce the behavior:
in tests/net/socket/tls_ext
1. mkdir build; cd build
2. cmake -DBOARD=frdm_k64f ..
3. make; make flash
4. See error
**Expected behavior**
test pass
**Impact**
https supporting
**Logs and console output**
```
W: You have 1 IPv6 net_if addresses but 2 network interfaces
W: Consider increasing CONFIG_NET_IF_MAX_IPV6_COUNT value.
W: You have 1 IPv4 net_if addresses but 2 network interfaces
W: Consider increasing CONFIG_NET_IF_MAX_IPV4_COUNT value.
*** Booting Zephyr OS build zephyr-v2.5.0-3919-gcd9421a45e72 ***
I: Initializing network
I: IPv4 address: 192.0.2.1
Running test suite tls_socket_api_extension
===================================================================
START - test_tls_peer_verify_none
E: TLS handshake error: -4e
Assertion failed at WEST_TOPDIR/zephyr/tests/net/socket/tls_ext/src/main.c:366: test_common: (r equal to -1)
failed to connect (113)
FAIL - test_tls_peer_verify_none in 0.19 seconds
===================================================================
START - test_tls_peer_verify_optional
E: Identical connection handler 0x20003a70 already found.
Assertion failed at WEST_TOPDIR/zephyr/tests/net/socket/tls_ext/src/main.c:281: test_common: (r equal to -1)
failed to listen (120)
FAIL - test_tls_peer_verify_optional in 0.18 seconds
===================================================================
START - test_tls_peer_verify_required
W: Failed to allocate TLS context
Assertion failed at WEST_TOPDIR/zephyr/tests/net/socket/tls_ext/src/main.c:217: test_common: (r equal to -1)
failed to create server socket (12)
FAIL - test_tls_peer_verify_required in 0.17 seconds
===================================================================
Test suite tls_socket_api_extension failed.
===================================================================
PROJECT EXECUTION FAILED
```
**Environment (please complete the following information):**
- OS: (e.g. Linux)
- Toolchain (e.g Zephyr SDK, ...)
- Commit SHA or Version used: zephyr-v2.5.0-3919-gcd9421a45e72 / v2.6.0-rc1-159-g8717a0f678
| 1.0 | net.socket.tls.tls_ext: frdm_k64f test failure - **Describe the bug**
all testing fails
**To Reproduce**
Steps to reproduce the behavior:
in tests/net/socket/tls_ext
1. mkdir build; cd build
2. cmake -DBOARD=frdm_k64f ..
3. make; make flash
4. See error
**Expected behavior**
test pass
**Impact**
https supporting
**Logs and console output**
```
W: You have 1 IPv6 net_if addresses but 2 network interfaces
W: Consider increasing CONFIG_NET_IF_MAX_IPV6_COUNT value.
W: You have 1 IPv4 net_if addresses but 2 network interfaces
W: Consider increasing CONFIG_NET_IF_MAX_IPV4_COUNT value.
*** Booting Zephyr OS build zephyr-v2.5.0-3919-gcd9421a45e72 ***
I: Initializing network
I: IPv4 address: 192.0.2.1
Running test suite tls_socket_api_extension
===================================================================
START - test_tls_peer_verify_none
E: TLS handshake error: -4e
Assertion failed at WEST_TOPDIR/zephyr/tests/net/socket/tls_ext/src/main.c:366: test_common: (r equal to -1)
failed to connect (113)
FAIL - test_tls_peer_verify_none in 0.19 seconds
===================================================================
START - test_tls_peer_verify_optional
E: Identical connection handler 0x20003a70 already found.
Assertion failed at WEST_TOPDIR/zephyr/tests/net/socket/tls_ext/src/main.c:281: test_common: (r equal to -1)
failed to listen (120)
FAIL - test_tls_peer_verify_optional in 0.18 seconds
===================================================================
START - test_tls_peer_verify_required
W: Failed to allocate TLS context
Assertion failed at WEST_TOPDIR/zephyr/tests/net/socket/tls_ext/src/main.c:217: test_common: (r equal to -1)
failed to create server socket (12)
FAIL - test_tls_peer_verify_required in 0.17 seconds
===================================================================
Test suite tls_socket_api_extension failed.
===================================================================
PROJECT EXECUTION FAILED
```
**Environment (please complete the following information):**
- OS: (e.g. Linux)
- Toolchain (e.g Zephyr SDK, ...)
- Commit SHA or Version used: zephyr-v2.5.0-3919-gcd9421a45e72 / v2.6.0-rc1-159-g8717a0f678
| priority | net socket tls tls ext frdm test failure describe the bug all testing fails to reproduce steps to reproduce the behavior in tests net socket tls ext mkdir build cd build cmake dboard frdm make make flash see error expected behavior test pass impact https supporting logs and console output w you have net if addresses but network interfaces w consider increasing config net if max count value w you have net if addresses but network interfaces w consider increasing config net if max count value booting zephyr os build zephyr i initializing network i address running test suite tls socket api extension start test tls peer verify none e tls handshake error assertion failed at west topdir zephyr tests net socket tls ext src main c test common r equal to failed to connect fail test tls peer verify none in seconds start test tls peer verify optional e identical connection handler already found assertion failed at west topdir zephyr tests net socket tls ext src main c test common r equal to failed to listen fail test tls peer verify optional in seconds start test tls peer verify required w failed to allocate tls context assertion failed at west topdir zephyr tests net socket tls ext src main c test common r equal to failed to create server socket fail test tls peer verify required in seconds test suite tls socket api extension failed project execution failed environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used zephyr | 1 |
128,451 | 5,065,072,480 | IssuesEvent | 2016-12-23 10:17:28 | dhowe/AdNauseamV1 | https://api.github.com/repos/dhowe/AdNauseamV1 | closed | Lack of regex for Google Ad text Ad in some sites | Bug PRIORITY: Medium | The Google Ad of the following sites sometimes shows text ad instead of image ad.
In case of showing text ads, ADN cannot identify them.
http://www.foxsports.com/
http://www.pcworld.com/
http://www.dailymotion.com/tw
e.g. on foxsports

title selector: .rh10c
text selector: .rh11c
site selector: .rh1110c
[This attached file](https://drive.google.com/file/d/0BzHBWuszZv5_NElCb080LUdMem8/view) is the saved webpage of foxsports.
| 1.0 | Lack of regex for Google Ad text Ad in some sites - The Google Ad of the following sites sometimes shows text ad instead of image ad.
In case of showing text ads, ADN cannot identify them.
http://www.foxsports.com/
http://www.pcworld.com/
http://www.dailymotion.com/tw
e.g. on foxsports

title selector: .rh10c
text selector: .rh11c
site selector: .rh1110c
[This attached file](https://drive.google.com/file/d/0BzHBWuszZv5_NElCb080LUdMem8/view) is the saved webpage of foxsports.
| priority | lack of regex for google ad text ad in some sites the google ad of the following sites sometimes shows text ad instead of image ad in case of showing text ads adn cannot identify them e g on foxsports title selector text selector site selector is the saved webpage of foxsports | 1 |
782,389 | 27,495,253,815 | IssuesEvent | 2023-03-05 03:40:35 | Baystation12/Baystation12 | https://api.github.com/repos/Baystation12/Baystation12 | closed | Force Feeding someone a chess piece will create a null empty version of the piece | Priority: Medium Could Reproduce | <!--
Anything inside tags like these is a comment and will not be displayed in the final issue.
Be careful not to write inside them!
Every field other than 'specific information for locating' is required.
If you do not fill out the 'specific information' field, please delete the header.
/!\ Omitting or not answering a required field will result in your issue being closed. /!\
Repeated violation of this rule, or joke or spam issues, will result in punishment.
PUT YOUR ANSWERS ON THE BLANK LINES BELOW THE HEADERS
(The lines with four #'s)
Don't edit them or delete them - it's part of the formatting
-->
#### Description of issue
This null spriteless item occupies your hand until you drop it or otherwise put it down.
#### Difference between expected and actual behavior
It should delete the item properly like if you eat it.
#### Steps to reproduce
Force feed someone, try to feed them again.
"None of the [Item] left!"
#### Specific information for locating
<!-- e.g. an object name, paste specific message outputs... -->
/obj/item/reagent_containers/food/snacks/checker
#### Length of time in which bug has been known to occur
<!--
Be specific if you approximately know the time it's been occurring
for—this can speed up finding the source. If you're not sure
about it, tell us too!
-->
At least 6/11/2022
#### Client version, Server revision & Game ID
<!-- Found with the "Show server revision" verb in the OOC tab in game. -->
Client Version: 514
Server Revision: [740b4f6f0738ac3265260aca619a48b47556b6f2](https://bay.ss13.me/github/commit/740b4f6f0738ac3265260aca619a48b47556b6f2) - dev - 2022-06-09
Game ID: cin-dgny
Current map: SEV Torch
#### Issue bingo
<!-- Check these by writing an x inside the [ ] (like this: [x])-->
<!-- Don't forget to remove the space between the brackets, or it won't work! -->
- [x] Issue could be reproduced at least once
- [x] Issue could be reproduced by different players
- [x] Issue could be reproduced in multiple rounds
- [x] Issue happened in a recent (less than 7 days ago) round
- [x] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
| 1.0 | Force Feeding someone a chess piece will create a null empty version of the piece - <!--
Anything inside tags like these is a comment and will not be displayed in the final issue.
Be careful not to write inside them!
Every field other than 'specific information for locating' is required.
If you do not fill out the 'specific information' field, please delete the header.
/!\ Omitting or not answering a required field will result in your issue being closed. /!\
Repeated violation of this rule, or joke or spam issues, will result in punishment.
PUT YOUR ANSWERS ON THE BLANK LINES BELOW THE HEADERS
(The lines with four #'s)
Don't edit them or delete them - it's part of the formatting
-->
#### Description of issue
This null spriteless item occupies your hand until you drop it or otherwise put it down.
#### Difference between expected and actual behavior
It should delete the item properly like if you eat it.
#### Steps to reproduce
Force feed someone, try to feed them again.
"None of the [Item] left!"
#### Specific information for locating
<!-- e.g. an object name, paste specific message outputs... -->
/obj/item/reagent_containers/food/snacks/checker
#### Length of time in which bug has been known to occur
<!--
Be specific if you approximately know the time it's been occurring
for—this can speed up finding the source. If you're not sure
about it, tell us too!
-->
At least 6/11/2022
#### Client version, Server revision & Game ID
<!-- Found with the "Show server revision" verb in the OOC tab in game. -->
Client Version: 514
Server Revision: [740b4f6f0738ac3265260aca619a48b47556b6f2](https://bay.ss13.me/github/commit/740b4f6f0738ac3265260aca619a48b47556b6f2) - dev - 2022-06-09
Game ID: cin-dgny
Current map: SEV Torch
#### Issue bingo
<!-- Check these by writing an x inside the [ ] (like this: [x])-->
<!-- Don't forget to remove the space between the brackets, or it won't work! -->
- [x] Issue could be reproduced at least once
- [x] Issue could be reproduced by different players
- [x] Issue could be reproduced in multiple rounds
- [x] Issue happened in a recent (less than 7 days ago) round
- [x] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
| priority | force feeding someone a chess piece will create a null empty version of the piece anything inside tags like these is a comment and will not be displayed in the final issue be careful not to write inside them every field other than specific information for locating is required if you do not fill out the specific information field please delete the header omitting or not answering a required field will result in your issue being closed repeated violation of this rule or joke or spam issues will result in punishment put your answers on the blank lines below the headers the lines with four s don t edit them or delete them it s part of the formatting description of issue this null spriteless item occupies your hand until you drop it or otherwise put it down difference between expected and actual behavior it should delete the item properly like if you eat it steps to reproduce force feed someone try to feed them again none of the left specific information for locating obj item reagent containers food snacks checker length of time in which bug has been known to occur be specific if you approximately know the time it s been occurring for—this can speed up finding the source if you re not sure about it tell us too at least client version server revision game id client version server revision dev game id cin dgny current map sev torch issue bingo issue could be reproduced at least once issue could be reproduced by different players issue could be reproduced in multiple rounds issue happened in a recent less than days ago round | 1 |
292,091 | 8,953,087,601 | IssuesEvent | 2019-01-25 18:24:08 | richelbilderbeek/djog_unos_2018 | https://api.github.com/repos/richelbilderbeek/djog_unos_2018 | reopened | Grass creates new grass | medium priority | **Is your feature request related to a problem? Please describe.**
Currently, grass just grows, but does not create new grasses.
**Describe the solution you'd like**
After a while, a grass agent should create a new grass agent.
Or, this test should pass:
```c++
//#define FIX_ISSUE_300
#ifdef FIX_ISSUE_300
//Grass creates new grasses
{
game g(create_default_tiles(), { agent(agent_type::grass) } );
assert(g.get_agents().size() == 1);
while (g.get_agents().size() == 1) //Wait until grass procreates
{
g.process_events();
}
assert(g.get_agents()[0].get_type() == agent_type::grass);
assert(g.get_agents()[1].get_type() == agent_type::grass);
}
#endif //FIX_ISSUE_300
```
**Describe alternatives you've considered**
None.
**Additional context**
Running the test now will freeze the game. | 1.0 | Grass creates new grass - **Is your feature request related to a problem? Please describe.**
Currently, grass just grows, but does not create new grasses.
**Describe the solution you'd like**
After a while, a grass agent should create a new grass agent.
Or, this test should pass:
```c++
//#define FIX_ISSUE_300
#ifdef FIX_ISSUE_300
//Grass creates new grasses
{
game g(create_default_tiles(), { agent(agent_type::grass) } );
assert(g.get_agents().size() == 1);
while (g.get_agents().size() == 1) //Wait until grass procreates
{
g.process_events();
}
assert(g.get_agents()[0].get_type() == agent_type::grass);
assert(g.get_agents()[1].get_type() == agent_type::grass);
}
#endif //FIX_ISSUE_300
```
**Describe alternatives you've considered**
None.
**Additional context**
Running the test now will freeze the game. | priority | grass creates new grass is your feature request related to a problem please describe currently grass just grows but does not create new grasses describe the solution you d like after a while a grass agent should create a new grass agent or this test should pass c define fix issue ifdef fix issue grass creates new grasses game g create default tiles agent agent type grass assert g get agents size while g get agents size wait until grass procreates g process events assert g get agents get type agent type grass assert g get agents get type agent type grass endif fix issue describe alternatives you ve considered none additional context running the test now will freeze the game | 1 |
735,280 | 25,387,677,533 | IssuesEvent | 2022-11-21 23:42:31 | bounswe/bounswe2022group7 | https://api.github.com/repos/bounswe/bounswe2022group7 | opened | Converter from String of List to List | Status: In Progress Priority: Low Type: Research Difficulty: Medium Type: Implementation Target: Backend | We are currently holding a few fields that are originally lists as their string version to reduce redundant database tables. Frontend is translating this string to list. We are to provide a list and remove their conversion code. | 1.0 | Converter from String of List to List - We are currently holding a few fields that are originally lists as their string version to reduce redundant database tables. Frontend is translating this string to list. We are to provide a list and remove their conversion code. | priority | converter from string of list to list we are currently holding a few fields that are originally lists as their string version to reduce redundant database tables frontend is translating this string to list we are to provide a list and remove their conversion code | 1 |
139,988 | 5,396,001,362 | IssuesEvent | 2017-02-27 10:21:58 | HPI-SWA-Lab/BP2016H1 | https://api.github.com/repos/HPI-SWA-Lab/BP2016H1 | opened | Render Glyphs from other Projects | priority medium | As a designer, when I visit a page of a furduck, I want to see the same furduck rendered multiple times but from several different projects in order get inspiration. | 1.0 | Render Glyphs from other Projects - As a designer, when I visit a page of a furduck, I want to see the same furduck rendered multiple times but from several different projects in order get inspiration. | priority | render glyphs from other projects as a designer when i visit a page of a furduck i want to see the same furduck rendered multiple times but from several different projects in order get inspiration | 1 |
78,213 | 3,509,519,401 | IssuesEvent | 2016-01-08 23:12:38 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Warrior and Druid charge bug (BB #984) | Category: Spells migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** everdie
**Original Date:** 05.06.2015 17:12:48 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** closed
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/984
<hr>
https://youtu.be/EX-DA1j8uy4
I fall down the texture, when there is something under me, exp.:cave | 1.0 | Warrior and Druid charge bug (BB #984) - This issue was migrated from bitbucket.
**Original Reporter:** everdie
**Original Date:** 05.06.2015 17:12:48 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** closed
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/984
<hr>
https://youtu.be/EX-DA1j8uy4
I fall down the texture, when there is something under me, exp.:cave | priority | warrior and druid charge bug bb this issue was migrated from bitbucket original reporter everdie original date gmt original priority major original type bug original state closed direct link i fall down the texture when there is something under me exp cave | 1 |
437,082 | 12,560,079,624 | IssuesEvent | 2020-06-07 20:58:20 | Tyresius92/le-chat | https://api.github.com/repos/Tyresius92/le-chat | closed | [Client] Integrate Apollo Client into the Client | Good First issue Priority: Medium Status: Ready for Review Type: Enhancement/Feature | install the `apollo-boost` package, and add the apollo provider to the app root. | 1.0 | [Client] Integrate Apollo Client into the Client - install the `apollo-boost` package, and add the apollo provider to the app root. | priority | integrate apollo client into the client install the apollo boost package and add the apollo provider to the app root | 1 |
758,978 | 26,576,046,565 | IssuesEvent | 2023-01-21 20:45:46 | space-wizards/RobustToolbox | https://api.github.com/repos/space-wizards/RobustToolbox | closed | Insert / remove sprites from Dynamictree when Visible changes | Issue: Performance Priority: 3-Not Required Area: Rendering Difficulty: 2-Medium | Probably ContainerOcclude as well.
Currently ProcessSpriteEntities does 2 bool checks for every single rendered spritecomponent but visibility rarely changes (especially for subfloor stuff on ss14) so it's *probably* better that we just do the insertions / removals into the tree instead. | 1.0 | Insert / remove sprites from Dynamictree when Visible changes - Probably ContainerOcclude as well.
Currently ProcessSpriteEntities does 2 bool checks for every single rendered spritecomponent but visibility rarely changes (especially for subfloor stuff on ss14) so it's *probably* better that we just do the insertions / removals into the tree instead. | priority | insert remove sprites from dynamictree when visible changes probably containerocclude as well currently processspriteentities does bool checks for every single rendered spritecomponent but visibility rarely changes especially for subfloor stuff on so it s probably better that we just do the insertions removals into the tree instead | 1 |
817,001 | 30,621,746,295 | IssuesEvent | 2023-07-24 08:44:56 | mantlenetworkio/blockscout | https://api.github.com/repos/mantlenetworkio/blockscout | closed | DA verifier may encounter a 'nonce too high' | Priority - Medium | **System Specs:**
- OS:mac(rde)
- mantel Version/branch: develop
**Describe the bug**
In on-chain transactions, the DA verifier may encounter a 'nonce too high' error.
**To Reproduce**
Steps to reproduce the behavior:
1. run 'rde'
2. shell 'l2 transactions'
3. open 'verifierda log'
4. See error
**Expected behavior**
The DA verifier synchronizes normally without experiencing any freezing or abnormal behavior.
**Screenshots**
`INFO [07-24|14:21:06.294] SyncService sync nextIndex=65 latestIndex=206
INFO [07-24|14:21:06.294] Syncing eigen transaction batch range start=65 end=206
INFO [07-24|14:21:06.295] Dtl rollup upgrade datastore id upgradeDatastoreId=0
INFO [07-24|14:21:06.301] CheckUpgrade chain id=17
ERROR[07-24|14:21:06.301] Could not verify error="verifier cannot sync transactions with BackendEigen: Cannot sync da dtl transaction batches to tip: cannot sync eigen batches: Cannot apply tx : nonce too high"`
`INFO [07-24|14:24:11.065] SyncService sync nextIndex=200 latestIndex=220
INFO [07-24|14:24:11.065] Syncing eigen transaction batch range start=200 end=220
INFO [07-24|14:24:11.065] Dtl rollup upgrade datastore id upgradeDatastoreId=0
INFO [07-24|14:24:11.070] CheckUpgrade chain id=17
ERROR[07-24|14:24:11.070] Could not verify error="verifier cannot sync transactions with BackendEigen: Cannot sync da dtl transaction batches to tip: cannot sync eigen batches: Cannot apply tx : nonce too high"` | 1.0 | DA verifier may encounter a 'nonce too high' - **System Specs:**
- OS:mac(rde)
- mantel Version/branch: develop
**Describe the bug**
In on-chain transactions, the DA verifier may encounter a 'nonce too high' error.
**To Reproduce**
Steps to reproduce the behavior:
1. run 'rde'
2. shell 'l2 transactions'
3. open 'verifierda log'
4. See error
**Expected behavior**
The DA verifier synchronizes normally without experiencing any freezing or abnormal behavior.
**Screenshots**
`INFO [07-24|14:21:06.294] SyncService sync nextIndex=65 latestIndex=206
INFO [07-24|14:21:06.294] Syncing eigen transaction batch range start=65 end=206
INFO [07-24|14:21:06.295] Dtl rollup upgrade datastore id upgradeDatastoreId=0
INFO [07-24|14:21:06.301] CheckUpgrade chain id=17
ERROR[07-24|14:21:06.301] Could not verify error="verifier cannot sync transactions with BackendEigen: Cannot sync da dtl transaction batches to tip: cannot sync eigen batches: Cannot apply tx : nonce too high"`
`INFO [07-24|14:24:11.065] SyncService sync nextIndex=200 latestIndex=220
INFO [07-24|14:24:11.065] Syncing eigen transaction batch range start=200 end=220
INFO [07-24|14:24:11.065] Dtl rollup upgrade datastore id upgradeDatastoreId=0
INFO [07-24|14:24:11.070] CheckUpgrade chain id=17
ERROR[07-24|14:24:11.070] Could not verify error="verifier cannot sync transactions with BackendEigen: Cannot sync da dtl transaction batches to tip: cannot sync eigen batches: Cannot apply tx : nonce too high"` | priority | da verifier may encounter a nonce too high system specs os mac rde mantel version branch develop describe the bug in on chain transactions the da verifier may encounter a nonce too high error to reproduce steps to reproduce the behavior run rde shell transactions open verifierda log see error expected behavior the da verifier synchronizes normally without experiencing any freezing or abnormal behavior screenshots info syncservice sync nextindex latestindex info syncing eigen transaction batch range start end info dtl rollup upgrade datastore id upgradedatastoreid info checkupgrade chain id error could not verify error verifier cannot sync transactions with backendeigen cannot sync da dtl transaction batches to tip cannot sync eigen batches cannot apply tx nonce too high info syncservice sync nextindex latestindex info syncing eigen transaction batch range start end info dtl rollup upgrade datastore id upgradedatastoreid info checkupgrade chain id error could not verify error verifier cannot sync transactions with backendeigen cannot sync da dtl transaction batches to tip cannot sync eigen batches cannot apply tx nonce too high | 1 |
55,756 | 3,074,297,066 | IssuesEvent | 2015-08-20 05:59:15 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | Waiter.getLog doesn't throw Exception when logcat process encounter error | bug imported Priority-Medium | _From [n...@orz.hk](https://code.google.com/u/115475815206789581674/) on January 07, 2014 20:05:07_
What steps will reproduce the problem? 1. Remove READ_LOGS permission from project under test. On my environment the problem would happens even READ_LOGS permission is present in AndroidManifest.
2. setup a test with memthod waitForLogMessage, e.g., mSolo.waitForLogMessage("some log")
3. When the test run waitForLogMessage would return false, although "some log" has dumped to logcat. Error stream from logcat process would report "android Unable to open log device '/dev/log/main': Permission denied". What is the expected output? What do you see instead? Expect it to throw Exception when return value from Process is non-zero. The Exception message would be the error message from error stream. What version of the product are you using? On what operating system? Robotium 4.3.1. Android 4.4.2 on Samsusng S2 with CyanFox.
_Original issue: http://code.google.com/p/robotium/issues/detail?id=569_ | 1.0 | Waiter.getLog doesn't throw Exception when logcat process encounter error - _From [n...@orz.hk](https://code.google.com/u/115475815206789581674/) on January 07, 2014 20:05:07_
What steps will reproduce the problem? 1. Remove READ_LOGS permission from project under test. On my environment the problem would happens even READ_LOGS permission is present in AndroidManifest.
2. setup a test with memthod waitForLogMessage, e.g., mSolo.waitForLogMessage("some log")
3. When the test run waitForLogMessage would return false, although "some log" has dumped to logcat. Error stream from logcat process would report "android Unable to open log device '/dev/log/main': Permission denied". What is the expected output? What do you see instead? Expect it to throw Exception when return value from Process is non-zero. The Exception message would be the error message from error stream. What version of the product are you using? On what operating system? Robotium 4.3.1. Android 4.4.2 on Samsusng S2 with CyanFox.
_Original issue: http://code.google.com/p/robotium/issues/detail?id=569_ | priority | waiter getlog doesn t throw exception when logcat process encounter error from on january what steps will reproduce the problem remove read logs permission from project under test on my environment the problem would happens even read logs permission is present in androidmanifest setup a test with memthod waitforlogmessage e g msolo waitforlogmessage some log when the test run waitforlogmessage would return false although some log has dumped to logcat error stream from logcat process would report android unable to open log device dev log main permission denied what is the expected output what do you see instead expect it to throw exception when return value from process is non zero the exception message would be the error message from error stream what version of the product are you using on what operating system robotium android on samsusng with cyanfox original issue | 1 |
146,426 | 5,621,338,837 | IssuesEvent | 2017-04-04 09:41:27 | linux-audit/audit-userspace | https://api.github.com/repos/linux-audit/audit-userspace | opened | Convert libaudit field parsing code to use macros instead of hard-coded error numbers | enhancement priority/medium | The audit_rule_interfield_comp_data() and audit_rule_fieldpair_data() option field parsers use over 30 hard-coded numerical error codes to select an error message when errors are detected. Use meaningfully named macros to make the code easier to understand and verify. | 1.0 | Convert libaudit field parsing code to use macros instead of hard-coded error numbers - The audit_rule_interfield_comp_data() and audit_rule_fieldpair_data() option field parsers use over 30 hard-coded numerical error codes to select an error message when errors are detected. Use meaningfully named macros to make the code easier to understand and verify. | priority | convert libaudit field parsing code to use macros instead of hard coded error numbers the audit rule interfield comp data and audit rule fieldpair data option field parsers use over hard coded numerical error codes to select an error message when errors are detected use meaningfully named macros to make the code easier to understand and verify | 1 |
712,746 | 24,506,234,409 | IssuesEvent | 2022-10-10 16:37:35 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [docdb] issues when doing a lot of splits for already existing big tablets | kind/bug area/docdb priority/medium kind/perf | Jira Link: [[DB-306]](https://yugabyte.atlassian.net/browse/DB-306)
We have been running YCQL benchmark CassandraBatchTimeseries with 50GB tablet split size for few days. After replacing flag to 5GB write ops have dropped significantly.
[DB-306]: https://yugabyte.atlassian.net/browse/DB-306?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [docdb] issues when doing a lot of splits for already existing big tablets - Jira Link: [[DB-306]](https://yugabyte.atlassian.net/browse/DB-306)
We have been running YCQL benchmark CassandraBatchTimeseries with 50GB tablet split size for few days. After replacing flag to 5GB write ops have dropped significantly.
[DB-306]: https://yugabyte.atlassian.net/browse/DB-306?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | issues when doing a lot of splits for already existing big tablets jira link we have been running ycql benchmark cassandrabatchtimeseries with tablet split size for few days after replacing flag to write ops have dropped significantly | 1 |
646,687 | 21,056,510,569 | IssuesEvent | 2022-04-01 04:15:32 | AY2122S2-CS2103T-W14-3/tp | https://api.github.com/repos/AY2122S2-CS2103T-W14-3/tp | closed | As a user keeping track of many events, I can delete all past events | priority: medium | so that I can avoid having to delete them one by one | 1.0 | As a user keeping track of many events, I can delete all past events - so that I can avoid having to delete them one by one | priority | as a user keeping track of many events i can delete all past events so that i can avoid having to delete them one by one | 1 |
287,917 | 8,822,855,314 | IssuesEvent | 2019-01-02 11:07:54 | commercialhaskell/hindent | https://api.github.com/repos/commercialhaskell/hindent | closed | Fails to create required indentation for infix | component: hindent priority: medium type: bug | hindent (5.2.0):
``` diff
- do
- pgsqlExecStmt conn x
- proceed savepoint
-
- `catch` handleSqlError savepoint
- `catch` handleQueryError savepoint
+ do pgsqlExecStmt conn x
+ proceed savepoint `catch`
+ handleSqlError savepoint `catch`
+ handleQueryError savepoint
```
ghc (7.10.3):
`parse error (possibly incorrect indentation or mismatched brackets)`
Complete function
``` hs
pgsqlExecStmtList :: Status -> [SqlStatement] -> [SqlStatement] -> Connection -> IO ()
pgsqlExecStmtList _ [] [] conn = commit conn
pgsqlExecStmtList Unchanged [] failed conn = pgsqlExecStmtHandled conn (head failed)
pgsqlExecStmtList Changed [] failed conn = void $ pgsqlExecStmtList Unchanged failed [] conn
pgsqlExecStmtList status (x:xs) failed conn = do
savepoint <- newSavepoint conn
do
pgsqlExecStmt conn x
proceed savepoint
`catch` handleSqlError savepoint
`catch` handleQueryError savepoint
where
handleSqlError savepoint SqlError{sqlState=errCode}
| errCode == "42P13" = skipQuery savepoint [stmtDropFunction' x, x]
| otherwise = skipQuery savepoint [x]
handleQueryError savepoint QueryError{} = proceed savepoint
proceed savepoint = do
releaseSavepoint conn savepoint
pgsqlExecStmtList Changed xs failed conn
skipQuery savepoint stmts = do
rollbackToSavepoint conn savepoint
releaseSavepoint conn savepoint
pgsqlExecStmtList status xs (failed ++ stmts) conn
```
| 1.0 | Fails to create required indentation for infix - hindent (5.2.0):
``` diff
- do
- pgsqlExecStmt conn x
- proceed savepoint
-
- `catch` handleSqlError savepoint
- `catch` handleQueryError savepoint
+ do pgsqlExecStmt conn x
+ proceed savepoint `catch`
+ handleSqlError savepoint `catch`
+ handleQueryError savepoint
```
ghc (7.10.3):
`parse error (possibly incorrect indentation or mismatched brackets)`
Complete function
``` hs
pgsqlExecStmtList :: Status -> [SqlStatement] -> [SqlStatement] -> Connection -> IO ()
pgsqlExecStmtList _ [] [] conn = commit conn
pgsqlExecStmtList Unchanged [] failed conn = pgsqlExecStmtHandled conn (head failed)
pgsqlExecStmtList Changed [] failed conn = void $ pgsqlExecStmtList Unchanged failed [] conn
pgsqlExecStmtList status (x:xs) failed conn = do
savepoint <- newSavepoint conn
do
pgsqlExecStmt conn x
proceed savepoint
`catch` handleSqlError savepoint
`catch` handleQueryError savepoint
where
handleSqlError savepoint SqlError{sqlState=errCode}
| errCode == "42P13" = skipQuery savepoint [stmtDropFunction' x, x]
| otherwise = skipQuery savepoint [x]
handleQueryError savepoint QueryError{} = proceed savepoint
proceed savepoint = do
releaseSavepoint conn savepoint
pgsqlExecStmtList Changed xs failed conn
skipQuery savepoint stmts = do
rollbackToSavepoint conn savepoint
releaseSavepoint conn savepoint
pgsqlExecStmtList status xs (failed ++ stmts) conn
```
| priority | fails to create required indentation for infix hindent diff do pgsqlexecstmt conn x proceed savepoint catch handlesqlerror savepoint catch handlequeryerror savepoint do pgsqlexecstmt conn x proceed savepoint catch handlesqlerror savepoint catch handlequeryerror savepoint ghc parse error possibly incorrect indentation or mismatched brackets complete function hs pgsqlexecstmtlist status connection io pgsqlexecstmtlist conn commit conn pgsqlexecstmtlist unchanged failed conn pgsqlexecstmthandled conn head failed pgsqlexecstmtlist changed failed conn void pgsqlexecstmtlist unchanged failed conn pgsqlexecstmtlist status x xs failed conn do savepoint newsavepoint conn do pgsqlexecstmt conn x proceed savepoint catch handlesqlerror savepoint catch handlequeryerror savepoint where handlesqlerror savepoint sqlerror sqlstate errcode errcode skipquery savepoint otherwise skipquery savepoint handlequeryerror savepoint queryerror proceed savepoint proceed savepoint do releasesavepoint conn savepoint pgsqlexecstmtlist changed xs failed conn skipquery savepoint stmts do rollbacktosavepoint conn savepoint releasesavepoint conn savepoint pgsqlexecstmtlist status xs failed stmts conn | 1 |
420,704 | 12,241,569,766 | IssuesEvent | 2020-05-05 04:33:25 | x13pixels/remedybg-issues | https://api.github.com/repos/x13pixels/remedybg-issues | closed | Memory ranges in the watch window | Component: Watch Window Priority: 4 (Medium) Type: Enhancement | Allow specifying a range of memory that can be displayed in the watch window. | 1.0 | Memory ranges in the watch window - Allow specifying a range of memory that can be displayed in the watch window. | priority | memory ranges in the watch window allow specifying a range of memory that can be displayed in the watch window | 1 |
795,589 | 28,078,559,714 | IssuesEvent | 2023-03-30 03:13:14 | AY2223S2-CS2113-F13-2/tp | https://api.github.com/repos/AY2223S2-CS2113-F13-2/tp | closed | Save/read information | priority.Medium type.Enhancement | As a user, I can save/read information so that I can load previously inputted expenses, budget, and wishlists. | 1.0 | Save/read information - As a user, I can save/read information so that I can load previously inputted expenses, budget, and wishlists. | priority | save read information as a user i can save read information so that i can load previously inputted expenses budget and wishlists | 1 |
534,787 | 15,649,019,125 | IssuesEvent | 2021-03-23 06:52:30 | jku-vds-lab/projection-space-explorer | https://api.github.com/repos/jku-vds-lab/projection-space-explorer | closed | Empty String as Category Label for Categorical Columns with Empty Cells | bug medium priority | Version: Develop
Date: 22-03-2021 13:30
## Problem
Dataset: [coral_usecase_no_structure_subsampled_normalized_no_one_hot.zip](https://github.com/jku-vds-lab/projection-space-explorer/files/6182135/coral_usecase_no_structure_subsampled_normalized_no_one_hot.zip)
Reproduction
1. Load the dataset abive
2. Go to the Encodings tab
3. Color by KRAS AA Mutated (categories: 'Mutated', 'Non-Mutated','')
Color coding works, but one label for the Color is empty. This is also the case for the summary.
The functionality is still given, there is just no text for the category.
GIF

## Possible Solution
Define a label for such a case: 'null', 'Missing Value', ..
Maybe check also if the label already exists in the column, e.g. column has these categories: 'null','CatA', 'CatB', and ''
| 1.0 | Empty String as Category Label for Categorical Columns with Empty Cells - Version: Develop
Date: 22-03-2021 13:30
## Problem
Dataset: [coral_usecase_no_structure_subsampled_normalized_no_one_hot.zip](https://github.com/jku-vds-lab/projection-space-explorer/files/6182135/coral_usecase_no_structure_subsampled_normalized_no_one_hot.zip)
Reproduction
1. Load the dataset abive
2. Go to the Encodings tab
3. Color by KRAS AA Mutated (categories: 'Mutated', 'Non-Mutated','')
Color coding works, but one label for the Color is empty. This is also the case for the summary.
The functionality is still given, there is just no text for the category.
GIF

## Possible Solution
Define a label for such a case: 'null', 'Missing Value', ..
Maybe check also if the label already exists in the column, e.g. column has these categories: 'null','CatA', 'CatB', and ''
| priority | empty string as category label for categorical columns with empty cells version develop date problem dataset reproduction load the dataset abive go to the encodings tab color by kras aa mutated categories mutated non mutated color coding works but one label for the color is empty this is also the case for the summary the functionality is still given there is just no text for the category gif possible solution define a label for such a case null missing value maybe check also if the label already exists in the column e g column has these categories null cata catb and | 1 |
563,131 | 16,676,709,102 | IssuesEvent | 2021-06-07 17:06:07 | minio/minio | https://api.github.com/repos/minio/minio | closed | Error: io: read/write on closed pipe (*errors.errorString) in healing | community priority: medium | <!--- Provide a general summary of the issue in the Title above -->
## Expected Behavior
Minio runs without such message in log
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
## Current Behavior
Thousands of messages like
```shell
API: SYSTEM()
Time: 12:51:27 UTC 06/01/2021
DeploymentID: 0bbd5e0d-fdb6-4bbd-b210-0c13a31d2e36
Error: io: read/write on closed pipe (*errors.errorString)
12: cmd/data-scanner.go:869:cmd.(*scannerItem).applyHealing()
11: cmd/data-scanner.go:972:cmd.(*scannerItem).applyActions()
10: cmd/xl-storage.go:459:cmd.(*xlStorage).NSScanner.func1()
9: cmd/data-scanner.go:467:cmd.(*folderScanner).scanFolder.func1()
8: cmd/os-readdir_unix.go:176:cmd.readDirFn()
7: cmd/data-scanner.go:402:cmd.(*folderScanner).scanFolder()
6: cmd/data-scanner.go:519:cmd.(*folderScanner).scanFolder.func2()
5: cmd/data-scanner.go:583:cmd.(*folderScanner).scanFolder()
4: cmd/data-scanner.go:314:cmd.scanDataFolder()
3: cmd/xl-storage.go:429:cmd.(*xlStorage).NSScanner()
2: cmd/xl-storage-disk-id-check.go:171:cmd.(*xlStorageDiskIDCheck).NSScanner()
1: cmd/erasure.go:500:cmd.erasureObjects.nsScanner.func3()
```
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## Possible Solution
Show what exactly happens, more clear description of problem
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
<!--- and make sure you have followed https://github.com/minio/minio/tree/release/docs/debugging to capture relevant logs -->
1. Update to `RELEASE.2021-05-26T00-22-46Z`
2. Run with `minio server /minio/minio{1...98}` in standalone mode
3. Use it with 1000+ buckets and 3,3M objects
## Context
There was no unusual behavior before upgrading, minio was working on `RELEASE.2021-03-26T00-00-41Z`
First time I faced with this message on `RELEASE.2021-05-22T02-34-39Z`
I suppose, that this error is not just a notice. Something bad is going inside. I'm not sure if my data in Minio is safe.
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
## Regression
<!-- Is this issue a regression? (Yes / No) -->
<!-- If Yes, optionally please include minio version or commit id or PR# that caused this regression, if you have these details. -->
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio --version`):
```shell
mc admin info storage11
● storage11.s3.domain.org
Uptime: 5 days
Version: 2021-05-26T00:22:46Z
Network: 1/1 OK
Drives: 98/98 OK
18 TiB Used, 1,194 Buckets, 3,343,906 Objects
98 drives online, 0 drives offline
```
* Server setup and configuration:
Running from simple daemonset
<details><summary>minio.yaml</summary>
<p>
```yaml
---
apiVersion: v1
data:
mount-check.sh: |
#!/bin/sh
for i in $(seq 1 99); do
if grep -qs "minio$i" /proc/mounts; then
echo "It's mounted."
else
echo "It's not mounted."
mkdir -p /minio/minio$i
mount /dev/disk/by-label/minio$i /minio/minio$i
if [ $? -eq 0 ]; then
echo "Mount minio$i success!"
else
echo "Something went wrong with the mount minio$i"
fi
fi; done
kind: ConfigMap
metadata:
annotations:
qbec.io/component: minio
labels:
qbec.io/application: minio
qbec.io/environment: prod
name: mount-check
namespace: minio
---
apiVersion: v1
data:
sysctl.sh: |
#!/bin/bash
cat > sysctl.conf <<EOF
# maximum number of open files/file descriptors
fs.file-max = 4194303
# use as little swap space as possible
vm.swappiness = 1
# prioritize application RAM against disk/swap cache
vm.vfs_cache_pressure = 10
# minimum free memory
vm.min_free_kbytes = 1000000
# maximum receive socket buffer (bytes)
net.core.rmem_max = 268435456
# maximum send buffer socket buffer (bytes)
net.core.wmem_max = 268435456
# default receive buffer socket size (bytes)
net.core.rmem_default = 67108864
# default send buffer socket size (bytes)
net.core.wmem_default = 67108864
# maximum number of packets in one poll cycle
net.core.netdev_budget = 1200
# maximum ancillary buffer size per socket
net.core.optmem_max = 134217728
# maximum number of incoming connections
net.core.somaxconn = 65535
# maximum number of packets queued
net.core.netdev_max_backlog = 250000
# maximum read buffer space
net.ipv4.tcp_rmem = 67108864 134217728 268435456
# maximum write buffer space
net.ipv4.tcp_wmem = 67108864 134217728 268435456
# enable low latency mode
net.ipv4.tcp_low_latency = 1
# socket buffer portion used for TCP window
net.ipv4.tcp_adv_win_scale = 1
# queue length of completely established sockets waiting for accept
net.ipv4.tcp_max_syn_backlog = 30000
# maximum number of sockets in TIME_WAIT state
net.ipv4.tcp_max_tw_buckets = 2000000
# reuse sockets in TIME_WAIT state when safe
net.ipv4.tcp_tw_reuse = 1
# time to wait (seconds) for FIN packet
net.ipv4.tcp_fin_timeout = 5
# disable icmp send redirects
net.ipv4.conf.all.send_redirects = 0
# disable icmp accept redirect
net.ipv4.conf.all.accept_redirects = 0
# drop packets with LSR or SSR
net.ipv4.conf.all.accept_source_route = 0
# MTU discovery, only enable when ICMP blackhole detected
net.ipv4.tcp_mtu_probing = 1
EOF
echo "Enabling system level tuning params"
sysctl --quiet --load sysctl.conf && rm -f sysctl.conf
# `Transparent Hugepage Support`*: This is a Linux kernel feature intended to improve
# performance by making more efficient use of processor’s memory-mapping hardware.
# But this may cause https://blogs.oracle.com/linux/performance-issues-with-transparent-huge-pages-thp
# for non-optimized applications. As most Linux distributions set it to `enabled=always` by default,
# we recommend changing this to `enabled=madvise`. This will allow applications optimized
# for transparent hugepages to obtain the performance benefits, while preventing the
# associated problems otherwise. Also, set `transparent_hugepage=madvise` on your kernel
# command line (e.g. in /etc/default/grub) to persistently set this value.
echo "Enabling THP madvise"
echo madvise | sudo tee /sys/kernel/mm/transparent_hugepage/enabled
kind: ConfigMap
metadata:
annotations:
qbec.io/component: minio
labels:
qbec.io/application: minio
qbec.io/environment: prod
name: sysctl
namespace: minio
---
apiVersion: apps/v1
kind: DaemonSet
metadata:
annotations:
qbec.io/component: minio
labels:
app: minio-storage11
headless: minio
qbec.io/application: minio
qbec.io/environment: prod
name: minio-storage11
namespace: minio
spec:
selector:
matchLabels:
app: minio-storage11
template:
metadata:
labels:
app: minio-storage11
headless: minio
spec:
containers:
- args:
- server
- /minio/minio{1...98}
env:
- name: MINIO_PROMETHEUS_AUTH_TYPE
value: public
- name: MINIO_STORAGE_CLASS_STANDARD
value: EC:2
- name: MINIO_ACCESS_KEY
value: access_key
- name: MINIO_SECRET_KEY
value: secret_key
image: minio/minio:RELEASE.2021-05-26T00-22-46Z
livenessProbe:
httpGet:
path: /minio/health/live
port: 9000
initialDelaySeconds: 120
periodSeconds: 20
name: minio
ports:
- containerPort: 9000
name: service
securityContext:
privileged: true
terminationMessagePath: /termination-log
volumeMounts:
- mountPath: /minio
mountPropagation: HostToContainer
name: data
initContainers:
- command:
- /bin/sh
- -c
- cp -fp /scripts/mount-check.sh /minio/mount-check.sh; chmod +x /minio/mount-check.sh; exec /minio/mount-check.sh; cp -fp /scripts/sysctl.sh /minio/sysctl.sh; chmod +x sysctl.sh; exec /minio/sysctl.sh
image: busybox:1.31.1
imagePullPolicy: IfNotPresent
name: mount
securityContext:
privileged: true
terminationMessagePath: /termination-log
volumeMounts:
- mountPath: /scripts/mount-check.sh
name: mount-check
subPath: mount-check.sh
- mountPath: /scripts/sysctl.sh
name: sysctl
subPath: sysctl.sh
- mountPath: /scripts
name: scripts
- mountPath: /minio
mountPropagation: Bidirectional
name: data
- mountPath: /dev
name: dev
nodeSelector:
kubernetes.io/hostname: storage11
terminationGracePeriodSeconds: 5
volumes:
- emptyDir: {}
name: scripts
- configMap:
defaultMode: 511
name: mount-check
name: mount-check
- configMap:
defaultMode: 511
name: sysctl
name: sysctl
- hostPath:
path: /minio
name: data
- hostPath:
path: /dev
name: dev
---
apiVersion: v1
kind: Service
metadata:
annotations:
qbec.io/component: minio
labels:
app: minio-headless
qbec.io/application: minio
qbec.io/environment: prod
name: minio-headless
namespace: minio
spec:
clusterIP: None
ports:
- name: service
port: 9000
protocol: TCP
targetPort: 9000
selector:
headless: minio
---
apiVersion: v1
kind: Service
metadata:
annotations:
qbec.io/component: minio
labels:
app: minio-svc
qbec.io/application: minio
qbec.io/environment: prod
name: minio-storage11
namespace: minio
spec:
ports:
- name: service
port: 9000
protocol: TCP
targetPort: 9000
selector:
app: minio-storage11
type: ClusterIP
```
</p>
</details>
* Operating System and version (`uname -a`):
```shell
root@storage11:/# uname -a
Linux storage11 5.4.0-73-generic #82-Ubuntu SMP Wed Apr 14 17:39:42 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
root@storage11:/# cat /etc/os-release
NAME="Ubuntu"
VERSION="20.04.2 LTS (Focal Fossa)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 20.04.2 LTS"
VERSION_ID="20.04"
HOME_URL="https://www.ubuntu.com/"
SUPPORT_URL="https://help.ubuntu.com/"
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
VERSION_CODENAME=focal
UBUNTU_CODENAME=focal
``` | 1.0 | Error: io: read/write on closed pipe (*errors.errorString) in healing - <!--- Provide a general summary of the issue in the Title above -->
## Expected Behavior
Minio runs without such message in log
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
## Current Behavior
Thousands of messages like
```shell
API: SYSTEM()
Time: 12:51:27 UTC 06/01/2021
DeploymentID: 0bbd5e0d-fdb6-4bbd-b210-0c13a31d2e36
Error: io: read/write on closed pipe (*errors.errorString)
12: cmd/data-scanner.go:869:cmd.(*scannerItem).applyHealing()
11: cmd/data-scanner.go:972:cmd.(*scannerItem).applyActions()
10: cmd/xl-storage.go:459:cmd.(*xlStorage).NSScanner.func1()
9: cmd/data-scanner.go:467:cmd.(*folderScanner).scanFolder.func1()
8: cmd/os-readdir_unix.go:176:cmd.readDirFn()
7: cmd/data-scanner.go:402:cmd.(*folderScanner).scanFolder()
6: cmd/data-scanner.go:519:cmd.(*folderScanner).scanFolder.func2()
5: cmd/data-scanner.go:583:cmd.(*folderScanner).scanFolder()
4: cmd/data-scanner.go:314:cmd.scanDataFolder()
3: cmd/xl-storage.go:429:cmd.(*xlStorage).NSScanner()
2: cmd/xl-storage-disk-id-check.go:171:cmd.(*xlStorageDiskIDCheck).NSScanner()
1: cmd/erasure.go:500:cmd.erasureObjects.nsScanner.func3()
```
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## Possible Solution
Show what exactly happens, more clear description of problem
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
<!--- and make sure you have followed https://github.com/minio/minio/tree/release/docs/debugging to capture relevant logs -->
1. Update to `RELEASE.2021-05-26T00-22-46Z`
2. Run with `minio server /minio/minio{1...98}` in standalone mode
3. Use it with 1000+ buckets and 3,3M objects
## Context
There was no unusual behavior before upgrading, minio was working on `RELEASE.2021-03-26T00-00-41Z`
First time I faced with this message on `RELEASE.2021-05-22T02-34-39Z`
I suppose, that this error is not just a notice. Something bad is going inside. I'm not sure if my data in Minio is safe.
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
## Regression
<!-- Is this issue a regression? (Yes / No) -->
<!-- If Yes, optionally please include minio version or commit id or PR# that caused this regression, if you have these details. -->
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio --version`):
```shell
mc admin info storage11
● storage11.s3.domain.org
Uptime: 5 days
Version: 2021-05-26T00:22:46Z
Network: 1/1 OK
Drives: 98/98 OK
18 TiB Used, 1,194 Buckets, 3,343,906 Objects
98 drives online, 0 drives offline
```
* Server setup and configuration:
Running from simple daemonset
<details><summary>minio.yaml</summary>
<p>
```yaml
---
apiVersion: v1
data:
mount-check.sh: |
#!/bin/sh
for i in $(seq 1 99); do
if grep -qs "minio$i" /proc/mounts; then
echo "It's mounted."
else
echo "It's not mounted."
mkdir -p /minio/minio$i
mount /dev/disk/by-label/minio$i /minio/minio$i
if [ $? -eq 0 ]; then
echo "Mount minio$i success!"
else
echo "Something went wrong with the mount minio$i"
fi
fi; done
kind: ConfigMap
metadata:
annotations:
qbec.io/component: minio
labels:
qbec.io/application: minio
qbec.io/environment: prod
name: mount-check
namespace: minio
---
apiVersion: v1
data:
sysctl.sh: |
#!/bin/bash
cat > sysctl.conf <<EOF
# maximum number of open files/file descriptors
fs.file-max = 4194303
# use as little swap space as possible
vm.swappiness = 1
# prioritize application RAM against disk/swap cache
vm.vfs_cache_pressure = 10
# minimum free memory
vm.min_free_kbytes = 1000000
# maximum receive socket buffer (bytes)
net.core.rmem_max = 268435456
# maximum send buffer socket buffer (bytes)
net.core.wmem_max = 268435456
# default receive buffer socket size (bytes)
net.core.rmem_default = 67108864
# default send buffer socket size (bytes)
net.core.wmem_default = 67108864
# maximum number of packets in one poll cycle
net.core.netdev_budget = 1200
# maximum ancillary buffer size per socket
net.core.optmem_max = 134217728
# maximum number of incoming connections
net.core.somaxconn = 65535
# maximum number of packets queued
net.core.netdev_max_backlog = 250000
# maximum read buffer space
net.ipv4.tcp_rmem = 67108864 134217728 268435456
# maximum write buffer space
net.ipv4.tcp_wmem = 67108864 134217728 268435456
# enable low latency mode
net.ipv4.tcp_low_latency = 1
# socket buffer portion used for TCP window
net.ipv4.tcp_adv_win_scale = 1
# queue length of completely established sockets waiting for accept
net.ipv4.tcp_max_syn_backlog = 30000
# maximum number of sockets in TIME_WAIT state
net.ipv4.tcp_max_tw_buckets = 2000000
# reuse sockets in TIME_WAIT state when safe
net.ipv4.tcp_tw_reuse = 1
# time to wait (seconds) for FIN packet
net.ipv4.tcp_fin_timeout = 5
# disable icmp send redirects
net.ipv4.conf.all.send_redirects = 0
# disable icmp accept redirect
net.ipv4.conf.all.accept_redirects = 0
# drop packets with LSR or SSR
net.ipv4.conf.all.accept_source_route = 0
# MTU discovery, only enable when ICMP blackhole detected
net.ipv4.tcp_mtu_probing = 1
EOF
echo "Enabling system level tuning params"
sysctl --quiet --load sysctl.conf && rm -f sysctl.conf
# `Transparent Hugepage Support`*: This is a Linux kernel feature intended to improve
# performance by making more efficient use of processor’s memory-mapping hardware.
# But this may cause https://blogs.oracle.com/linux/performance-issues-with-transparent-huge-pages-thp
# for non-optimized applications. As most Linux distributions set it to `enabled=always` by default,
# we recommend changing this to `enabled=madvise`. This will allow applications optimized
# for transparent hugepages to obtain the performance benefits, while preventing the
# associated problems otherwise. Also, set `transparent_hugepage=madvise` on your kernel
# command line (e.g. in /etc/default/grub) to persistently set this value.
echo "Enabling THP madvise"
echo madvise | sudo tee /sys/kernel/mm/transparent_hugepage/enabled
kind: ConfigMap
metadata:
annotations:
qbec.io/component: minio
labels:
qbec.io/application: minio
qbec.io/environment: prod
name: sysctl
namespace: minio
---
apiVersion: apps/v1
kind: DaemonSet
metadata:
annotations:
qbec.io/component: minio
labels:
app: minio-storage11
headless: minio
qbec.io/application: minio
qbec.io/environment: prod
name: minio-storage11
namespace: minio
spec:
selector:
matchLabels:
app: minio-storage11
template:
metadata:
labels:
app: minio-storage11
headless: minio
spec:
containers:
- args:
- server
- /minio/minio{1...98}
env:
- name: MINIO_PROMETHEUS_AUTH_TYPE
value: public
- name: MINIO_STORAGE_CLASS_STANDARD
value: EC:2
- name: MINIO_ACCESS_KEY
value: access_key
- name: MINIO_SECRET_KEY
value: secret_key
image: minio/minio:RELEASE.2021-05-26T00-22-46Z
livenessProbe:
httpGet:
path: /minio/health/live
port: 9000
initialDelaySeconds: 120
periodSeconds: 20
name: minio
ports:
- containerPort: 9000
name: service
securityContext:
privileged: true
terminationMessagePath: /termination-log
volumeMounts:
- mountPath: /minio
mountPropagation: HostToContainer
name: data
initContainers:
- command:
- /bin/sh
- -c
- cp -fp /scripts/mount-check.sh /minio/mount-check.sh; chmod +x /minio/mount-check.sh; exec /minio/mount-check.sh; cp -fp /scripts/sysctl.sh /minio/sysctl.sh; chmod +x sysctl.sh; exec /minio/sysctl.sh
image: busybox:1.31.1
imagePullPolicy: IfNotPresent
name: mount
securityContext:
privileged: true
terminationMessagePath: /termination-log
volumeMounts:
- mountPath: /scripts/mount-check.sh
name: mount-check
subPath: mount-check.sh
- mountPath: /scripts/sysctl.sh
name: sysctl
subPath: sysctl.sh
- mountPath: /scripts
name: scripts
- mountPath: /minio
mountPropagation: Bidirectional
name: data
- mountPath: /dev
name: dev
nodeSelector:
kubernetes.io/hostname: storage11
terminationGracePeriodSeconds: 5
volumes:
- emptyDir: {}
name: scripts
- configMap:
defaultMode: 511
name: mount-check
name: mount-check
- configMap:
defaultMode: 511
name: sysctl
name: sysctl
- hostPath:
path: /minio
name: data
- hostPath:
path: /dev
name: dev
---
apiVersion: v1
kind: Service
metadata:
annotations:
qbec.io/component: minio
labels:
app: minio-headless
qbec.io/application: minio
qbec.io/environment: prod
name: minio-headless
namespace: minio
spec:
clusterIP: None
ports:
- name: service
port: 9000
protocol: TCP
targetPort: 9000
selector:
headless: minio
---
apiVersion: v1
kind: Service
metadata:
annotations:
qbec.io/component: minio
labels:
app: minio-svc
qbec.io/application: minio
qbec.io/environment: prod
name: minio-storage11
namespace: minio
spec:
ports:
- name: service
port: 9000
protocol: TCP
targetPort: 9000
selector:
app: minio-storage11
type: ClusterIP
```
</p>
</details>
* Operating System and version (`uname -a`):
```shell
root@storage11:/# uname -a
Linux storage11 5.4.0-73-generic #82-Ubuntu SMP Wed Apr 14 17:39:42 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
root@storage11:/# cat /etc/os-release
NAME="Ubuntu"
VERSION="20.04.2 LTS (Focal Fossa)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 20.04.2 LTS"
VERSION_ID="20.04"
HOME_URL="https://www.ubuntu.com/"
SUPPORT_URL="https://help.ubuntu.com/"
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
VERSION_CODENAME=focal
UBUNTU_CODENAME=focal
``` | priority | error io read write on closed pipe errors errorstring in healing expected behavior minio runs without such message in log current behavior thousands of messages like shell api system time utc deploymentid error io read write on closed pipe errors errorstring cmd data scanner go cmd scanneritem applyhealing cmd data scanner go cmd scanneritem applyactions cmd xl storage go cmd xlstorage nsscanner cmd data scanner go cmd folderscanner scanfolder cmd os readdir unix go cmd readdirfn cmd data scanner go cmd folderscanner scanfolder cmd data scanner go cmd folderscanner scanfolder cmd data scanner go cmd folderscanner scanfolder cmd data scanner go cmd scandatafolder cmd xl storage go cmd xlstorage nsscanner cmd xl storage disk id check go cmd xlstoragediskidcheck nsscanner cmd erasure go cmd erasureobjects nsscanner possible solution show what exactly happens more clear description of problem steps to reproduce for bugs update to release run with minio server minio minio in standalone mode use it with buckets and objects context there was no unusual behavior before upgrading minio was working on release first time i faced with this message on release i suppose that this error is not just a notice something bad is going inside i m not sure if my data in minio is safe regression your environment version used minio version shell mc admin info ● domain org uptime days version network ok drives ok tib used buckets objects drives online drives offline server setup and configuration running from simple daemonset minio yaml yaml apiversion data mount check sh bin sh for i in seq do if grep qs minio i proc mounts then echo it s mounted else echo it s not mounted mkdir p minio minio i mount dev disk by label minio i minio minio i if then echo mount minio i success else echo something went wrong with the mount minio i fi fi done kind configmap metadata annotations qbec io component minio labels qbec io application minio qbec io environment prod name mount check namespace minio apiversion data sysctl sh bin bash cat sysctl conf eof maximum number of open files file descriptors fs file max use as little swap space as possible vm swappiness prioritize application ram against disk swap cache vm vfs cache pressure minimum free memory vm min free kbytes maximum receive socket buffer bytes net core rmem max maximum send buffer socket buffer bytes net core wmem max default receive buffer socket size bytes net core rmem default default send buffer socket size bytes net core wmem default maximum number of packets in one poll cycle net core netdev budget maximum ancillary buffer size per socket net core optmem max maximum number of incoming connections net core somaxconn maximum number of packets queued net core netdev max backlog maximum read buffer space net tcp rmem maximum write buffer space net tcp wmem enable low latency mode net tcp low latency socket buffer portion used for tcp window net tcp adv win scale queue length of completely established sockets waiting for accept net tcp max syn backlog maximum number of sockets in time wait state net tcp max tw buckets reuse sockets in time wait state when safe net tcp tw reuse time to wait seconds for fin packet net tcp fin timeout disable icmp send redirects net conf all send redirects disable icmp accept redirect net conf all accept redirects drop packets with lsr or ssr net conf all accept source route mtu discovery only enable when icmp blackhole detected net tcp mtu probing eof echo enabling system level tuning params sysctl quiet load sysctl conf rm f sysctl conf transparent hugepage support this is a linux kernel feature intended to improve performance by making more efficient use of processor’s memory mapping hardware but this may cause for non optimized applications as most linux distributions set it to enabled always by default we recommend changing this to enabled madvise this will allow applications optimized for transparent hugepages to obtain the performance benefits while preventing the associated problems otherwise also set transparent hugepage madvise on your kernel command line e g in etc default grub to persistently set this value echo enabling thp madvise echo madvise sudo tee sys kernel mm transparent hugepage enabled kind configmap metadata annotations qbec io component minio labels qbec io application minio qbec io environment prod name sysctl namespace minio apiversion apps kind daemonset metadata annotations qbec io component minio labels app minio headless minio qbec io application minio qbec io environment prod name minio namespace minio spec selector matchlabels app minio template metadata labels app minio headless minio spec containers args server minio minio env name minio prometheus auth type value public name minio storage class standard value ec name minio access key value access key name minio secret key value secret key image minio minio release livenessprobe httpget path minio health live port initialdelayseconds periodseconds name minio ports containerport name service securitycontext privileged true terminationmessagepath termination log volumemounts mountpath minio mountpropagation hosttocontainer name data initcontainers command bin sh c cp fp scripts mount check sh minio mount check sh chmod x minio mount check sh exec minio mount check sh cp fp scripts sysctl sh minio sysctl sh chmod x sysctl sh exec minio sysctl sh image busybox imagepullpolicy ifnotpresent name mount securitycontext privileged true terminationmessagepath termination log volumemounts mountpath scripts mount check sh name mount check subpath mount check sh mountpath scripts sysctl sh name sysctl subpath sysctl sh mountpath scripts name scripts mountpath minio mountpropagation bidirectional name data mountpath dev name dev nodeselector kubernetes io hostname terminationgraceperiodseconds volumes emptydir name scripts configmap defaultmode name mount check name mount check configmap defaultmode name sysctl name sysctl hostpath path minio name data hostpath path dev name dev apiversion kind service metadata annotations qbec io component minio labels app minio headless qbec io application minio qbec io environment prod name minio headless namespace minio spec clusterip none ports name service port protocol tcp targetport selector headless minio apiversion kind service metadata annotations qbec io component minio labels app minio svc qbec io application minio qbec io environment prod name minio namespace minio spec ports name service port protocol tcp targetport selector app minio type clusterip operating system and version uname a shell root uname a linux generic ubuntu smp wed apr utc gnu linux root cat etc os release name ubuntu version lts focal fossa id ubuntu id like debian pretty name ubuntu lts version id home url support url bug report url privacy policy url version codename focal ubuntu codename focal | 1 |
15,174 | 2,611,431,339 | IssuesEvent | 2015-02-27 04:44:00 | OpenConceptLab/oclapi | https://api.github.com/repos/OpenConceptLab/oclapi | opened | Add purge functionality to concepts (just like mappings) | enhancement medium-priority | Assigning this to @MisterNando as FYI | 1.0 | Add purge functionality to concepts (just like mappings) - Assigning this to @MisterNando as FYI | priority | add purge functionality to concepts just like mappings assigning this to misternando as fyi | 1 |
675,163 | 23,082,079,126 | IssuesEvent | 2022-07-26 08:10:03 | canonical-web-and-design/maas-ui | https://api.github.com/repos/canonical-web-and-design/maas-ui | closed | Investigate large MAAS UI bundle size | Priority: Medium | MAAS UI bundle weighs 2.01 MB - which is larger than in version 3.1 (~1.9 MB). In the latest version we dropped Angular and we should see a significant decrease in size for that reason alone.
This seems to be caused by https://github.com/juju/bakeryjs/issues/40 but there might be more to this.
Needs investigation.
<img width="1497" alt="image" src="https://user-images.githubusercontent.com/7452681/179235637-2e0f73bf-a58d-485f-903c-5bb4e8ba338a.png">
| 1.0 | Investigate large MAAS UI bundle size - MAAS UI bundle weighs 2.01 MB - which is larger than in version 3.1 (~1.9 MB). In the latest version we dropped Angular and we should see a significant decrease in size for that reason alone.
This seems to be caused by https://github.com/juju/bakeryjs/issues/40 but there might be more to this.
Needs investigation.
<img width="1497" alt="image" src="https://user-images.githubusercontent.com/7452681/179235637-2e0f73bf-a58d-485f-903c-5bb4e8ba338a.png">
| priority | investigate large maas ui bundle size maas ui bundle weighs mb which is larger than in version mb in the latest version we dropped angular and we should see a significant decrease in size for that reason alone this seems to be caused by but there might be more to this needs investigation img width alt image src | 1 |
560,698 | 16,601,852,180 | IssuesEvent | 2021-06-01 20:40:34 | dtcenter/METplus | https://api.github.com/repos/dtcenter/METplus | closed | Improve logic of TCPairs wrapper | component: use case wrapper priority: blocker priority: medium requestor: NCAR requestor: NOAA/EMC required: FOR DEVELOPMENT RELEASE type: enhancement | There are two main things that could be improved in this wrapper.
1) Set cyclone to current cyclone if processing individually
Currently the wrapper passes in the full list of cyclones to process from the METplus config for every run. This works fine if the input files are sorted by cyclone since the input file contains only that cyclone. If the input file contains many cyclones and a list of cyclones are specified in the METplus config, the wrapper will loop over these values and process once for each. The same input file is read and the output files contain the cyclone in the filename (if included in the template), but each file will contain all cyclones listed in the config file.
If we are looping over a list of cyclones and processing once for each, the environment variable CYCLONE should only contain the current cyclone being processed, not the full list of cyclones.
2) \* added to output filename if processing all cyclones
If no cyclones are specified in the METplus config, all of the available cyclones will be processed. However, if the cyclone template is not specified in the input file template, the output file will contain * as the cyclone value. If * is set for the cyclone value in this case, it should be replaced with "all" or something similar to avoid created a problematic filename.
## Describe the Enhancement ##
See above
### Time Estimate ###
1 day
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
- [X] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
None
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [X] Select **component(s)**
- [X] Select **priority**
- [X] Select **requestor(s)**
### Projects and Milestone ###
- [X] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label
- [X] Select **milestone** to next major version milestone or "Future Versions"
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [X] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Enhancement Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [x] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [x] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues**
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 2.0 | Improve logic of TCPairs wrapper - There are two main things that could be improved in this wrapper.
1) Set cyclone to current cyclone if processing individually
Currently the wrapper passes in the full list of cyclones to process from the METplus config for every run. This works fine if the input files are sorted by cyclone since the input file contains only that cyclone. If the input file contains many cyclones and a list of cyclones are specified in the METplus config, the wrapper will loop over these values and process once for each. The same input file is read and the output files contain the cyclone in the filename (if included in the template), but each file will contain all cyclones listed in the config file.
If we are looping over a list of cyclones and processing once for each, the environment variable CYCLONE should only contain the current cyclone being processed, not the full list of cyclones.
2) \* added to output filename if processing all cyclones
If no cyclones are specified in the METplus config, all of the available cyclones will be processed. However, if the cyclone template is not specified in the input file template, the output file will contain * as the cyclone value. If * is set for the cyclone value in this case, it should be replaced with "all" or something similar to avoid created a problematic filename.
## Describe the Enhancement ##
See above
### Time Estimate ###
1 day
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
- [X] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
None
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [X] Select **component(s)**
- [X] Select **priority**
- [X] Select **requestor(s)**
### Projects and Milestone ###
- [X] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label
- [X] Select **milestone** to next major version milestone or "Future Versions"
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [X] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Enhancement Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [x] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [x] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues**
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| priority | improve logic of tcpairs wrapper there are two main things that could be improved in this wrapper set cyclone to current cyclone if processing individually currently the wrapper passes in the full list of cyclones to process from the metplus config for every run this works fine if the input files are sorted by cyclone since the input file contains only that cyclone if the input file contains many cyclones and a list of cyclones are specified in the metplus config the wrapper will loop over these values and process once for each the same input file is read and the output files contain the cyclone in the filename if included in the template but each file will contain all cyclones listed in the config file if we are looping over a list of cyclones and processing once for each the environment variable cyclone should only contain the current cyclone being processed not the full list of cyclones added to output filename if processing all cyclones if no cyclones are specified in the metplus config all of the available cyclones will be processed however if the cyclone template is not specified in the input file template the output file will contain as the cyclone value if is set for the cyclone value in this case it should be replaced with all or something similar to avoid created a problematic filename describe the enhancement see above time estimate day sub issues consider breaking the enhancement down into sub issues add a checkbox for each sub issue here relevant deadlines none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone review projects and select relevant repository and organization ones or add alert need project assignment label select milestone to next major version milestone or future versions define related issue s consider the impact to the other metplus components enhancement checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s project s milestone and linked issues iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
68,052 | 3,284,035,627 | IssuesEvent | 2015-10-28 15:12:34 | CenterForOpenScience/osf.io | https://api.github.com/repos/CenterForOpenScience/osf.io | closed | Wiki snippets appear for every word | Addon: Wiki bug: production discuss JIRA priority - medium | In the wiki editor, the snippets suggestions popup appears for virtually every word the user enters. This can be pretty annoying.
Expected:
The snippets popup only appears after entering at least a few characters of a word. | 1.0 | Wiki snippets appear for every word - In the wiki editor, the snippets suggestions popup appears for virtually every word the user enters. This can be pretty annoying.
Expected:
The snippets popup only appears after entering at least a few characters of a word. | priority | wiki snippets appear for every word in the wiki editor the snippets suggestions popup appears for virtually every word the user enters this can be pretty annoying expected the snippets popup only appears after entering at least a few characters of a word | 1 |
464,937 | 13,348,512,608 | IssuesEvent | 2020-08-29 18:55:26 | garden-io/garden | https://api.github.com/repos/garden-io/garden | closed | Garden throws wrong error message when it can't fetch the logs of a pod | bug priority:medium stale | ## Bug
### Current Behavior
After a failed deployment (due to the in-cluster builder being unable to access certain images in a registry), Garden attempts to fetch the latest logs in order to provide debugging information. It fails to catch a specific error while fetching the logs, and instead simply throws the following error:
```
Failed deleting service 'REDACTED' (from module 'REDACTED'). Here is the output:
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
HTTP request failed
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
```
Further inspection of the pod reveals it's in ImagePullBackOff state:
```
NAME READY STATUS RESTARTS AGE
REDACTED-v-fad73e89a1-58cd94587b-lqkzj 0/1 ImagePullBackOff 0 3h47m
```
### Expected behavior
Garden should accurately display the state of the pod, and default back to a different error message when it can't find anything in the logs to handle.
### Reproducible example
In a project, have one image in a separate registry that isn't authenticated against yet. Attempt to deploy the stack, which will fail and show the unhelpful error message mentioned above. Note that setting the correct secret in order to fetch from the registry at a later point will alleviate the problem (which might be a separate issue).
### Workaround
Manually deleting the namespace and re-deploying (with working authentication against the external registry) 'solved' the issue.
### Suggested solution(s)
<!-- How could we solve this bug? What changes would need to made? -->
### Additional context
<!-- Add any other context about the problem here. -->
### Your environment
<!-- PLEASE FILL THIS OUT -->
<!-- Please run and copy and paste the results -->
`garden version`
`kubectl version`
Client Version: version.Info{Major:"1", Minor:"16+", GitVersion:"v1.16.6-beta.0", GitCommit:"e7f962ba86f4ce7033828210ca3556393c377bcc", GitTreeState:"clean", BuildDate:"2020-01-15T08:26:26Z", GoVersion:"go1.13.5", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"15+", GitVersion:"v1.15.9-gke.24", GitCommit:"39e41a8d6b7221b901a95d3af358dea6994b4a40", GitTreeState:"clean", BuildDate:"2020-02-29T01:24:35Z", GoVersion:"go1.12.12b4", Compiler:"gc", Platform:"linux/amd64"}
`docker version`
Client: Docker Engine - Community
Version: 19.03.8
API version: 1.40
Go version: go1.12.17
Git commit: afacb8b
Built: Wed Mar 11 01:21:11 2020
OS/Arch: darwin/amd64
Experimental: true
Server: Docker Engine - Community
Engine:
Version: 19.03.8
API version: 1.40 (minimum version 1.12)
Go version: go1.12.17
Git commit: afacb8b
Built: Wed Mar 11 01:29:16 2020
OS/Arch: linux/amd64
Experimental: true
containerd:
Version: v1.2.13
GitCommit: 7ad184331fa3e55e52b890ea95e65ba581ae3429
runc:
Version: 1.0.0-rc10
GitCommit: dc9208a3303feef5b3839f4323d9beb36df0a9dd
docker-init:
Version: 0.18.0
GitCommit: fec3683
CC @edvald | 1.0 | Garden throws wrong error message when it can't fetch the logs of a pod - ## Bug
### Current Behavior
After a failed deployment (due to the in-cluster builder being unable to access certain images in a registry), Garden attempts to fetch the latest logs in order to provide debugging information. It fails to catch a specific error while fetching the logs, and instead simply throws the following error:
```
Failed deleting service 'REDACTED' (from module 'REDACTED'). Here is the output:
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
HTTP request failed
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
```
Further inspection of the pod reveals it's in ImagePullBackOff state:
```
NAME READY STATUS RESTARTS AGE
REDACTED-v-fad73e89a1-58cd94587b-lqkzj 0/1 ImagePullBackOff 0 3h47m
```
### Expected behavior
Garden should accurately display the state of the pod, and default back to a different error message when it can't find anything in the logs to handle.
### Reproducible example
In a project, have one image in a separate registry that isn't authenticated against yet. Attempt to deploy the stack, which will fail and show the unhelpful error message mentioned above. Note that setting the correct secret in order to fetch from the registry at a later point will alleviate the problem (which might be a separate issue).
### Workaround
Manually deleting the namespace and re-deploying (with working authentication against the external registry) 'solved' the issue.
### Suggested solution(s)
<!-- How could we solve this bug? What changes would need to made? -->
### Additional context
<!-- Add any other context about the problem here. -->
### Your environment
<!-- PLEASE FILL THIS OUT -->
<!-- Please run and copy and paste the results -->
`garden version`
`kubectl version`
Client Version: version.Info{Major:"1", Minor:"16+", GitVersion:"v1.16.6-beta.0", GitCommit:"e7f962ba86f4ce7033828210ca3556393c377bcc", GitTreeState:"clean", BuildDate:"2020-01-15T08:26:26Z", GoVersion:"go1.13.5", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"15+", GitVersion:"v1.15.9-gke.24", GitCommit:"39e41a8d6b7221b901a95d3af358dea6994b4a40", GitTreeState:"clean", BuildDate:"2020-02-29T01:24:35Z", GoVersion:"go1.12.12b4", Compiler:"gc", Platform:"linux/amd64"}
`docker version`
Client: Docker Engine - Community
Version: 19.03.8
API version: 1.40
Go version: go1.12.17
Git commit: afacb8b
Built: Wed Mar 11 01:21:11 2020
OS/Arch: darwin/amd64
Experimental: true
Server: Docker Engine - Community
Engine:
Version: 19.03.8
API version: 1.40 (minimum version 1.12)
Go version: go1.12.17
Git commit: afacb8b
Built: Wed Mar 11 01:29:16 2020
OS/Arch: linux/amd64
Experimental: true
containerd:
Version: v1.2.13
GitCommit: 7ad184331fa3e55e52b890ea95e65ba581ae3429
runc:
Version: 1.0.0-rc10
GitCommit: dc9208a3303feef5b3839f4323d9beb36df0a9dd
docker-init:
Version: 0.18.0
GitCommit: fec3683
CC @edvald | priority | garden throws wrong error message when it can t fetch the logs of a pod bug current behavior after a failed deployment due to the in cluster builder being unable to access certain images in a registry garden attempts to fetch the latest logs in order to provide debugging information it fails to catch a specific error while fetching the logs and instead simply throws the following error failed deleting service redacted from module redacted here is the output ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ http request failed ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ further inspection of the pod reveals it s in imagepullbackoff state name ready status restarts age redacted v lqkzj imagepullbackoff expected behavior garden should accurately display the state of the pod and default back to a different error message when it can t find anything in the logs to handle reproducible example in a project have one image in a separate registry that isn t authenticated against yet attempt to deploy the stack which will fail and show the unhelpful error message mentioned above note that setting the correct secret in order to fetch from the registry at a later point will alleviate the problem which might be a separate issue workaround manually deleting the namespace and re deploying with working authentication against the external registry solved the issue suggested solution s additional context your environment garden version kubectl version client version version info major minor gitversion beta gitcommit gittreestate clean builddate goversion compiler gc platform darwin server version version info major minor gitversion gke gitcommit gittreestate clean builddate goversion compiler gc platform linux docker version client docker engine community version api version go version git commit built wed mar os arch darwin experimental true server docker engine community engine version api version minimum version go version git commit built wed mar os arch linux experimental true containerd version gitcommit runc version gitcommit docker init version gitcommit cc edvald | 1 |
258,197 | 8,166,031,494 | IssuesEvent | 2018-08-25 03:16:46 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Move more static files to Webpack | help wanted priority: medium section: other | ### Description
[//]: # (Describe bug in detail here. Include screenshots if helpful.)
@paglias made the following observations in our testing document:
>the new client production file (js, css, sprites, audio, ..) are not cached by the server. Anything going through webpack can be cached as the filenames are hashed but not the sprites or audio files. Need to hash the filenames of /sprites and /audio files
>Files in dist-client/{js|img|css} are easily cacheable as they’re renamed after every change. But not the sprites and audio files as the filenames are not hashed (just copied into the dist folder without going through webpack)
If possible, we should get these files into the build process for uniform caching of static content.
| 1.0 | Move more static files to Webpack - ### Description
[//]: # (Describe bug in detail here. Include screenshots if helpful.)
@paglias made the following observations in our testing document:
>the new client production file (js, css, sprites, audio, ..) are not cached by the server. Anything going through webpack can be cached as the filenames are hashed but not the sprites or audio files. Need to hash the filenames of /sprites and /audio files
>Files in dist-client/{js|img|css} are easily cacheable as they’re renamed after every change. But not the sprites and audio files as the filenames are not hashed (just copied into the dist folder without going through webpack)
If possible, we should get these files into the build process for uniform caching of static content.
| priority | move more static files to webpack description describe bug in detail here include screenshots if helpful paglias made the following observations in our testing document the new client production file js css sprites audio are not cached by the server anything going through webpack can be cached as the filenames are hashed but not the sprites or audio files need to hash the filenames of sprites and audio files files in dist client js img css are easily cacheable as they’re renamed after every change but not the sprites and audio files as the filenames are not hashed just copied into the dist folder without going through webpack if possible we should get these files into the build process for uniform caching of static content | 1 |
750,230 | 26,194,002,022 | IssuesEvent | 2023-01-03 11:45:31 | software-mansion/starknet.py | https://api.github.com/repos/software-mansion/starknet.py | closed | Create module for hash calculation | enhancement refactor priority: medium | ### Feature Request
Currently, we use functions from the cairo lang:
- calculate_transaction_hash_common
- calculate_declare_transaction_hash
- calculate_deploy_account_transaction_hash
They should be replaced by our implementation | 1.0 | Create module for hash calculation - ### Feature Request
Currently, we use functions from the cairo lang:
- calculate_transaction_hash_common
- calculate_declare_transaction_hash
- calculate_deploy_account_transaction_hash
They should be replaced by our implementation | priority | create module for hash calculation feature request currently we use functions from the cairo lang calculate transaction hash common calculate declare transaction hash calculate deploy account transaction hash they should be replaced by our implementation | 1 |
470,599 | 13,541,289,241 | IssuesEvent | 2020-09-16 15:41:31 | NREL/EnergyPlus | https://api.github.com/repos/NREL/EnergyPlus | closed | Sizing:System with FlowPerCoolingCapacity and CoolingDesignCapacity not working as expected | PriorityLow SeverityMedium | ## Issue overview
User file Sizing:System object has these inputs for system airflow and coil capacities:
```
FlowPerCoolingCapacity, !- Cooling Supply Air Flow Rate Method
0, !- Cooling Supply Air Flow Rate {m3/s}
0.00006041, !- Cooling Supply Air Flow Rate Per Unit Cooling Capacity {m3/s-W}
FlowPerHeatingCapacity, !- Heating Supply Air Flow Rate Method
0, !- Heating Supply Air Flow Rate {m3/s}
0.00006041, !- Heating Supply Air Flow Rate Per Unit Heating Capacity {m3/s-W}
ZoneSum, !- System Outdoor Air Method
1.0, !- Zone Maximum Outdoor Air Fraction {dimensionless}
CoolingDesignCapacity, !- Cooling Design Capacity Method
12500, !- Cooling Design Capacity {W}
HeatingDesignCapacity, !- Heating Design Capacity Method
14400, !- Heating Design Capacity {W}
OnOff; !- Central Cooling Capacity Control Method
```
Expected airflow rate for this system is 0.00006041_14400=0.8699 for heating and 0.00006041_12500=0.755124 for cooling, but the eio reports:
```
System Sizing Information, AC01-02, Calculated Cooling Design Air Flow Rate [m3/s], 0.23372
System Sizing Information, AC01-02, User Cooling Design Air Flow Rate [m3/s], 0.19308
System Sizing Information, AC01-02, Calculated Heating Design Air Flow Rate [m3/s], 0.24154
System Sizing Information, AC01-02, User Heating Design Air Flow Rate [m3/s], 0.24722
```
Appears the the FlowPer...Capacity is not checking if the user also specified a hard value for heating/cooling capacity.
### Details
Some additional details for this issue (if relevant):
- Platform ()
- Version of EnergyPlus v8.5.0, didn't recheck with current develop, but expect the same
- Helpdesk ticket number11503
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [x] Defect file added (EnergyPlusDevSupport\DefectFiles)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| 1.0 | Sizing:System with FlowPerCoolingCapacity and CoolingDesignCapacity not working as expected - ## Issue overview
User file Sizing:System object has these inputs for system airflow and coil capacities:
```
FlowPerCoolingCapacity, !- Cooling Supply Air Flow Rate Method
0, !- Cooling Supply Air Flow Rate {m3/s}
0.00006041, !- Cooling Supply Air Flow Rate Per Unit Cooling Capacity {m3/s-W}
FlowPerHeatingCapacity, !- Heating Supply Air Flow Rate Method
0, !- Heating Supply Air Flow Rate {m3/s}
0.00006041, !- Heating Supply Air Flow Rate Per Unit Heating Capacity {m3/s-W}
ZoneSum, !- System Outdoor Air Method
1.0, !- Zone Maximum Outdoor Air Fraction {dimensionless}
CoolingDesignCapacity, !- Cooling Design Capacity Method
12500, !- Cooling Design Capacity {W}
HeatingDesignCapacity, !- Heating Design Capacity Method
14400, !- Heating Design Capacity {W}
OnOff; !- Central Cooling Capacity Control Method
```
Expected airflow rate for this system is 0.00006041_14400=0.8699 for heating and 0.00006041_12500=0.755124 for cooling, but the eio reports:
```
System Sizing Information, AC01-02, Calculated Cooling Design Air Flow Rate [m3/s], 0.23372
System Sizing Information, AC01-02, User Cooling Design Air Flow Rate [m3/s], 0.19308
System Sizing Information, AC01-02, Calculated Heating Design Air Flow Rate [m3/s], 0.24154
System Sizing Information, AC01-02, User Heating Design Air Flow Rate [m3/s], 0.24722
```
Appears the the FlowPer...Capacity is not checking if the user also specified a hard value for heating/cooling capacity.
### Details
Some additional details for this issue (if relevant):
- Platform ()
- Version of EnergyPlus v8.5.0, didn't recheck with current develop, but expect the same
- Helpdesk ticket number11503
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [x] Defect file added (EnergyPlusDevSupport\DefectFiles)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| priority | sizing system with flowpercoolingcapacity and coolingdesigncapacity not working as expected issue overview user file sizing system object has these inputs for system airflow and coil capacities flowpercoolingcapacity cooling supply air flow rate method cooling supply air flow rate s cooling supply air flow rate per unit cooling capacity s w flowperheatingcapacity heating supply air flow rate method heating supply air flow rate s heating supply air flow rate per unit heating capacity s w zonesum system outdoor air method zone maximum outdoor air fraction dimensionless coolingdesigncapacity cooling design capacity method cooling design capacity w heatingdesigncapacity heating design capacity method heating design capacity w onoff central cooling capacity control method expected airflow rate for this system is for heating and for cooling but the eio reports system sizing information calculated cooling design air flow rate system sizing information user cooling design air flow rate system sizing information calculated heating design air flow rate system sizing information user heating design air flow rate appears the the flowper capacity is not checking if the user also specified a hard value for heating cooling capacity details some additional details for this issue if relevant platform version of energyplus didn t recheck with current develop but expect the same helpdesk ticket checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added energyplusdevsupport defectfiles ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect | 1 |
554,378 | 16,419,174,442 | IssuesEvent | 2021-05-19 10:25:23 | ita-social-projects/TeachUA | https://api.github.com/repos/ita-social-projects/TeachUA | closed | [Додати гурток] A user can move on to the next step without filling a mandatory field 'Телефон' | Priority: Medium bug | **Environment:** Windows, Google Chrome 88.0.4324.190 (64-bit).
**Reproducible:** always.
**Build found:** last commit from https://speak-ukrainian.org.ua/dev/clubs
**Steps to reproduce**
1. Go to https://speak-ukrainian.org.ua/dev/
2. Click on 'Додати гурток' button on the main page or in user's private office
3. Fill all mandatory parameters on 'Основна інформація' step and move to 'Контакти' step
4. Leave 'Телефон' field empty and click 'Наступний крок' button
**Actual result**
The 'Опис' step is opened
**Expected result**
The 'Контакти' step should still be opened. A 'Телефон' field should become red, because it's a mandatory field
**User story and test case links**
User story #70
**Labels to be added**
"Bug", Priority ("pri: high"), Severity ("severity: medium"), Type ("Functional").
| 1.0 | [Додати гурток] A user can move on to the next step without filling a mandatory field 'Телефон' - **Environment:** Windows, Google Chrome 88.0.4324.190 (64-bit).
**Reproducible:** always.
**Build found:** last commit from https://speak-ukrainian.org.ua/dev/clubs
**Steps to reproduce**
1. Go to https://speak-ukrainian.org.ua/dev/
2. Click on 'Додати гурток' button on the main page or in user's private office
3. Fill all mandatory parameters on 'Основна інформація' step and move to 'Контакти' step
4. Leave 'Телефон' field empty and click 'Наступний крок' button
**Actual result**
The 'Опис' step is opened
**Expected result**
The 'Контакти' step should still be opened. A 'Телефон' field should become red, because it's a mandatory field
**User story and test case links**
User story #70
**Labels to be added**
"Bug", Priority ("pri: high"), Severity ("severity: medium"), Type ("Functional").
| priority | a user can move on to the next step without filling a mandatory field телефон environment windows google chrome bit reproducible always build found last commit from steps to reproduce go to click on додати гурток button on the main page or in user s private office fill all mandatory parameters on основна інформація step and move to контакти step leave телефон field empty and click наступний крок button actual result the опис step is opened expected result the контакти step should still be opened a телефон field should become red because it s a mandatory field user story and test case links user story labels to be added bug priority pri high severity severity medium type functional | 1 |
419,481 | 12,224,026,407 | IssuesEvent | 2020-05-02 20:23:19 | hochreutenerl/camp-dictionary | https://api.github.com/repos/hochreutenerl/camp-dictionary | opened | Falsche Begriffe melden | enhancement feedback medium priority medium task | Funktion um falsch übersetzte Begriffe zu melden und ev. Einen Vorschlag für eine Alternativübersetzung zu machen. | 1.0 | Falsche Begriffe melden - Funktion um falsch übersetzte Begriffe zu melden und ev. Einen Vorschlag für eine Alternativübersetzung zu machen. | priority | falsche begriffe melden funktion um falsch übersetzte begriffe zu melden und ev einen vorschlag für eine alternativübersetzung zu machen | 1 |
464 | 2,497,112,989 | IssuesEvent | 2015-01-07 01:26:32 | leonbloy/pngj | https://api.github.com/repos/leonbloy/pngj | opened | Some errors could not be fatal | bug imported Priority-Medium | _From [hjg.com.ar@gmail.com](https://code.google.com/u/hjg.com.ar@gmail.com/) on January 22, 2014 23:18:10_
In particular, if an ancillary chunk has some internal problem (this does not include its outer structure: length - but could include crc), the reader should, optionally, omit them.
_Original issue: http://code.google.com/p/pngj/issues/detail?id=28_ | 1.0 | Some errors could not be fatal - _From [hjg.com.ar@gmail.com](https://code.google.com/u/hjg.com.ar@gmail.com/) on January 22, 2014 23:18:10_
In particular, if an ancillary chunk has some internal problem (this does not include its outer structure: length - but could include crc), the reader should, optionally, omit them.
_Original issue: http://code.google.com/p/pngj/issues/detail?id=28_ | priority | some errors could not be fatal from on january in particular if an ancillary chunk has some internal problem this does not include its outer structure length but could include crc the reader should optionally omit them original issue | 1 |
268,398 | 8,406,499,590 | IssuesEvent | 2018-10-11 18:07:22 | IBM/carbon-components | https://api.github.com/repos/IBM/carbon-components | closed | UI Shell: Header Nav links font weight | priority: medium project: ui shell type: bug 🐛 | UI Shell: Type Font Weight bug.
Font weight should be 400. The Link and Dropdown both appear to have 400 when inspecting the elements but are not showing the right weight on the screen.
The dropdown seems to not be inheriting the font styles.
<img width="608" alt="screen shot 2018-09-20 at 1 30 40 pm" src="https://user-images.githubusercontent.com/12533409/45839239-6d270900-bcd9-11e8-8cb0-b9afc67f0ff5.png">
| 1.0 | UI Shell: Header Nav links font weight - UI Shell: Type Font Weight bug.
Font weight should be 400. The Link and Dropdown both appear to have 400 when inspecting the elements but are not showing the right weight on the screen.
The dropdown seems to not be inheriting the font styles.
<img width="608" alt="screen shot 2018-09-20 at 1 30 40 pm" src="https://user-images.githubusercontent.com/12533409/45839239-6d270900-bcd9-11e8-8cb0-b9afc67f0ff5.png">
| priority | ui shell header nav links font weight ui shell type font weight bug font weight should be the link and dropdown both appear to have when inspecting the elements but are not showing the right weight on the screen the dropdown seems to not be inheriting the font styles img width alt screen shot at pm src | 1 |
444,061 | 12,805,889,671 | IssuesEvent | 2020-07-03 08:25:22 | GrottoCenter/Grottocenter3 | https://api.github.com/repos/GrottoCenter/Grottocenter3 | opened | [DOCUMENT] Validation page | Priority: Medium Status: Proposal Type: Feature | Un document est soumis par n'importe quel utilisateur authentifié de Grottocenter. Une fois le document soumis, il est dans un état "en attente de validation".
Les coordinateurs / validateurs peuvent, à tout moment, consulter la liste des soumissions de documents par les utilisateurs. Ils ont alors le choix de valider ou de refuser la soumission, avec un commentaire motivant ce choix (commentaire facultatif en cas de validation, obligatoire en cas de refus).

| 1.0 | [DOCUMENT] Validation page - Un document est soumis par n'importe quel utilisateur authentifié de Grottocenter. Une fois le document soumis, il est dans un état "en attente de validation".
Les coordinateurs / validateurs peuvent, à tout moment, consulter la liste des soumissions de documents par les utilisateurs. Ils ont alors le choix de valider ou de refuser la soumission, avec un commentaire motivant ce choix (commentaire facultatif en cas de validation, obligatoire en cas de refus).

| priority | validation page un document est soumis par n importe quel utilisateur authentifié de grottocenter une fois le document soumis il est dans un état en attente de validation les coordinateurs validateurs peuvent à tout moment consulter la liste des soumissions de documents par les utilisateurs ils ont alors le choix de valider ou de refuser la soumission avec un commentaire motivant ce choix commentaire facultatif en cas de validation obligatoire en cas de refus | 1 |
241,077 | 7,808,713,886 | IssuesEvent | 2018-06-11 21:06:07 | Stivius/XiboLinuxStack | https://api.github.com/repos/Stivius/XiboLinuxStack | closed | Unit-testing | medium priority task | We need unit-tests for several modules for now:
- [x] ImageParser
- [x] AudioParser
- [x] WebViewParser
- [x] VideoParser
- [x] LayoutParser
- [x] RegionParser
- [x] CommandLineParser
- [x] MainLayout
- [x] Background
| 1.0 | Unit-testing - We need unit-tests for several modules for now:
- [x] ImageParser
- [x] AudioParser
- [x] WebViewParser
- [x] VideoParser
- [x] LayoutParser
- [x] RegionParser
- [x] CommandLineParser
- [x] MainLayout
- [x] Background
| priority | unit testing we need unit tests for several modules for now imageparser audioparser webviewparser videoparser layoutparser regionparser commandlineparser mainlayout background | 1 |
198,567 | 6,974,149,088 | IssuesEvent | 2017-12-11 23:14:58 | minio/mc | https://api.github.com/repos/minio/mc | closed | mc --insecure flag not effective with config host add cmd | priority: medium | ## Expected behaviour
mc `--insecure` flag should allow self signed certificates.
## Actual behaviour
I created a self signed certificate with domain - `domain.abc`. The certificate is configured with Nginx proxy, supposed to pass on the request to a Minio server on the same host.
Minio browser is accessible via `https://domain.abc/minio` (after ignoring the security warning). When I try to set up an alias with mc, I get the below error
```sh
$ mc --insecure config host add myminio https://domain.abc minio minio123
mc: <ERROR> Unable to initialize new config from the provided credentials Get https://domain.abc/probe-bucket-sign/?location=: x509: certificate signed by unknown authority
```
## Steps to reproduce the behaviour
- Generate self signed certificate
```sh
$ openssl genrsa -out private.key 2048
$ openssl req -new -x509 -days 3650 -key private.key -out public.crt -subj "/C=US/ST=state/L=location/O=organization/CN=domain.abc"
```
- Add certs to dir `/etc/nginx/ssl/domain.abc/`
- Configure Nginx with below config
```sh
upstream minio_servers {
server 127.0.0.1:9000;
}
server {
listen 443 ssl;
server_name domain.abc www.domain.abc;
ssl on;
ssl_certificate /etc/nginx/ssl/domain.abc/public.crt;
ssl_certificate_key /etc/nginx/ssl/domain.abc/private.key;
location / {
proxy_set_header Host $http_host;
proxy_pass http://minio_servers;
}
}
```
- Try adding mc alias
## mc version
- 2017-10-31T10:13:16Z
| 1.0 | mc --insecure flag not effective with config host add cmd - ## Expected behaviour
mc `--insecure` flag should allow self signed certificates.
## Actual behaviour
I created a self signed certificate with domain - `domain.abc`. The certificate is configured with Nginx proxy, supposed to pass on the request to a Minio server on the same host.
Minio browser is accessible via `https://domain.abc/minio` (after ignoring the security warning). When I try to set up an alias with mc, I get the below error
```sh
$ mc --insecure config host add myminio https://domain.abc minio minio123
mc: <ERROR> Unable to initialize new config from the provided credentials Get https://domain.abc/probe-bucket-sign/?location=: x509: certificate signed by unknown authority
```
## Steps to reproduce the behaviour
- Generate self signed certificate
```sh
$ openssl genrsa -out private.key 2048
$ openssl req -new -x509 -days 3650 -key private.key -out public.crt -subj "/C=US/ST=state/L=location/O=organization/CN=domain.abc"
```
- Add certs to dir `/etc/nginx/ssl/domain.abc/`
- Configure Nginx with below config
```sh
upstream minio_servers {
server 127.0.0.1:9000;
}
server {
listen 443 ssl;
server_name domain.abc www.domain.abc;
ssl on;
ssl_certificate /etc/nginx/ssl/domain.abc/public.crt;
ssl_certificate_key /etc/nginx/ssl/domain.abc/private.key;
location / {
proxy_set_header Host $http_host;
proxy_pass http://minio_servers;
}
}
```
- Try adding mc alias
## mc version
- 2017-10-31T10:13:16Z
| priority | mc insecure flag not effective with config host add cmd expected behaviour mc insecure flag should allow self signed certificates actual behaviour i created a self signed certificate with domain domain abc the certificate is configured with nginx proxy supposed to pass on the request to a minio server on the same host minio browser is accessible via after ignoring the security warning when i try to set up an alias with mc i get the below error sh mc insecure config host add myminio minio mc unable to initialize new config from the provided credentials get certificate signed by unknown authority steps to reproduce the behaviour generate self signed certificate sh openssl genrsa out private key openssl req new days key private key out public crt subj c us st state l location o organization cn domain abc add certs to dir etc nginx ssl domain abc configure nginx with below config sh upstream minio servers server server listen ssl server name domain abc ssl on ssl certificate etc nginx ssl domain abc public crt ssl certificate key etc nginx ssl domain abc private key location proxy set header host http host proxy pass try adding mc alias mc version | 1 |
113,178 | 4,544,145,123 | IssuesEvent | 2016-09-10 14:33:11 | 4-20ma/ModbusMaster | https://api.github.com/repos/4-20ma/ModbusMaster | closed | Add Code of Conduct | Priority: Medium Status: In Progress Type: Enhancement | <!----------------------------------------------------------------------------
Title - ensure the issue title is clear & concise
- QUESTIONS - describe the specific question
- BUG REPORTS - describe an activity
- FEATURE REQUESTS - describe an activity
-->
<!----------------------------------------------------------------------------
Provide the following information for all issues. Replace [brackets] and placeholder text with your responses.
(QUESTIONS, BUG REPORTS, FEATURE REQUESTS)
-->
### ModbusMaster version
0.11.0
### Arduino IDE version
1.6
### Arduino Hardware
any
### Platform Details
any
---
<!----------------------------------------------------------------------------
Provide the following for FEATURE REQUESTS. Replace [brackets] and placeholder text with your responses.
Refer to [What's in a Story?](https://dannorth.net/whats-in-a-story/)
-->
### Feature Request
#### Narrative:
<!-- Replace role, feature, benefit. -->
As a project owner
I want to contributors to have clear expectations related to interaction with other contributors
So that we foster an open and welcoming environment
| 1.0 | Add Code of Conduct - <!----------------------------------------------------------------------------
Title - ensure the issue title is clear & concise
- QUESTIONS - describe the specific question
- BUG REPORTS - describe an activity
- FEATURE REQUESTS - describe an activity
-->
<!----------------------------------------------------------------------------
Provide the following information for all issues. Replace [brackets] and placeholder text with your responses.
(QUESTIONS, BUG REPORTS, FEATURE REQUESTS)
-->
### ModbusMaster version
0.11.0
### Arduino IDE version
1.6
### Arduino Hardware
any
### Platform Details
any
---
<!----------------------------------------------------------------------------
Provide the following for FEATURE REQUESTS. Replace [brackets] and placeholder text with your responses.
Refer to [What's in a Story?](https://dannorth.net/whats-in-a-story/)
-->
### Feature Request
#### Narrative:
<!-- Replace role, feature, benefit. -->
As a project owner
I want to contributors to have clear expectations related to interaction with other contributors
So that we foster an open and welcoming environment
| priority | add code of conduct title ensure the issue title is clear concise questions describe the specific question bug reports describe an activity feature requests describe an activity provide the following information for all issues replace and placeholder text with your responses questions bug reports feature requests modbusmaster version arduino ide version arduino hardware any platform details any provide the following for feature requests replace and placeholder text with your responses refer to feature request narrative as a project owner i want to contributors to have clear expectations related to interaction with other contributors so that we foster an open and welcoming environment | 1 |
319,454 | 9,744,449,916 | IssuesEvent | 2019-06-03 07:00:00 | minio/minio-dotnet | https://api.github.com/repos/minio/minio-dotnet | closed | Disable duplicate nuget package | community priority: medium | Currently there are two packages on nuget but one i think is outdated.
https://www.nuget.org/profiles/minio
Add an info in the description or disable the package in the nuget management. For me it seems at first sight that the old package should be the new one because it is the core version. | 1.0 | Disable duplicate nuget package - Currently there are two packages on nuget but one i think is outdated.
https://www.nuget.org/profiles/minio
Add an info in the description or disable the package in the nuget management. For me it seems at first sight that the old package should be the new one because it is the core version. | priority | disable duplicate nuget package currently there are two packages on nuget but one i think is outdated add an info in the description or disable the package in the nuget management for me it seems at first sight that the old package should be the new one because it is the core version | 1 |
30,297 | 2,723,436,989 | IssuesEvent | 2015-04-14 12:37:17 | minj/foxtrick | https://api.github.com/repos/minj/foxtrick | closed | ContextMenuCopy no longer copies link text when in html mode | bug Priority-Medium starred started | E. g.

**Gives:** [playerid=388543265]
**Expected:** Jaroslav List [playerid=388543265]
| 1.0 | ContextMenuCopy no longer copies link text when in html mode - E. g.

**Gives:** [playerid=388543265]
**Expected:** Jaroslav List [playerid=388543265]
| priority | contextmenucopy no longer copies link text when in html mode e g gives expected jaroslav list | 1 |
192,330 | 6,848,745,615 | IssuesEvent | 2017-11-13 19:35:15 | osuosl/streamwebs | https://api.github.com/repos/osuosl/streamwebs | closed | Implement field notes in Soil Survey & Rip Aquatic Survey | medium priority missing data | As stated in #500, we need to implement the field notes in the Soil Survey model and then import the old data.
Examples can be seen here:
http://drupal.streamwebs.org/data/soil/tenino-court-site/11-09-2011
http://streamwebs-staging.osuosl.org/sites/tenino-court-site/soil/8/ | 1.0 | Implement field notes in Soil Survey & Rip Aquatic Survey - As stated in #500, we need to implement the field notes in the Soil Survey model and then import the old data.
Examples can be seen here:
http://drupal.streamwebs.org/data/soil/tenino-court-site/11-09-2011
http://streamwebs-staging.osuosl.org/sites/tenino-court-site/soil/8/ | priority | implement field notes in soil survey rip aquatic survey as stated in we need to implement the field notes in the soil survey model and then import the old data examples can be seen here | 1 |
40,464 | 2,868,921,264 | IssuesEvent | 2015-06-05 21:58:30 | dart-lang/pub | https://api.github.com/repos/dart-lang/pub | closed | Show package descriptions on pub.dartlang.org | bug Fixed Priority-Medium | <a href="https://github.com/munificent"><img src="https://avatars.githubusercontent.com/u/46275?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [munificent](https://github.com/munificent)**
_Originally opened as dart-lang/sdk#4966_
----
Right now, pubsite doesn't parse descriptions from packages. It should. On the page for a package version, it should show the description associated with it. On the main package page (and other pages that show packages) it should use the description from the most recent (non-prerelease) version. | 1.0 | Show package descriptions on pub.dartlang.org - <a href="https://github.com/munificent"><img src="https://avatars.githubusercontent.com/u/46275?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [munificent](https://github.com/munificent)**
_Originally opened as dart-lang/sdk#4966_
----
Right now, pubsite doesn't parse descriptions from packages. It should. On the page for a package version, it should show the description associated with it. On the main package page (and other pages that show packages) it should use the description from the most recent (non-prerelease) version. | priority | show package descriptions on pub dartlang org issue by originally opened as dart lang sdk right now pubsite doesn t parse descriptions from packages it should on the page for a package version it should show the description associated with it on the main package page and other pages that show packages it should use the description from the most recent non prerelease version | 1 |
412,842 | 12,056,915,670 | IssuesEvent | 2020-04-15 15:06:22 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Configuration: Write behind interval after failure | Module: IMap Priority: Medium Source: Community Team: Core Type: Enhancement | Hi @ahmetmircik ,
when using write-behind map stores, I'd like to have a way to configure the retry time interval after an exception in store().
Currently, when a store operation fails (e.g. due to a database which is not available), the store call is retried every 3 seconds. This might lead to huge amounts of log output.
What I would like to have is
- either some exponential back-off after a exception, e.g. retry after 2, 4, 8, 16, 32, ... seconds or
- a way to configure a much longer retry interval, e.g. several minutes.
I'm using the latest 3.3 version.
thanks and best,
Lukas
| 1.0 | Configuration: Write behind interval after failure - Hi @ahmetmircik ,
when using write-behind map stores, I'd like to have a way to configure the retry time interval after an exception in store().
Currently, when a store operation fails (e.g. due to a database which is not available), the store call is retried every 3 seconds. This might lead to huge amounts of log output.
What I would like to have is
- either some exponential back-off after a exception, e.g. retry after 2, 4, 8, 16, 32, ... seconds or
- a way to configure a much longer retry interval, e.g. several minutes.
I'm using the latest 3.3 version.
thanks and best,
Lukas
| priority | configuration write behind interval after failure hi ahmetmircik when using write behind map stores i d like to have a way to configure the retry time interval after an exception in store currently when a store operation fails e g due to a database which is not available the store call is retried every seconds this might lead to huge amounts of log output what i would like to have is either some exponential back off after a exception e g retry after seconds or a way to configure a much longer retry interval e g several minutes i m using the latest version thanks and best lukas | 1 |
35,568 | 2,790,541,588 | IssuesEvent | 2015-05-09 09:54:45 | timdown/rangy | https://api.github.com/repos/timdown/rangy | closed | CSS Class Applier problem | bug imported Priority-Medium | _From [dave11...@gmail.com](https://code.google.com/u/101095142428116831402/) on April 13, 2013 11:58:47_
What steps will reproduce the problem? 1. On http://rangy.googlecode.com/svn/trunk/demos/cssclassapplier.html page, select "played between two " (include trailing space) and click "toggle bold and red"
2. Select the word "two " (include trailing space) in the previous selection, and click "toggle italic and yellow bg"
3. Select word "teams" and click "toggle bold and red" What is the expected output? What do you see instead? I expect to see "played between " in red color and bold, "two " in red color and bold, yellow bg and italic, and "teams in red color bold only.
What happens is that "teams" gets the yellow bg and italic. What version of the product are you using? On what operating system? I'm using chrome 26.0.1410.64 m, on windows 7. Please provide any additional information below.
_Original issue: http://code.google.com/p/rangy/issues/detail?id=161_ | 1.0 | CSS Class Applier problem - _From [dave11...@gmail.com](https://code.google.com/u/101095142428116831402/) on April 13, 2013 11:58:47_
What steps will reproduce the problem? 1. On http://rangy.googlecode.com/svn/trunk/demos/cssclassapplier.html page, select "played between two " (include trailing space) and click "toggle bold and red"
2. Select the word "two " (include trailing space) in the previous selection, and click "toggle italic and yellow bg"
3. Select word "teams" and click "toggle bold and red" What is the expected output? What do you see instead? I expect to see "played between " in red color and bold, "two " in red color and bold, yellow bg and italic, and "teams in red color bold only.
What happens is that "teams" gets the yellow bg and italic. What version of the product are you using? On what operating system? I'm using chrome 26.0.1410.64 m, on windows 7. Please provide any additional information below.
_Original issue: http://code.google.com/p/rangy/issues/detail?id=161_ | priority | css class applier problem from on april what steps will reproduce the problem on page select played between two include trailing space and click toggle bold and red select the word two include trailing space in the previous selection and click toggle italic and yellow bg select word teams and click toggle bold and red what is the expected output what do you see instead i expect to see played between in red color and bold two in red color and bold yellow bg and italic and teams in red color bold only what happens is that teams gets the yellow bg and italic what version of the product are you using on what operating system i m using chrome m on windows please provide any additional information below original issue | 1 |
396,010 | 11,700,033,311 | IssuesEvent | 2020-03-06 16:42:05 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | Import fails without any further explanation when running out of space on SD Card | Bug Priority-Medium Reproduced | Originally reported on Google Code with ID 2443
```
* Import a deck with lots of media on the emulator with small SD card set.
* "Import failed" message pops up
SD card capacity low message should appear
```
Reported by `perceptualchaos2` on 2014-12-18 06:52:35
| 1.0 | Import fails without any further explanation when running out of space on SD Card - Originally reported on Google Code with ID 2443
```
* Import a deck with lots of media on the emulator with small SD card set.
* "Import failed" message pops up
SD card capacity low message should appear
```
Reported by `perceptualchaos2` on 2014-12-18 06:52:35
| priority | import fails without any further explanation when running out of space on sd card originally reported on google code with id import a deck with lots of media on the emulator with small sd card set import failed message pops up sd card capacity low message should appear reported by on | 1 |
160,885 | 6,104,656,391 | IssuesEvent | 2017-06-20 21:27:29 | dollabs/pamela | https://api.github.com/repos/dollabs/pamela | reopened | HTN/TPN generation doesn't include bounds (or other metadata) declared for primitive methods | bug Priority: Medium | In the following example, the `:bounds` declared for the primitive methods (e.g., `do-a`) aren't included in the generated TPN (generated using the `htn` action)
```
(defpclass plant []
:meta {:doc "The Plant API"}
:methods [(defpmethod do-a {:bounds [2 4]} [])
(defpmethod do-b {:bounds [3 6]} [])
(defpmethod do-c {:bounds [4 8]} [])
(defpmethod do-d [])
(defpmethod do-e {:bounds [5 10]} [])])
(defpclass infeasible-sequence [plnt]
:meta {:doc "An example of infeasible sequence of activties"}
:methods [(defpmethod start
{:doc "Simple TPN with constraints"}
[]
(sequence :bounds [10 30]
(plnt.do-a)
(plnt.do-b)
(plnt.do-c)
(plnt.do-d)
(plnt.do-e)))])
``` | 1.0 | HTN/TPN generation doesn't include bounds (or other metadata) declared for primitive methods - In the following example, the `:bounds` declared for the primitive methods (e.g., `do-a`) aren't included in the generated TPN (generated using the `htn` action)
```
(defpclass plant []
:meta {:doc "The Plant API"}
:methods [(defpmethod do-a {:bounds [2 4]} [])
(defpmethod do-b {:bounds [3 6]} [])
(defpmethod do-c {:bounds [4 8]} [])
(defpmethod do-d [])
(defpmethod do-e {:bounds [5 10]} [])])
(defpclass infeasible-sequence [plnt]
:meta {:doc "An example of infeasible sequence of activties"}
:methods [(defpmethod start
{:doc "Simple TPN with constraints"}
[]
(sequence :bounds [10 30]
(plnt.do-a)
(plnt.do-b)
(plnt.do-c)
(plnt.do-d)
(plnt.do-e)))])
``` | priority | htn tpn generation doesn t include bounds or other metadata declared for primitive methods in the following example the bounds declared for the primitive methods e g do a aren t included in the generated tpn generated using the htn action defpclass plant meta doc the plant api methods defpmethod do b bounds defpmethod do c bounds defpmethod do d defpmethod do e bounds defpclass infeasible sequence meta doc an example of infeasible sequence of activties methods defpmethod start doc simple tpn with constraints sequence bounds plnt do a plnt do b plnt do c plnt do d plnt do e | 1 |
136,010 | 5,268,040,608 | IssuesEvent | 2017-02-05 06:11:07 | pixijs/pixi.js | https://api.github.com/repos/pixijs/pixi.js | closed | Parse sprite sheet image using object | Difficulty: Easy Domain: API Plugin: Loader Priority: Medium Status: Accepting PRs Type: Enhancement Version: v4.x | Loading the JSON for a spritesheet loads the image and creates frames automatically, but when using something like Webpack, it would be great to save a roundtrip request and build the spritesheet json data into the bundle.
Unfortunately, there's no easy way to use PIXI's built in spritesheet parser to make frames from the data when the image is loaded. A built-in method along the lines of `PIXI.utils.framesFromData( jsonData, image or Texture )` would be quite helpful!
@danielberndt [built a method](https://gist.github.com/danielberndt/98dd90adb1a623fc93e3a7d7daafd675/) to do this, and I adapted it use the built in `spritesheetParser` middleware, but this is easily broken if `spritesheetParser` changes.
```
import PIXI from "pixi.js";
export default function loadFromJson(name, pathToImage, data, resolution = parseInt(data.meta.scale, 10)) {
var loader = new PIXI.loaders.Loader();
PIXI.loaders.spritesheetParser().call(loader, {
name: name,
url: pathToImage,
data: data,
isJson: true,
metadata: {}
}, function(){ console.log('next', arguments, PIXI.utils.TextureCache); });
return loader;
}
```
Usage:
```
import loadFromJson from "./loadFromJson.js";
var loader = loadFromJson(
'spritesheet',
require("../spritesheets/spritesheet.png"),
require("../spritesheets/spritesheet.json")
);
function onAssetsLoaded(){
var body = new PIXI.Sprite.fromFrame('body.png');
}
loader.load(onAssetsLoaded);
```
| 1.0 | Parse sprite sheet image using object - Loading the JSON for a spritesheet loads the image and creates frames automatically, but when using something like Webpack, it would be great to save a roundtrip request and build the spritesheet json data into the bundle.
Unfortunately, there's no easy way to use PIXI's built in spritesheet parser to make frames from the data when the image is loaded. A built-in method along the lines of `PIXI.utils.framesFromData( jsonData, image or Texture )` would be quite helpful!
@danielberndt [built a method](https://gist.github.com/danielberndt/98dd90adb1a623fc93e3a7d7daafd675/) to do this, and I adapted it use the built in `spritesheetParser` middleware, but this is easily broken if `spritesheetParser` changes.
```
import PIXI from "pixi.js";
export default function loadFromJson(name, pathToImage, data, resolution = parseInt(data.meta.scale, 10)) {
var loader = new PIXI.loaders.Loader();
PIXI.loaders.spritesheetParser().call(loader, {
name: name,
url: pathToImage,
data: data,
isJson: true,
metadata: {}
}, function(){ console.log('next', arguments, PIXI.utils.TextureCache); });
return loader;
}
```
Usage:
```
import loadFromJson from "./loadFromJson.js";
var loader = loadFromJson(
'spritesheet',
require("../spritesheets/spritesheet.png"),
require("../spritesheets/spritesheet.json")
);
function onAssetsLoaded(){
var body = new PIXI.Sprite.fromFrame('body.png');
}
loader.load(onAssetsLoaded);
```
| priority | parse sprite sheet image using object loading the json for a spritesheet loads the image and creates frames automatically but when using something like webpack it would be great to save a roundtrip request and build the spritesheet json data into the bundle unfortunately there s no easy way to use pixi s built in spritesheet parser to make frames from the data when the image is loaded a built in method along the lines of pixi utils framesfromdata jsondata image or texture would be quite helpful danielberndt to do this and i adapted it use the built in spritesheetparser middleware but this is easily broken if spritesheetparser changes import pixi from pixi js export default function loadfromjson name pathtoimage data resolution parseint data meta scale var loader new pixi loaders loader pixi loaders spritesheetparser call loader name name url pathtoimage data data isjson true metadata function console log next arguments pixi utils texturecache return loader usage import loadfromjson from loadfromjson js var loader loadfromjson spritesheet require spritesheets spritesheet png require spritesheets spritesheet json function onassetsloaded var body new pixi sprite fromframe body png loader load onassetsloaded | 1 |
362,413 | 10,727,208,842 | IssuesEvent | 2019-10-28 11:07:30 | code4romania/mon-vot-android-kotlin | https://api.github.com/repos/code4romania/mon-vot-android-kotlin | closed | Implement buttons states | enhancement good first issue hacktoberfest help wanted kotlin medium-priority | - Choose polling station screen: 'continue' button should be enabled only if the county and polling station number are filled in
- Polling station details screen: 'continue to forms' button should be enabled only if the area, president and arrival time info are filled in
- Notes Screen: the send note button should be enabled if the user added some text

complete UI in figma: https://www.figma.com/file/21x1ui3YZJrGnpZNVnEuuQ/DRAFT-MV---mobile
| 1.0 | Implement buttons states - - Choose polling station screen: 'continue' button should be enabled only if the county and polling station number are filled in
- Polling station details screen: 'continue to forms' button should be enabled only if the area, president and arrival time info are filled in
- Notes Screen: the send note button should be enabled if the user added some text

complete UI in figma: https://www.figma.com/file/21x1ui3YZJrGnpZNVnEuuQ/DRAFT-MV---mobile
| priority | implement buttons states choose polling station screen continue button should be enabled only if the county and polling station number are filled in polling station details screen continue to forms button should be enabled only if the area president and arrival time info are filled in notes screen the send note button should be enabled if the user added some text complete ui in figma | 1 |
148,537 | 5,687,861,964 | IssuesEvent | 2017-04-14 00:32:54 | minetest-mods/technic | https://api.github.com/repos/minetest-mods/technic | closed | Nuclear fuel lasts too long | Medium priority | Nuclear fuel is supposed to last a week according to the manual. I assume this is 1 real-time week, not 1 in-game week. Either way, I have always had nuclear fuel last for a month or more, so something seems to be a bit broken there.
| 1.0 | Nuclear fuel lasts too long - Nuclear fuel is supposed to last a week according to the manual. I assume this is 1 real-time week, not 1 in-game week. Either way, I have always had nuclear fuel last for a month or more, so something seems to be a bit broken there.
| priority | nuclear fuel lasts too long nuclear fuel is supposed to last a week according to the manual i assume this is real time week not in game week either way i have always had nuclear fuel last for a month or more so something seems to be a bit broken there | 1 |
350,274 | 10,481,587,762 | IssuesEvent | 2019-09-24 09:59:50 | project-march/gait-generation | https://api.github.com/repos/project-march/gait-generation | closed | Preview doesn't work after the latest setpoint | priority-medium | The visualization doesn't work after the last setpoint for each joint. The gait is only played till the moment of the last setpoint. | 1.0 | Preview doesn't work after the latest setpoint - The visualization doesn't work after the last setpoint for each joint. The gait is only played till the moment of the last setpoint. | priority | preview doesn t work after the latest setpoint the visualization doesn t work after the last setpoint for each joint the gait is only played till the moment of the last setpoint | 1 |
193,649 | 6,886,699,356 | IssuesEvent | 2017-11-21 20:26:56 | unfoldingWord-dev/translationCore | https://api.github.com/repos/unfoldingWord-dev/translationCore | closed | Search for online projects does not return results | Epic Kind/Bug Priority/Medium QA/Pass | .beta63
There are multiple French projects for Ephesians on Doo43, but tC does not return any results when performing that search. | 1.0 | Search for online projects does not return results - .beta63
There are multiple French projects for Ephesians on Doo43, but tC does not return any results when performing that search. | priority | search for online projects does not return results there are multiple french projects for ephesians on but tc does not return any results when performing that search | 1 |
20,643 | 2,622,855,543 | IssuesEvent | 2015-03-04 08:07:32 | max99x/pagemon-chrome-ext | https://api.github.com/repos/max99x/pagemon-chrome-ext | opened | Change detected, but page with changes does not load | auto-migrated Priority-Medium | ```
Added http://www.knkv.nl/competitie/tuchtzaken/Tuchtuitspraken/ to be monitored.
Get message that page has changed, but if I want to see the changes, the page
does not load.
Chrome Version 40.0.2214.115 m, Windows 7 operating system.
Can't find the next info. (What are the error log messages? To view them, go to
Wrench -> Tools -> Extensions -> Developer Mode -> expand Page Monitor ->
"background.htm" -> Console.
```
Original issue reported on code.google.com by `japo...@gmail.com` on 23 Feb 2015 at 7:14 | 1.0 | Change detected, but page with changes does not load - ```
Added http://www.knkv.nl/competitie/tuchtzaken/Tuchtuitspraken/ to be monitored.
Get message that page has changed, but if I want to see the changes, the page
does not load.
Chrome Version 40.0.2214.115 m, Windows 7 operating system.
Can't find the next info. (What are the error log messages? To view them, go to
Wrench -> Tools -> Extensions -> Developer Mode -> expand Page Monitor ->
"background.htm" -> Console.
```
Original issue reported on code.google.com by `japo...@gmail.com` on 23 Feb 2015 at 7:14 | priority | change detected but page with changes does not load added to be monitored get message that page has changed but if i want to see the changes the page does not load chrome version m windows operating system can t find the next info what are the error log messages to view them go to wrench tools extensions developer mode expand page monitor background htm console original issue reported on code google com by japo gmail com on feb at | 1 |
91,945 | 3,863,517,822 | IssuesEvent | 2016-04-08 09:45:57 | iamxavier/elmah | https://api.github.com/repos/iamxavier/elmah | closed | Strong name the assembly | auto-migrated Priority-Medium Type-Enhancement | ```
What new or enhanced feature are you proposing?
Strong naming the assembly.
What goal would this enhancement help you achieve?
As is, it can not be linked into strong named projects without a custom
re-compile with a strong name added. The key needn't be secure, just
present so that the compile doesn't complain about a non-strong named
assembly being linked in.
```
Original issue reported on code.google.com by `BrennanA...@gmail.com` on 1 Jun 2009 at 2:42 | 1.0 | Strong name the assembly - ```
What new or enhanced feature are you proposing?
Strong naming the assembly.
What goal would this enhancement help you achieve?
As is, it can not be linked into strong named projects without a custom
re-compile with a strong name added. The key needn't be secure, just
present so that the compile doesn't complain about a non-strong named
assembly being linked in.
```
Original issue reported on code.google.com by `BrennanA...@gmail.com` on 1 Jun 2009 at 2:42 | priority | strong name the assembly what new or enhanced feature are you proposing strong naming the assembly what goal would this enhancement help you achieve as is it can not be linked into strong named projects without a custom re compile with a strong name added the key needn t be secure just present so that the compile doesn t complain about a non strong named assembly being linked in original issue reported on code google com by brennana gmail com on jun at | 1 |
747,927 | 26,102,244,220 | IssuesEvent | 2022-12-27 08:42:38 | bounswe/bounswe2022group2 | https://api.github.com/repos/bounswe/bounswe2022group2 | closed | Learning space reccomendation endpoint | priority-medium Status: Completed back-end | ### Issue Description
We are to show reccomended learining spaces to the user. This issue is about implementation of a GET endpoint regarding reccomended learning spaces.
### Step Details
Steps that will be performed:
- [x] Implement the semantic server logic
- [x] Implement the backend server endpoint
- [x] Add the endpoint to the doccumentation
### Final Actions
Upon this endpoints making to the deployment, it should be utilized to show recommended learning spaces.
### Deadline of the Issue
26.12.2022 21:00
### Reviewer
Ecenur Sezer
### Deadline for the Review
26.12.2022 24:00 | 1.0 | Learning space reccomendation endpoint - ### Issue Description
We are to show reccomended learining spaces to the user. This issue is about implementation of a GET endpoint regarding reccomended learning spaces.
### Step Details
Steps that will be performed:
- [x] Implement the semantic server logic
- [x] Implement the backend server endpoint
- [x] Add the endpoint to the doccumentation
### Final Actions
Upon this endpoints making to the deployment, it should be utilized to show recommended learning spaces.
### Deadline of the Issue
26.12.2022 21:00
### Reviewer
Ecenur Sezer
### Deadline for the Review
26.12.2022 24:00 | priority | learning space reccomendation endpoint issue description we are to show reccomended learining spaces to the user this issue is about implementation of a get endpoint regarding reccomended learning spaces step details steps that will be performed implement the semantic server logic implement the backend server endpoint add the endpoint to the doccumentation final actions upon this endpoints making to the deployment it should be utilized to show recommended learning spaces deadline of the issue reviewer ecenur sezer deadline for the review | 1 |
397,167 | 11,724,441,432 | IssuesEvent | 2020-03-10 10:58:03 | level73/membernet | https://api.github.com/repos/level73/membernet | closed | good practices not linking to Platformpage | Platform: Membernet Priority: Medium Type: Enhancement | for NES ALBANIA we have a good practice, but i dont know whether its not showing up on the front because im missing some tag, or because they are not synched. I tagged the country and everything in the back end in manage good practices but perhaps we need to add the platform tag to the back end of that entity as well? please let me know. thanks | 1.0 | good practices not linking to Platformpage - for NES ALBANIA we have a good practice, but i dont know whether its not showing up on the front because im missing some tag, or because they are not synched. I tagged the country and everything in the back end in manage good practices but perhaps we need to add the platform tag to the back end of that entity as well? please let me know. thanks | priority | good practices not linking to platformpage for nes albania we have a good practice but i dont know whether its not showing up on the front because im missing some tag or because they are not synched i tagged the country and everything in the back end in manage good practices but perhaps we need to add the platform tag to the back end of that entity as well please let me know thanks | 1 |
658,080 | 21,877,718,068 | IssuesEvent | 2022-05-19 11:46:49 | unfoldingWord/tc-create-app | https://api.github.com/repos/unfoldingWord/tc-create-app | closed | Turn off Preview editing in Markdown files. | Priority/Medium | Issue:
---
Users are having trouble with editing in preview mode. It should be view only. The default view should be markdown.
DoD:
---
The user should not be allowed to edit markdown files in preview mode. | 1.0 | Turn off Preview editing in Markdown files. - Issue:
---
Users are having trouble with editing in preview mode. It should be view only. The default view should be markdown.
DoD:
---
The user should not be allowed to edit markdown files in preview mode. | priority | turn off preview editing in markdown files issue users are having trouble with editing in preview mode it should be view only the default view should be markdown dod the user should not be allowed to edit markdown files in preview mode | 1 |
148,871 | 5,697,588,384 | IssuesEvent | 2017-04-16 23:14:00 | ncssar/buckshot | https://api.github.com/repos/ncssar/buckshot | closed | mark the "exact match"; rank the possibilities | enhancement help wanted Priority:Medium | It would be very helpful to have a quick visual key as to which possibility is an "exact" match based on what you were told; this would require having some punctuation in the input string, and making assumptions based on the punctuation. It may be that there is not enough information in the punctuation to make an "exact" match.
Example:
"three nine point two six five four one one by one two zero point three five two one seven six"
The string you would enter, with punctuation:
39.265411x120.352176
That should be clear enough for buckshot to determine an "exact" match. This could be indicated by different font, bold font, a colored marker next to the list item, etc; and should also be indicated by something unique about that marker in sarsoft and in the gpx. Even when the exact match is easily identified, it would be good to rank the next tier of possibilities based on similarity / likelihood.
Example:
"three nine by two six five point four one point one by one two zero three five two one seven six"
This grammar is jibberish, and is not unlikely. Based on that jibberish, the string you would enter:
39x265.41.1x120352176
buckshot would know to not mark any "exact match". But, it could still provide some type of visual ranking scheme to the possibilities (maybe just the most likely few possibilities) based on whether zeros had to be inserted or other criteria - generally based on how much the possibility differs from the entered jibberish.
This issue is fairly open-ended: whoever tackles it would need to come up with the ranking concept, criteria, and algorithm, as well as coding it up. | 1.0 | mark the "exact match"; rank the possibilities - It would be very helpful to have a quick visual key as to which possibility is an "exact" match based on what you were told; this would require having some punctuation in the input string, and making assumptions based on the punctuation. It may be that there is not enough information in the punctuation to make an "exact" match.
Example:
"three nine point two six five four one one by one two zero point three five two one seven six"
The string you would enter, with punctuation:
39.265411x120.352176
That should be clear enough for buckshot to determine an "exact" match. This could be indicated by different font, bold font, a colored marker next to the list item, etc; and should also be indicated by something unique about that marker in sarsoft and in the gpx. Even when the exact match is easily identified, it would be good to rank the next tier of possibilities based on similarity / likelihood.
Example:
"three nine by two six five point four one point one by one two zero three five two one seven six"
This grammar is jibberish, and is not unlikely. Based on that jibberish, the string you would enter:
39x265.41.1x120352176
buckshot would know to not mark any "exact match". But, it could still provide some type of visual ranking scheme to the possibilities (maybe just the most likely few possibilities) based on whether zeros had to be inserted or other criteria - generally based on how much the possibility differs from the entered jibberish.
This issue is fairly open-ended: whoever tackles it would need to come up with the ranking concept, criteria, and algorithm, as well as coding it up. | priority | mark the exact match rank the possibilities it would be very helpful to have a quick visual key as to which possibility is an exact match based on what you were told this would require having some punctuation in the input string and making assumptions based on the punctuation it may be that there is not enough information in the punctuation to make an exact match example three nine point two six five four one one by one two zero point three five two one seven six the string you would enter with punctuation that should be clear enough for buckshot to determine an exact match this could be indicated by different font bold font a colored marker next to the list item etc and should also be indicated by something unique about that marker in sarsoft and in the gpx even when the exact match is easily identified it would be good to rank the next tier of possibilities based on similarity likelihood example three nine by two six five point four one point one by one two zero three five two one seven six this grammar is jibberish and is not unlikely based on that jibberish the string you would enter buckshot would know to not mark any exact match but it could still provide some type of visual ranking scheme to the possibilities maybe just the most likely few possibilities based on whether zeros had to be inserted or other criteria generally based on how much the possibility differs from the entered jibberish this issue is fairly open ended whoever tackles it would need to come up with the ranking concept criteria and algorithm as well as coding it up | 1 |
274,062 | 8,556,602,572 | IssuesEvent | 2018-11-08 13:40:10 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | closed | Unable to create application | SEV3-medium area/launcher priority/P4 team/app-creation type/bug | I am trying to create an application on openshift.io and every time getting message "**There was a problem fetching results, please try again**"
1. Select space
2. Click on create an application.
3. Select mission as REST Api Level 0
4. Select run time as Spring Boot
5. Select Red Hat suggested pipeline
6. Click on Select application.
I repeated this multiple time and observed that every time when I was putting underscore in app name issue was reproducible. e.g demo_app_1
But when I tried same thing with a name like demo-app-1 worked fine.
| 1.0 | Unable to create application - I am trying to create an application on openshift.io and every time getting message "**There was a problem fetching results, please try again**"
1. Select space
2. Click on create an application.
3. Select mission as REST Api Level 0
4. Select run time as Spring Boot
5. Select Red Hat suggested pipeline
6. Click on Select application.
I repeated this multiple time and observed that every time when I was putting underscore in app name issue was reproducible. e.g demo_app_1
But when I tried same thing with a name like demo-app-1 worked fine.
| priority | unable to create application i am trying to create an application on openshift io and every time getting message there was a problem fetching results please try again select space click on create an application select mission as rest api level select run time as spring boot select red hat suggested pipeline click on select application i repeated this multiple time and observed that every time when i was putting underscore in app name issue was reproducible e g demo app but when i tried same thing with a name like demo app worked fine | 1 |
373,583 | 11,045,780,558 | IssuesEvent | 2019-12-09 15:42:26 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | [Coverity CID :205651]Uninitialized variables in /drivers/dma/dma_stm32.c | Coverity area: DMA area: Drivers bug has-pr priority: medium | Static code scan issues seen in File: /drivers/dma/dma_stm32.c
Category: Uninitialized variables
Function: dma_stm32_configure
Component: Drivers
CID: 205651
Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996 | 1.0 | [Coverity CID :205651]Uninitialized variables in /drivers/dma/dma_stm32.c - Static code scan issues seen in File: /drivers/dma/dma_stm32.c
Category: Uninitialized variables
Function: dma_stm32_configure
Component: Drivers
CID: 205651
Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996 | priority | uninitialized variables in drivers dma dma c static code scan issues seen in file drivers dma dma c category uninitialized variables function dma configure component drivers cid please fix or provide comments to square it off in coverity in the link | 1 |
157,978 | 6,019,808,124 | IssuesEvent | 2017-06-07 15:12:26 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | opened | App Canvas | Enable a "per application" out of service mode | bug-medium-priority tech-improvement | Sometimes only one part of *Caseflow* will have an issue. But others are just fine, so we need a way to disable one applications, while others function:
1. Break Dispatch
1. Certification is fine
1. Get scared because we need to take Dispatch out of service
1. Not want to stop Certifications from happening
1. Get sad.
- Tagged all the tech leads for visibility.
## Related Issue
https://github.com/department-of-veterans-affairs/caseflow/issues/2211
| 1.0 | App Canvas | Enable a "per application" out of service mode - Sometimes only one part of *Caseflow* will have an issue. But others are just fine, so we need a way to disable one applications, while others function:
1. Break Dispatch
1. Certification is fine
1. Get scared because we need to take Dispatch out of service
1. Not want to stop Certifications from happening
1. Get sad.
- Tagged all the tech leads for visibility.
## Related Issue
https://github.com/department-of-veterans-affairs/caseflow/issues/2211
| priority | app canvas enable a per application out of service mode sometimes only one part of caseflow will have an issue but others are just fine so we need a way to disable one applications while others function break dispatch certification is fine get scared because we need to take dispatch out of service not want to stop certifications from happening get sad tagged all the tech leads for visibility related issue | 1 |
754,357 | 26,383,686,239 | IssuesEvent | 2023-01-12 10:34:31 | calcom/cal.com | https://api.github.com/repos/calcom/cal.com | closed | [CAL-655] locale of date/month/day in the confirmation email is not in the receiver's locale | 🐛 bug Medium priority | 
that should be in en_US even if the person sending the invite does not have their locale set to en_US
<sub>From [SyncLinear.com](https://synclinear.com) | [CAL-655](https://linear.app/calcom/issue/CAL-655/locale-of-datemonthday-in-the-confirmation-email-is-not-in-the)</sub> | 1.0 | [CAL-655] locale of date/month/day in the confirmation email is not in the receiver's locale - 
that should be in en_US even if the person sending the invite does not have their locale set to en_US
<sub>From [SyncLinear.com](https://synclinear.com) | [CAL-655](https://linear.app/calcom/issue/CAL-655/locale-of-datemonthday-in-the-confirmation-email-is-not-in-the)</sub> | priority | locale of date month day in the confirmation email is not in the receiver s locale that should be in en us even if the person sending the invite does not have their locale set to en us from | 1 |
687,355 | 23,523,351,542 | IssuesEvent | 2022-08-19 08:27:12 | ansible-collections/azure | https://api.github.com/repos/ansible-collections/azure | closed | Feature Request: Virtual Wan Network Connection (vHub Network Connection) | new_module_issue has_pr medium_priority | ##### SUMMARY
Create module to allow Virtual WAN Network Connections to the Virtual WAN Hub.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
azure_rm_virtualwan_connection
##### ADDITIONAL INFORMATION
Here's how to do it in Powershell
Select-AzSubscription -SubscriptionId "{{azr_remote_subid}}"
$remote = Get-AzVirtualNetwork -Name "{{azr_remote_vnet_name}}" -ResourceGroupName "{{azr_remote_rg_name}}"
Select-AzSubscription -SubscriptionId "{{azr_subid}}"
New-AzVirtualHubVnetConnection -ResourceGroupName "{{azr_rg_name}}" -VirtualHubName "{{azr_vhub_name}}" -Name "con-{{azr_remote_vnet_name}}" -RemoteVirtualNetwork $remote
Ansible:
- name: Create a Virtual Wan Network Connection to a Virtual Hub (Like vNET Peering but different)
azure_rm_vwan_connection:
subscription_id: "{{azr_subid}}"
resource_group: "{{azr_rg_name}}"
virtual_network_name: "{{azr_vnet_name}}"
name: "{azr_vwan_con_name}}"
virtual_wan_name: "{{azr_vwan_name"
remote_network_id: "/subscriptions/{{azr_subid}}/resourceGroups/{{azr_rg_name}}/providers/Microsoft.Network/virtualNetworks/{{azr_remote_vnet_name}}"
#Optional Items
associated_route_table_id:
"/subscriptions/MySub/resourceGroups/MyRG/providers/Microsoft.Network/virtualHubs/MyHub/hubRouteTables/RouteTable1"
propagated_route_tables_ids:
-
-
"/subscriptions/MySub/resourceGroups/MyRG/providers/Microsoft.Network/virtualHubs/MyHub/hubRouteTables/RouteTable1"
-
-
"/subscriptions/MySub/resourceGroups/MyRG/providers/Microsoft.Network/virtualHubs/MyHub/hubRouteTables/RouteTable2"
labels:
- "label1"
- "label2"
route_name: "route1"
next_hop: "70.0.0.2"
address_prefixes:
- "10.80.0.0/16"
- "10.90.0.0/16"
| 1.0 | Feature Request: Virtual Wan Network Connection (vHub Network Connection) - ##### SUMMARY
Create module to allow Virtual WAN Network Connections to the Virtual WAN Hub.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
azure_rm_virtualwan_connection
##### ADDITIONAL INFORMATION
Here's how to do it in Powershell
Select-AzSubscription -SubscriptionId "{{azr_remote_subid}}"
$remote = Get-AzVirtualNetwork -Name "{{azr_remote_vnet_name}}" -ResourceGroupName "{{azr_remote_rg_name}}"
Select-AzSubscription -SubscriptionId "{{azr_subid}}"
New-AzVirtualHubVnetConnection -ResourceGroupName "{{azr_rg_name}}" -VirtualHubName "{{azr_vhub_name}}" -Name "con-{{azr_remote_vnet_name}}" -RemoteVirtualNetwork $remote
Ansible:
- name: Create a Virtual Wan Network Connection to a Virtual Hub (Like vNET Peering but different)
azure_rm_vwan_connection:
subscription_id: "{{azr_subid}}"
resource_group: "{{azr_rg_name}}"
virtual_network_name: "{{azr_vnet_name}}"
name: "{azr_vwan_con_name}}"
virtual_wan_name: "{{azr_vwan_name"
remote_network_id: "/subscriptions/{{azr_subid}}/resourceGroups/{{azr_rg_name}}/providers/Microsoft.Network/virtualNetworks/{{azr_remote_vnet_name}}"
#Optional Items
associated_route_table_id:
"/subscriptions/MySub/resourceGroups/MyRG/providers/Microsoft.Network/virtualHubs/MyHub/hubRouteTables/RouteTable1"
propagated_route_tables_ids:
-
-
"/subscriptions/MySub/resourceGroups/MyRG/providers/Microsoft.Network/virtualHubs/MyHub/hubRouteTables/RouteTable1"
-
-
"/subscriptions/MySub/resourceGroups/MyRG/providers/Microsoft.Network/virtualHubs/MyHub/hubRouteTables/RouteTable2"
labels:
- "label1"
- "label2"
route_name: "route1"
next_hop: "70.0.0.2"
address_prefixes:
- "10.80.0.0/16"
- "10.90.0.0/16"
| priority | feature request virtual wan network connection vhub network connection summary create module to allow virtual wan network connections to the virtual wan hub issue type feature idea component name azure rm virtualwan connection additional information here s how to do it in powershell select azsubscription subscriptionid azr remote subid remote get azvirtualnetwork name azr remote vnet name resourcegroupname azr remote rg name select azsubscription subscriptionid azr subid new azvirtualhubvnetconnection resourcegroupname azr rg name virtualhubname azr vhub name name con azr remote vnet name remotevirtualnetwork remote ansible name create a virtual wan network connection to a virtual hub like vnet peering but different azure rm vwan connection subscription id azr subid resource group azr rg name virtual network name azr vnet name name azr vwan con name virtual wan name azr vwan name remote network id subscriptions azr subid resourcegroups azr rg name providers microsoft network virtualnetworks azr remote vnet name optional items associated route table id subscriptions mysub resourcegroups myrg providers microsoft network virtualhubs myhub hubroutetables propagated route tables ids subscriptions mysub resourcegroups myrg providers microsoft network virtualhubs myhub hubroutetables subscriptions mysub resourcegroups myrg providers microsoft network virtualhubs myhub hubroutetables labels route name next hop address prefixes | 1 |
166,846 | 6,312,986,368 | IssuesEvent | 2017-07-24 06:34:23 | vmware/harbor | https://api.github.com/repos/vmware/harbor | closed | Cannot scan image by search a image | area/ui priority/medium target/1.2.0 |
when view an image by search it from the search box, there is no vulnerability column.
https://192.168.1.210/harbor/tags/2/myvv%2Falanbueno8%2Fdocker-android-atom-x86_64-6.0_non-interactive
| 1.0 | Cannot scan image by search a image -
when view an image by search it from the search box, there is no vulnerability column.
https://192.168.1.210/harbor/tags/2/myvv%2Falanbueno8%2Fdocker-android-atom-x86_64-6.0_non-interactive
| priority | cannot scan image by search a image when view an image by search it from the search box there is no vulnerability column | 1 |
148,610 | 5,693,505,712 | IssuesEvent | 2017-04-15 02:13:08 | OperationCode/operationcode | https://api.github.com/repos/OperationCode/operationcode | closed | Links within README.md's "Setting up your environment" section | Needs: More Detail Priority: Medium Type: Maintenance | As a VISITOR,
When I open [readme.md], I find trouble setting up the environment because certain links pointing to the location of required software is not up and running. Can we have an alternative for that, like a different method to set up the required environment.
It should make it easy for anyone who is coming to the repository for the first time and wants to get started with the code. | 1.0 | Links within README.md's "Setting up your environment" section - As a VISITOR,
When I open [readme.md], I find trouble setting up the environment because certain links pointing to the location of required software is not up and running. Can we have an alternative for that, like a different method to set up the required environment.
It should make it easy for anyone who is coming to the repository for the first time and wants to get started with the code. | priority | links within readme md s setting up your environment section as a visitor when i open i find trouble setting up the environment because certain links pointing to the location of required software is not up and running can we have an alternative for that like a different method to set up the required environment it should make it easy for anyone who is coming to the repository for the first time and wants to get started with the code | 1 |
163,569 | 6,201,011,043 | IssuesEvent | 2017-07-06 03:53:37 | vmware/vic | https://api.github.com/repos/vmware/vic | closed | Upgrade Angular and Clarity | area/ui priority/medium | **User Statement:**
As a UI engineer I'd like to upgrade Angular 2.4 to the latest Angular using Angular CLI, as the Clarity Wizard component has been released for 0.9.5 and 0.9.x requires that Angular 4.0.0 or higher be used.
**Acceptance Criteria:**
1. App works the same as before
2. Build process works the same as before | 1.0 | Upgrade Angular and Clarity - **User Statement:**
As a UI engineer I'd like to upgrade Angular 2.4 to the latest Angular using Angular CLI, as the Clarity Wizard component has been released for 0.9.5 and 0.9.x requires that Angular 4.0.0 or higher be used.
**Acceptance Criteria:**
1. App works the same as before
2. Build process works the same as before | priority | upgrade angular and clarity user statement as a ui engineer i d like to upgrade angular to the latest angular using angular cli as the clarity wizard component has been released for and x requires that angular or higher be used acceptance criteria app works the same as before build process works the same as before | 1 |
93,911 | 3,916,760,444 | IssuesEvent | 2016-04-21 04:02:21 | dmusican/Elegit | https://api.github.com/repos/dmusican/Elegit | closed | Support of tracking new remote branches | enhancement priority medium | It seems that currently if someone else pushes a new branch to a remote repo (which you've cloned already), there is no way of showing/tracking that new branch in Elegit. | 1.0 | Support of tracking new remote branches - It seems that currently if someone else pushes a new branch to a remote repo (which you've cloned already), there is no way of showing/tracking that new branch in Elegit. | priority | support of tracking new remote branches it seems that currently if someone else pushes a new branch to a remote repo which you ve cloned already there is no way of showing tracking that new branch in elegit | 1 |
165,944 | 6,288,575,363 | IssuesEvent | 2017-07-19 17:17:12 | ESAPI/esapi-java-legacy | https://api.github.com/repos/ESAPI/esapi-java-legacy | closed | Difference between encodeForHTMLAttribute and encodeForHTML | bug imported Priority-Medium | _From [wettstei...@gmail.com](https://code.google.com/u/108417551973747153004/) on July 18, 2012 06:31:53_
As I have seen in the source code the single difference between
ESAPI.encoder().encodeForHTML() and ESAPI.encoder().encodeForHTMLAttribute()
is that ' ' is part of the whitelist in encodeForHTML() and therefore not encoded.
ESAPI.encoder().encodeForHTMLAttribute() does indeed neither encode ' ' as it is part of the UNENCODED_SET (line 73 HTMLEntityCode).
Since the two methods do apparently the same (or did I miss some point?), why is there a need to make a distinction?
_Original issue: http://code.google.com/p/owasp-esapi-java/issues/detail?id=276_
| 1.0 | Difference between encodeForHTMLAttribute and encodeForHTML - _From [wettstei...@gmail.com](https://code.google.com/u/108417551973747153004/) on July 18, 2012 06:31:53_
As I have seen in the source code the single difference between
ESAPI.encoder().encodeForHTML() and ESAPI.encoder().encodeForHTMLAttribute()
is that ' ' is part of the whitelist in encodeForHTML() and therefore not encoded.
ESAPI.encoder().encodeForHTMLAttribute() does indeed neither encode ' ' as it is part of the UNENCODED_SET (line 73 HTMLEntityCode).
Since the two methods do apparently the same (or did I miss some point?), why is there a need to make a distinction?
_Original issue: http://code.google.com/p/owasp-esapi-java/issues/detail?id=276_
| priority | difference between encodeforhtmlattribute and encodeforhtml from on july as i have seen in the source code the single difference between esapi encoder encodeforhtml and esapi encoder encodeforhtmlattribute is that is part of the whitelist in encodeforhtml and therefore not encoded esapi encoder encodeforhtmlattribute does indeed neither encode as it is part of the unencoded set line htmlentitycode since the two methods do apparently the same or did i miss some point why is there a need to make a distinction original issue | 1 |
57,829 | 3,084,029,445 | IssuesEvent | 2015-08-24 12:59:17 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | [Рефакторинг] Параметры "TO" и "FD" перенесены в 8 битное значение - | bug imported Priority-Medium | _From [Pavel.Pimenov@gmail.com](https://code.google.com/u/Pavel.Pimenov@gmail.com/) on July 29, 2013 12:07:51_
Это счетчик - раньше хранился в int
Посмотреть на что это влияет и узнать у Ежика зачем он сократил счетчик до 8 бит. может это ошибка?
Оригинал:
int connectionTimeouts = Util::toInt(ou.getIdentity().get("TO")) + 1;
ou.getIdentity().set("TO", Util::toString(connectionTimeouts));
int fileListDisconnects = Util::toInt(ou.getIdentity().get("FD")) + 1;
ou.getIdentity().set("FD", Util::toString(fileListDisconnects));
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1118_ | 1.0 | [Рефакторинг] Параметры "TO" и "FD" перенесены в 8 битное значение - - _From [Pavel.Pimenov@gmail.com](https://code.google.com/u/Pavel.Pimenov@gmail.com/) on July 29, 2013 12:07:51_
Это счетчик - раньше хранился в int
Посмотреть на что это влияет и узнать у Ежика зачем он сократил счетчик до 8 бит. может это ошибка?
Оригинал:
int connectionTimeouts = Util::toInt(ou.getIdentity().get("TO")) + 1;
ou.getIdentity().set("TO", Util::toString(connectionTimeouts));
int fileListDisconnects = Util::toInt(ou.getIdentity().get("FD")) + 1;
ou.getIdentity().set("FD", Util::toString(fileListDisconnects));
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1118_ | priority | параметры to и fd перенесены в битное значение from on july это счетчик раньше хранился в int посмотреть на что это влияет и узнать у ежика зачем он сократил счетчик до бит может это ошибка оригинал int connectiontimeouts util toint ou getidentity get to ou getidentity set to util tostring connectiontimeouts int filelistdisconnects util toint ou getidentity get fd ou getidentity set fd util tostring filelistdisconnects original issue | 1 |
509,202 | 14,724,053,888 | IssuesEvent | 2021-01-06 01:44:08 | googlefonts/noto-fonts | https://api.github.com/repos/googlefonts/noto-fonts | closed | Malay Jawi letter ڽ 'NYA' appeared as wrong glyph in Noto Sans Arabic (UI) | Fix-in-progress Kufi Priority-Medium Sans Script-Arabic hinted wrong glyph(s) | ## Defect Report
### Title
Malay Jawi letter ڽ 'NYA' appeared as wrong glyph in Noto Sans Arabic (UI)
### Font
1. NotoSansArabic-Regular.ttf
2. NotoSansArabicUI-Regular.ttf
### Where the font came from, and when
Site: https://noto-website-2.storage.googleapis.com/pkgs/NotoSansArabic-hinted.zip
Date: 2018-05-19 (GMT+8)
### Font Version
NotoSansArabic-Regular.ttf - `Version 2.000;GOOG;noto-source:20170915:90ef993387c0; ttfautohint (v1.7)`
NotoSansArabicUI-Regular.ttf - `Version 2.000;GOOG;noto-source:20170915:90ef993387c0; ttfautohint (v1.7)`
### OS name and version
Windows 8.1 Single Language
### Application name and version
No specific app, the issue is global.
### Issue
The [Jawi letter](https://en.wikipedia.org/wiki/Jawi_alphabet) NYA ([ڽ](https://en.wiktionary.org/wiki/%DA%BD)) which is used for the /ɲ/ sound in writing Jawi-scripted Malay language is appearing as wrong glyph in the initial form and medial form when I use both fonts.
In Noto Sans Arabic, the letter just uses the isolated form in place of initial and medial form.
In Noto Sans Arabic UI, the placement of the three dots are wrong, it should be below and not above.
For comparison, Noto Kufi Arabic and Noto Naskh Arabic UI displayed the letter correctly, and non-Noto font Times New Roman also displays correctly. (This is shown at Screenshot section below)
1. Steps to reproduce: just type as usual
2. Observed results: see screenshot, on first two rows
3. Expected results: see screenshot, on last three rows
4. Additional information: none
### Character data
The unicode codepoint for 'NYA' (ڽ) is 06BD.
### Screenshot

## Tools for reporting bugs
### Harfbuzz hb-view
Using the [windows executable](https://www.freedesktop.org/software/harfbuzz/release/harfbuzz-1.7.6-win32.zip) of Harfbuzz with the following input and command gives the corresponding output.
Input: in.txt
https://gist.github.com/MuhdNurHidayat/8630c720b064b6f3bcb9bad9fc768819

First command to test Noto Sans Arabic:
`hb-view.exe --font-file "C:\Users\User\Desktop\NotoSansArabic-Regular.ttf" --text-file "in.txt" --output-file "out1.png"`
Output:

Second command to test Noto Sans Arabic UI:
`hb-view.exe --font-file "C:\Users\User\Desktop\NotoSansArabicUI-Regular.ttf" --text-file "in.txt" --output-file "out2.png"`
Output:

### Fontview
The [tool](https://github.com/googlei18n/fontview/releases/tag/v0.3.1) is only released for OS X, I'm not a developer and I don't have the time to build the tool from source just for reporting wrong glyph.
### Fontdiff
The [tool](https://github.com/googlei18n/fontdiff/releases/tag/v0.2.3) is only released for OS X and Linux, I'm not a developer and I don't have the time to build the tool from source just for reporting wrong glyph. | 1.0 | Malay Jawi letter ڽ 'NYA' appeared as wrong glyph in Noto Sans Arabic (UI) - ## Defect Report
### Title
Malay Jawi letter ڽ 'NYA' appeared as wrong glyph in Noto Sans Arabic (UI)
### Font
1. NotoSansArabic-Regular.ttf
2. NotoSansArabicUI-Regular.ttf
### Where the font came from, and when
Site: https://noto-website-2.storage.googleapis.com/pkgs/NotoSansArabic-hinted.zip
Date: 2018-05-19 (GMT+8)
### Font Version
NotoSansArabic-Regular.ttf - `Version 2.000;GOOG;noto-source:20170915:90ef993387c0; ttfautohint (v1.7)`
NotoSansArabicUI-Regular.ttf - `Version 2.000;GOOG;noto-source:20170915:90ef993387c0; ttfautohint (v1.7)`
### OS name and version
Windows 8.1 Single Language
### Application name and version
No specific app, the issue is global.
### Issue
The [Jawi letter](https://en.wikipedia.org/wiki/Jawi_alphabet) NYA ([ڽ](https://en.wiktionary.org/wiki/%DA%BD)) which is used for the /ɲ/ sound in writing Jawi-scripted Malay language is appearing as wrong glyph in the initial form and medial form when I use both fonts.
In Noto Sans Arabic, the letter just uses the isolated form in place of initial and medial form.
In Noto Sans Arabic UI, the placement of the three dots are wrong, it should be below and not above.
For comparison, Noto Kufi Arabic and Noto Naskh Arabic UI displayed the letter correctly, and non-Noto font Times New Roman also displays correctly. (This is shown at Screenshot section below)
1. Steps to reproduce: just type as usual
2. Observed results: see screenshot, on first two rows
3. Expected results: see screenshot, on last three rows
4. Additional information: none
### Character data
The unicode codepoint for 'NYA' (ڽ) is 06BD.
### Screenshot

## Tools for reporting bugs
### Harfbuzz hb-view
Using the [windows executable](https://www.freedesktop.org/software/harfbuzz/release/harfbuzz-1.7.6-win32.zip) of Harfbuzz with the following input and command gives the corresponding output.
Input: in.txt
https://gist.github.com/MuhdNurHidayat/8630c720b064b6f3bcb9bad9fc768819

First command to test Noto Sans Arabic:
`hb-view.exe --font-file "C:\Users\User\Desktop\NotoSansArabic-Regular.ttf" --text-file "in.txt" --output-file "out1.png"`
Output:

Second command to test Noto Sans Arabic UI:
`hb-view.exe --font-file "C:\Users\User\Desktop\NotoSansArabicUI-Regular.ttf" --text-file "in.txt" --output-file "out2.png"`
Output:

### Fontview
The [tool](https://github.com/googlei18n/fontview/releases/tag/v0.3.1) is only released for OS X, I'm not a developer and I don't have the time to build the tool from source just for reporting wrong glyph.
### Fontdiff
The [tool](https://github.com/googlei18n/fontdiff/releases/tag/v0.2.3) is only released for OS X and Linux, I'm not a developer and I don't have the time to build the tool from source just for reporting wrong glyph. | priority | malay jawi letter ڽ nya appeared as wrong glyph in noto sans arabic ui defect report title malay jawi letter ڽ nya appeared as wrong glyph in noto sans arabic ui font notosansarabic regular ttf notosansarabicui regular ttf where the font came from and when site date gmt font version notosansarabic regular ttf version goog noto source ttfautohint notosansarabicui regular ttf version goog noto source ttfautohint os name and version windows single language application name and version no specific app the issue is global issue the nya which is used for the ɲ sound in writing jawi scripted malay language is appearing as wrong glyph in the initial form and medial form when i use both fonts in noto sans arabic the letter just uses the isolated form in place of initial and medial form in noto sans arabic ui the placement of the three dots are wrong it should be below and not above for comparison noto kufi arabic and noto naskh arabic ui displayed the letter correctly and non noto font times new roman also displays correctly this is shown at screenshot section below steps to reproduce just type as usual observed results see screenshot on first two rows expected results see screenshot on last three rows additional information none character data the unicode codepoint for nya ڽ is screenshot tools for reporting bugs harfbuzz hb view using the of harfbuzz with the following input and command gives the corresponding output input in txt first command to test noto sans arabic hb view exe font file c users user desktop notosansarabic regular ttf text file in txt output file png output second command to test noto sans arabic ui hb view exe font file c users user desktop notosansarabicui regular ttf text file in txt output file png output fontview the is only released for os x i m not a developer and i don t have the time to build the tool from source just for reporting wrong glyph fontdiff the is only released for os x and linux i m not a developer and i don t have the time to build the tool from source just for reporting wrong glyph | 1 |
319,958 | 9,762,632,703 | IssuesEvent | 2019-06-05 11:56:13 | EthVM/EthVM | https://api.github.com/repos/EthVM/EthVM | closed | Handle syncing correctly within the frontend | enhancement priority:medium project:api project:explorer | When syncing we should provide a visual indicator of the sync status and disconnect the live update subscription to prevent flooding of the for UI. When syncing has finished we then need to re-connect the subscritions of live update. | 1.0 | Handle syncing correctly within the frontend - When syncing we should provide a visual indicator of the sync status and disconnect the live update subscription to prevent flooding of the for UI. When syncing has finished we then need to re-connect the subscritions of live update. | priority | handle syncing correctly within the frontend when syncing we should provide a visual indicator of the sync status and disconnect the live update subscription to prevent flooding of the for ui when syncing has finished we then need to re connect the subscritions of live update | 1 |
77,108 | 3,506,261,670 | IssuesEvent | 2016-01-08 05:04:16 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Talent Res wen Player Reloged (BB #169) | migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 08.06.2010 09:06:49 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** invalid
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/169
<hr>
some Time All Talent res in palayer logOut | 1.0 | Talent Res wen Player Reloged (BB #169) - This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 08.06.2010 09:06:49 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** invalid
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/169
<hr>
some Time All Talent res in palayer logOut | priority | talent res wen player reloged bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state invalid direct link some time all talent res in palayer logout | 1 |
813,263 | 30,450,185,296 | IssuesEvent | 2023-07-16 07:29:05 | open-source-uc/planner | https://api.github.com/repos/open-source-uc/planner | closed | [PAN-196] [PAN-111] Acoplar errores similares (como desambiguación de cursos) | Medium priority |
<sub>From [SyncLinear.com](https://synclinear.com) | [PAN-111](https://linear.app/planner-uc/issue/PAN-111/acoplar-errores-similares-como-desambiguacion-de-cursos)</sub>
<sub>[PAN-196](https://linear.app/planner-uc/issue/PAN-196/[pan-111]-acoplar-errores-similares-como-desambiguacion-de-cursos)</sub> | 1.0 | [PAN-196] [PAN-111] Acoplar errores similares (como desambiguación de cursos) -
<sub>From [SyncLinear.com](https://synclinear.com) | [PAN-111](https://linear.app/planner-uc/issue/PAN-111/acoplar-errores-similares-como-desambiguacion-de-cursos)</sub>
<sub>[PAN-196](https://linear.app/planner-uc/issue/PAN-196/[pan-111]-acoplar-errores-similares-como-desambiguacion-de-cursos)</sub> | priority | acoplar errores similares como desambiguación de cursos from acoplar errores similares como desambiguacion de cursos | 1 |
274,680 | 8,564,168,794 | IssuesEvent | 2018-11-09 16:01:13 | CS2103-AY1819S1-W17-1/main | https://api.github.com/repos/CS2103-AY1819S1-W17-1/main | closed | Refactor ContextCommand to reduce coupling | priority.Medium type.Task | Goal: `ContextCommand` should not know about `AddEntryCommand`. It should only know about `Command`
Implementation idea:
1. Create a CommandSupplier interface
2. ContextCommand will accept a CommandSupplier via its constructor and will execute the Command supplied
3. ContextCommandParser will return a ContextCommand created using a CommandSupplier that depends on the exact input (i.e. the CommandSupplier may supply `AddEntryCommand` or any other command) | 1.0 | Refactor ContextCommand to reduce coupling - Goal: `ContextCommand` should not know about `AddEntryCommand`. It should only know about `Command`
Implementation idea:
1. Create a CommandSupplier interface
2. ContextCommand will accept a CommandSupplier via its constructor and will execute the Command supplied
3. ContextCommandParser will return a ContextCommand created using a CommandSupplier that depends on the exact input (i.e. the CommandSupplier may supply `AddEntryCommand` or any other command) | priority | refactor contextcommand to reduce coupling goal contextcommand should not know about addentrycommand it should only know about command implementation idea create a commandsupplier interface contextcommand will accept a commandsupplier via its constructor and will execute the command supplied contextcommandparser will return a contextcommand created using a commandsupplier that depends on the exact input i e the commandsupplier may supply addentrycommand or any other command | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.