Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
10,728
8,697,726,229
IssuesEvent
2018-12-04 21:07:59
conversationai/conversationai-models
https://api.github.com/repos/conversationai/conversationai-models
closed
Write subset of predictions to HTML file
infrastructure
Reading CSV files is tough, but it's often useful to look through the test data and predictions beyond just looking at the accuracy metrics. One solution is to write a sample of the predictions in a HTML format that we can add some basic styling to so it's easy to read. That way we can go from new model -> analyzing results really quickly.
1.0
Write subset of predictions to HTML file - Reading CSV files is tough, but it's often useful to look through the test data and predictions beyond just looking at the accuracy metrics. One solution is to write a sample of the predictions in a HTML format that we can add some basic styling to so it's easy to read. That way we can go from new model -> analyzing results really quickly.
non_main
write subset of predictions to html file reading csv files is tough but it s often useful to look through the test data and predictions beyond just looking at the accuracy metrics one solution is to write a sample of the predictions in a html format that we can add some basic styling to so it s easy to read that way we can go from new model analyzing results really quickly
0
5,056
25,892,470,280
IssuesEvent
2022-12-14 19:10:44
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
SortFieldNotFound error when linking to a newly-created table
type: bug work: backend status: ready restricted: maintainers
## Steps to reproduce 1. Begin with our Library Schema and the code in my PR #2080. 1. From the schema page, choose "New Table" > "From Scratch". 1. When the Table Page loads, rename the table to "Reviews". 1. Note that we get an error when loading the records. That error is reported in #2089. I'm not sure if it's related to the error I'm reporting now. 1. Create a new "Many to One" link to the Publications table. (Note that since recording the video below, the "Many to One" option has been relocated to the first option.) 1. Attempt to add a new record to "Reviews". 1. Observe the records POST request to fail with the following error <details> <summary>Traceback</summary> ``` Environment: Request Method: POST Request URL: http://localhost:8000/api/db/v0/tables/89/records/ Django Version: 3.1.14 Python Version: 3.9.16 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', 'mathesar.middleware.CursorClosedHandlerMiddleware', 'mathesar.middleware.PasswordChangeNeededMiddleware', 'django_userforeignkey.middleware.UserForeignKeyMiddleware', 'django_request_cache.middleware.RequestCacheMiddleware'] Traceback (most recent call last): File "/code/db/columns/utils.py", line 70, in find_column_by_name_in_relation return relation.columns[col_name_string] File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1158, in __getitem__ return self._index[key] During handling of the above exception (Column('439__323__col__323', INTEGER(), table=<anon_1>, primary_key=True, nullable=False)), another exception occurred: File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 826, in __getattr__ return getattr(self.comparator, key) The above exception ('Comparator' object has no attribute 'split') was the direct cause of the following exception: File "/code/db/columns/utils.py", line 54, in get_column_obj_from_relation column = find_column_by_name_in_relation(relation, column) File "/code/db/columns/utils.py", line 72, in find_column_by_name_in_relation col_name_split = col_name_string.split(sep='.', maxsplit=1) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 828, in __getattr__ util.raise_( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception During handling of the above exception (Neither 'Column' object nor 'Comparator' object has an attribute 'split'), another exception occurred: File "/code/db/records/operations/sort.py", line 57, in _get_sorted_column_obj_from_spec column = col_utils.get_column_obj_from_relation(relation, sort_spec.field) File "/code/db/columns/utils.py", line 56, in get_column_obj_from_relation column = relation.columns[column.name] File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1158, in __getitem__ return self._index[key] During handling of the above exception ('439__323__col__323'), another exception occurred: File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/code/mathesar/api/db/viewsets/records.py", line 155, in create records = paginator.paginate_queryset( File "/code/mathesar/api/pagination.py", line 89, in paginate_queryset records = query.get_records( File "/code/mathesar/models/query.py", line 172, in get_records return self.db_query.get_records(**kwargs) File "/code/db/queries/base.py", line 108, in get_records return records_select.get_records_with_default_order( File "/code/db/records/operations/select.py", line 28, in get_records_with_default_order return get_records( File "/code/db/records/operations/select.py", line 70, in get_records relation = apply_transformations_deprecated( File "/code/db/transforms/operations/apply.py", line 55, in apply_transformations_deprecated relation = apply_transformations(relation, transforms) File "/code/db/transforms/operations/apply.py", line 8, in apply_transformations relation = _apply_transform(relation, transform) File "/code/db/transforms/operations/apply.py", line 14, in _apply_transform relation = transform.apply_to_relation(relation) File "/code/db/transforms/base.py", line 124, in apply_to_relation executable = rec_sort.apply_relation_sorting(relation, order_by) File "/code/db/records/operations/sort.py", line 44, in apply_relation_sorting order_by_list = [ File "/code/db/records/operations/sort.py", line 45, in <listcomp> _get_sorted_column_obj_from_spec(relation, spec) for spec in sort_spec File "/code/db/records/operations/sort.py", line 59, in _get_sorted_column_obj_from_spec raise SortFieldNotFound(e) Exception Type: SortFieldNotFound at /api/db/v0/tables/89/records/ Exception Value: '439__323__col__323' ``` </details> https://user-images.githubusercontent.com/42411/207400140-bc4fdd90-6655-4213-95ce-a5cb111f6749.mp4
True
SortFieldNotFound error when linking to a newly-created table - ## Steps to reproduce 1. Begin with our Library Schema and the code in my PR #2080. 1. From the schema page, choose "New Table" > "From Scratch". 1. When the Table Page loads, rename the table to "Reviews". 1. Note that we get an error when loading the records. That error is reported in #2089. I'm not sure if it's related to the error I'm reporting now. 1. Create a new "Many to One" link to the Publications table. (Note that since recording the video below, the "Many to One" option has been relocated to the first option.) 1. Attempt to add a new record to "Reviews". 1. Observe the records POST request to fail with the following error <details> <summary>Traceback</summary> ``` Environment: Request Method: POST Request URL: http://localhost:8000/api/db/v0/tables/89/records/ Django Version: 3.1.14 Python Version: 3.9.16 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', 'mathesar.middleware.CursorClosedHandlerMiddleware', 'mathesar.middleware.PasswordChangeNeededMiddleware', 'django_userforeignkey.middleware.UserForeignKeyMiddleware', 'django_request_cache.middleware.RequestCacheMiddleware'] Traceback (most recent call last): File "/code/db/columns/utils.py", line 70, in find_column_by_name_in_relation return relation.columns[col_name_string] File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1158, in __getitem__ return self._index[key] During handling of the above exception (Column('439__323__col__323', INTEGER(), table=<anon_1>, primary_key=True, nullable=False)), another exception occurred: File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 826, in __getattr__ return getattr(self.comparator, key) The above exception ('Comparator' object has no attribute 'split') was the direct cause of the following exception: File "/code/db/columns/utils.py", line 54, in get_column_obj_from_relation column = find_column_by_name_in_relation(relation, column) File "/code/db/columns/utils.py", line 72, in find_column_by_name_in_relation col_name_split = col_name_string.split(sep='.', maxsplit=1) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 828, in __getattr__ util.raise_( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception During handling of the above exception (Neither 'Column' object nor 'Comparator' object has an attribute 'split'), another exception occurred: File "/code/db/records/operations/sort.py", line 57, in _get_sorted_column_obj_from_spec column = col_utils.get_column_obj_from_relation(relation, sort_spec.field) File "/code/db/columns/utils.py", line 56, in get_column_obj_from_relation column = relation.columns[column.name] File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1158, in __getitem__ return self._index[key] During handling of the above exception ('439__323__col__323'), another exception occurred: File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/code/mathesar/api/db/viewsets/records.py", line 155, in create records = paginator.paginate_queryset( File "/code/mathesar/api/pagination.py", line 89, in paginate_queryset records = query.get_records( File "/code/mathesar/models/query.py", line 172, in get_records return self.db_query.get_records(**kwargs) File "/code/db/queries/base.py", line 108, in get_records return records_select.get_records_with_default_order( File "/code/db/records/operations/select.py", line 28, in get_records_with_default_order return get_records( File "/code/db/records/operations/select.py", line 70, in get_records relation = apply_transformations_deprecated( File "/code/db/transforms/operations/apply.py", line 55, in apply_transformations_deprecated relation = apply_transformations(relation, transforms) File "/code/db/transforms/operations/apply.py", line 8, in apply_transformations relation = _apply_transform(relation, transform) File "/code/db/transforms/operations/apply.py", line 14, in _apply_transform relation = transform.apply_to_relation(relation) File "/code/db/transforms/base.py", line 124, in apply_to_relation executable = rec_sort.apply_relation_sorting(relation, order_by) File "/code/db/records/operations/sort.py", line 44, in apply_relation_sorting order_by_list = [ File "/code/db/records/operations/sort.py", line 45, in <listcomp> _get_sorted_column_obj_from_spec(relation, spec) for spec in sort_spec File "/code/db/records/operations/sort.py", line 59, in _get_sorted_column_obj_from_spec raise SortFieldNotFound(e) Exception Type: SortFieldNotFound at /api/db/v0/tables/89/records/ Exception Value: '439__323__col__323' ``` </details> https://user-images.githubusercontent.com/42411/207400140-bc4fdd90-6655-4213-95ce-a5cb111f6749.mp4
main
sortfieldnotfound error when linking to a newly created table steps to reproduce begin with our library schema and the code in my pr from the schema page choose new table from scratch when the table page loads rename the table to reviews note that we get an error when loading the records that error is reported in i m not sure if it s related to the error i m reporting now create a new many to one link to the publications table note that since recording the video below the many to one option has been relocated to the first option attempt to add a new record to reviews observe the records post request to fail with the following error traceback environment request method post request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware mathesar middleware cursorclosedhandlermiddleware mathesar middleware passwordchangeneededmiddleware django userforeignkey middleware userforeignkeymiddleware django request cache middleware requestcachemiddleware traceback most recent call last file code db columns utils py line in find column by name in relation return relation columns file usr local lib site packages sqlalchemy sql base py line in getitem return self index during handling of the above exception column col integer table primary key true nullable false another exception occurred file usr local lib site packages sqlalchemy sql elements py line in getattr return getattr self comparator key the above exception comparator object has no attribute split was the direct cause of the following exception file code db columns utils py line in get column obj from relation column find column by name in relation relation column file code db columns utils py line in find column by name in relation col name split col name string split sep maxsplit file usr local lib site packages sqlalchemy sql elements py line in getattr util raise file usr local lib site packages sqlalchemy util compat py line in raise raise exception during handling of the above exception neither column object nor comparator object has an attribute split another exception occurred file code db records operations sort py line in get sorted column obj from spec column col utils get column obj from relation relation sort spec field file code db columns utils py line in get column obj from relation column relation columns file usr local lib site packages sqlalchemy sql base py line in getitem return self index during handling of the above exception col another exception occurred file usr local lib site packages django core handlers exception py line in inner response get response request file usr local lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr local lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file usr local lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file usr local lib site packages rest framework views py line in dispatch response self handle exception exc file usr local lib site packages rest framework views py line in handle exception response exception handler exc context file code mathesar exception handlers py line in mathesar exception handler raise exc file usr local lib site packages rest framework views py line in dispatch response handler request args kwargs file code mathesar api db viewsets records py line in create records paginator paginate queryset file code mathesar api pagination py line in paginate queryset records query get records file code mathesar models query py line in get records return self db query get records kwargs file code db queries base py line in get records return records select get records with default order file code db records operations select py line in get records with default order return get records file code db records operations select py line in get records relation apply transformations deprecated file code db transforms operations apply py line in apply transformations deprecated relation apply transformations relation transforms file code db transforms operations apply py line in apply transformations relation apply transform relation transform file code db transforms operations apply py line in apply transform relation transform apply to relation relation file code db transforms base py line in apply to relation executable rec sort apply relation sorting relation order by file code db records operations sort py line in apply relation sorting order by list file code db records operations sort py line in get sorted column obj from spec relation spec for spec in sort spec file code db records operations sort py line in get sorted column obj from spec raise sortfieldnotfound e exception type sortfieldnotfound at api db tables records exception value col
1
82,034
23,649,117,146
IssuesEvent
2022-08-26 03:42:23
curl/curl
https://api.github.com/repos/curl/curl
closed
Cannot build curl statically and use static dependencies - Mac OS X
build not-a-bug
Hi there, I'm trying to build libcurl on Apple Mac OS X 11. I'm doing a minimal build with HTTPS and HTTP/2. I have built out openssl, zlib, and nghttp2 as static libraries -> libssl.a, libcrypto.a, libz.a. The files exist in that path as needed and the configure script picks them up correctly. However, despite passing --disable-shared, the configure script errors out with: `configure: error: one or more libs available at link-time are not available run-time. Libs used at link-time: -lnghttp2 -lssl -lcrypto -lz` And of course the .dylibs for those don't exist - they're static. I'm trying to do a fully 100% independent build of curl that has no reliance on any other dependencies - except for mac os x frameworks. I want my static openssl, my static nghttp2, and my static zlib. Here is my full configure command: `./configure --disable-shared --with-secure-transport --without-librtmp --without-libidn2 --with-zlib=/Users/jammerxd/Desktop/Git/cplusplusdependencies/zlib-1.2.11/MacOSX/x86_64/ --with-openssl=/Users/jammerxd/Desktop/Git/cplusplusdependencies/openssl-1.1.1m/MacOSX/x86_64 --disable-ldaps --disable-ntlm --disable-telnet --disable-proxy --disable-debug --enable-optimize --disable-curldebug --enable-symbol-hiding --disable-ftp --disable-rtsp --disable-dict --disable-tftp --disable-pop3 --disable-imap --disable-smb --disable-smtp --disable-gopher --disable-mqtt --disable-ntlm-wb --disable-tls-srp --disable-http-auth --disable-doh --disable-netrc --enable-progress-meter --disable-alt-svc --disable-ldap --disable-file --disable-dependency-tracking --with-nghttp2=/Users/jammerxd/Desktop/Git/cplusplusdependencies/nghttp2-1.49.0/MacOSX/x86_64/` If I try to prefix the above command with: `LDFLAGS="-static" PKG_CONFIG="pkg-config --static"` the configure tool errors out with: ``` checking whether the C compiler works... no configure: error: in `/Users/jammerxd/Desktop/Git/cplusplusdependencies/curl-7.84.0': configure: error: C compiler cannot create executables See `config.log' for more details ``` Looking on how I'm supposed to be able to build curl statically then have it link to the dependencies statically as well. Thanks, jammerxd
1.0
Cannot build curl statically and use static dependencies - Mac OS X - Hi there, I'm trying to build libcurl on Apple Mac OS X 11. I'm doing a minimal build with HTTPS and HTTP/2. I have built out openssl, zlib, and nghttp2 as static libraries -> libssl.a, libcrypto.a, libz.a. The files exist in that path as needed and the configure script picks them up correctly. However, despite passing --disable-shared, the configure script errors out with: `configure: error: one or more libs available at link-time are not available run-time. Libs used at link-time: -lnghttp2 -lssl -lcrypto -lz` And of course the .dylibs for those don't exist - they're static. I'm trying to do a fully 100% independent build of curl that has no reliance on any other dependencies - except for mac os x frameworks. I want my static openssl, my static nghttp2, and my static zlib. Here is my full configure command: `./configure --disable-shared --with-secure-transport --without-librtmp --without-libidn2 --with-zlib=/Users/jammerxd/Desktop/Git/cplusplusdependencies/zlib-1.2.11/MacOSX/x86_64/ --with-openssl=/Users/jammerxd/Desktop/Git/cplusplusdependencies/openssl-1.1.1m/MacOSX/x86_64 --disable-ldaps --disable-ntlm --disable-telnet --disable-proxy --disable-debug --enable-optimize --disable-curldebug --enable-symbol-hiding --disable-ftp --disable-rtsp --disable-dict --disable-tftp --disable-pop3 --disable-imap --disable-smb --disable-smtp --disable-gopher --disable-mqtt --disable-ntlm-wb --disable-tls-srp --disable-http-auth --disable-doh --disable-netrc --enable-progress-meter --disable-alt-svc --disable-ldap --disable-file --disable-dependency-tracking --with-nghttp2=/Users/jammerxd/Desktop/Git/cplusplusdependencies/nghttp2-1.49.0/MacOSX/x86_64/` If I try to prefix the above command with: `LDFLAGS="-static" PKG_CONFIG="pkg-config --static"` the configure tool errors out with: ``` checking whether the C compiler works... no configure: error: in `/Users/jammerxd/Desktop/Git/cplusplusdependencies/curl-7.84.0': configure: error: C compiler cannot create executables See `config.log' for more details ``` Looking on how I'm supposed to be able to build curl statically then have it link to the dependencies statically as well. Thanks, jammerxd
non_main
cannot build curl statically and use static dependencies mac os x hi there i m trying to build libcurl on apple mac os x i m doing a minimal build with https and http i have built out openssl zlib and as static libraries libssl a libcrypto a libz a the files exist in that path as needed and the configure script picks them up correctly however despite passing disable shared the configure script errors out with configure error one or more libs available at link time are not available run time libs used at link time lssl lcrypto lz and of course the dylibs for those don t exist they re static i m trying to do a fully independent build of curl that has no reliance on any other dependencies except for mac os x frameworks i want my static openssl my static and my static zlib here is my full configure command configure disable shared with secure transport without librtmp without with zlib users jammerxd desktop git cplusplusdependencies zlib macosx with openssl users jammerxd desktop git cplusplusdependencies openssl macosx disable ldaps disable ntlm disable telnet disable proxy disable debug enable optimize disable curldebug enable symbol hiding disable ftp disable rtsp disable dict disable tftp disable disable imap disable smb disable smtp disable gopher disable mqtt disable ntlm wb disable tls srp disable http auth disable doh disable netrc enable progress meter disable alt svc disable ldap disable file disable dependency tracking with users jammerxd desktop git cplusplusdependencies macosx if i try to prefix the above command with ldflags static pkg config pkg config static the configure tool errors out with checking whether the c compiler works no configure error in users jammerxd desktop git cplusplusdependencies curl configure error c compiler cannot create executables see config log for more details looking on how i m supposed to be able to build curl statically then have it link to the dependencies statically as well thanks jammerxd
0
74,221
14,220,967,715
IssuesEvent
2020-11-17 15:09:16
paulevsGitch/BetterNether
https://api.github.com/repos/paulevsGitch/BetterNether
closed
Player on seats snaps to south-facing position, and body lags behind head when turning
bug in dev code
Whilst testing the three chair offerings added by Better Nether, I noticed that the player model visible via F5 or another player's perspective has issues with the player's body not facing the same way the head is. The player's body seems to always want to face south, and, when the player turns their camera, the head follows perfectly, but the body lags behind more and more the further the camera goes from dead south. When facing dead north, there is a point where the body flips 180* instantly when continuing to turn around. This issue also means that chairs that don't face south have the player seated with their legs clipping through the back of the chair by default. I would hope that this issue can be fixed for a lovelier sitting-down experience.~ **Attachments** Facing dead south. ![2020-11-02_18 30 46](https://user-images.githubusercontent.com/69029346/97944889-c546eb00-1d3a-11eb-91ea-3f16fdd14a84.png) Turning away from south, the body lags behind, about halfway between where the head is facing and south. ![2020-11-02_18 30 52](https://user-images.githubusercontent.com/69029346/97944890-c710ae80-1d3a-11eb-8cdc-82238fc12306.png) This picture is right on the edge of where the body would flip, facing north. ![2020-11-02_18 31 03](https://user-images.githubusercontent.com/69029346/97944892-c7a94500-1d3a-11eb-8753-68be2b683bba.png)
1.0
Player on seats snaps to south-facing position, and body lags behind head when turning - Whilst testing the three chair offerings added by Better Nether, I noticed that the player model visible via F5 or another player's perspective has issues with the player's body not facing the same way the head is. The player's body seems to always want to face south, and, when the player turns their camera, the head follows perfectly, but the body lags behind more and more the further the camera goes from dead south. When facing dead north, there is a point where the body flips 180* instantly when continuing to turn around. This issue also means that chairs that don't face south have the player seated with their legs clipping through the back of the chair by default. I would hope that this issue can be fixed for a lovelier sitting-down experience.~ **Attachments** Facing dead south. ![2020-11-02_18 30 46](https://user-images.githubusercontent.com/69029346/97944889-c546eb00-1d3a-11eb-91ea-3f16fdd14a84.png) Turning away from south, the body lags behind, about halfway between where the head is facing and south. ![2020-11-02_18 30 52](https://user-images.githubusercontent.com/69029346/97944890-c710ae80-1d3a-11eb-8cdc-82238fc12306.png) This picture is right on the edge of where the body would flip, facing north. ![2020-11-02_18 31 03](https://user-images.githubusercontent.com/69029346/97944892-c7a94500-1d3a-11eb-8753-68be2b683bba.png)
non_main
player on seats snaps to south facing position and body lags behind head when turning whilst testing the three chair offerings added by better nether i noticed that the player model visible via or another player s perspective has issues with the player s body not facing the same way the head is the player s body seems to always want to face south and when the player turns their camera the head follows perfectly but the body lags behind more and more the further the camera goes from dead south when facing dead north there is a point where the body flips instantly when continuing to turn around this issue also means that chairs that don t face south have the player seated with their legs clipping through the back of the chair by default i would hope that this issue can be fixed for a lovelier sitting down experience attachments facing dead south turning away from south the body lags behind about halfway between where the head is facing and south this picture is right on the edge of where the body would flip facing north
0
4,383
22,304,098,274
IssuesEvent
2022-06-13 11:27:08
precice/precice
https://api.github.com/repos/precice/precice
opened
Clarify mesh API
enhancement usability maintainability breaking change
**Please describe the problem you are trying to solve.** The API for setting mesh primitives is confusing and tedious. * Triangles have to be set using edges * `setMeshTriangleWithEdges` sounds like it would take edges, but it actually takes vertices. #1057 * There are no bulk functions for setting edges and triangles. #465 * Adding `setMeshTetrahedron` requiring triangles would be a huge pain for users. #1314 * Exposing handles to connectivity (EdgeID) prevents us from optimizing meshes #1313 **Describe the solution you propose.** 1. Change the API to a vertex-only style. **:warning: breaking** 2. Remove `XWithEdges` 3. Add bulk functions Function | Inputs + MeshID | Outputs --- | --- | --- setMeshVertex | Coords | VertexID setMeshVertices | Count, Coords | VertexIDs setMeshEdge | VertexIDs | | setMeshEdges | Count, VertexIDs | | setMeshTriangle | VertexIDs | | setMeshTriangles | Count, VertexIDs | | setMeshQuad | VertexIDs | | setMeshQuads | Count, VertexIDs | | setMeshTetrahedron | VertexIDs | | setMeshTetrahedra | Count, VertexIDs | | **Describe alternatives you've considered** Leave it as it and end up with an increasingly confusing API.
True
Clarify mesh API - **Please describe the problem you are trying to solve.** The API for setting mesh primitives is confusing and tedious. * Triangles have to be set using edges * `setMeshTriangleWithEdges` sounds like it would take edges, but it actually takes vertices. #1057 * There are no bulk functions for setting edges and triangles. #465 * Adding `setMeshTetrahedron` requiring triangles would be a huge pain for users. #1314 * Exposing handles to connectivity (EdgeID) prevents us from optimizing meshes #1313 **Describe the solution you propose.** 1. Change the API to a vertex-only style. **:warning: breaking** 2. Remove `XWithEdges` 3. Add bulk functions Function | Inputs + MeshID | Outputs --- | --- | --- setMeshVertex | Coords | VertexID setMeshVertices | Count, Coords | VertexIDs setMeshEdge | VertexIDs | | setMeshEdges | Count, VertexIDs | | setMeshTriangle | VertexIDs | | setMeshTriangles | Count, VertexIDs | | setMeshQuad | VertexIDs | | setMeshQuads | Count, VertexIDs | | setMeshTetrahedron | VertexIDs | | setMeshTetrahedra | Count, VertexIDs | | **Describe alternatives you've considered** Leave it as it and end up with an increasingly confusing API.
main
clarify mesh api please describe the problem you are trying to solve the api for setting mesh primitives is confusing and tedious triangles have to be set using edges setmeshtrianglewithedges sounds like it would take edges but it actually takes vertices there are no bulk functions for setting edges and triangles adding setmeshtetrahedron requiring triangles would be a huge pain for users exposing handles to connectivity edgeid prevents us from optimizing meshes describe the solution you propose change the api to a vertex only style warning breaking remove xwithedges add bulk functions function inputs meshid outputs setmeshvertex coords vertexid setmeshvertices count coords vertexids setmeshedge vertexids setmeshedges count vertexids setmeshtriangle vertexids setmeshtriangles count vertexids setmeshquad vertexids setmeshquads count vertexids setmeshtetrahedron vertexids setmeshtetrahedra count vertexids describe alternatives you ve considered leave it as it and end up with an increasingly confusing api
1
294
3,072,891,779
IssuesEvent
2015-08-19 19:09:39
cattolyst/datafinisher
https://api.github.com/repos/cattolyst/datafinisher
opened
Test DF on i2b2 server
maintainability qc
Since the anticipated production environment would be the i2b2 server at our site (because that's where the JSON files live) it's kind of important that it be able to run there. However, ticket #22 causes an error and prevents us from finding out if anything else will break further on down. Once #22 is closed, need to test DF on i2b2.
True
Test DF on i2b2 server - Since the anticipated production environment would be the i2b2 server at our site (because that's where the JSON files live) it's kind of important that it be able to run there. However, ticket #22 causes an error and prevents us from finding out if anything else will break further on down. Once #22 is closed, need to test DF on i2b2.
main
test df on server since the anticipated production environment would be the server at our site because that s where the json files live it s kind of important that it be able to run there however ticket causes an error and prevents us from finding out if anything else will break further on down once is closed need to test df on
1
107,529
4,310,062,460
IssuesEvent
2016-07-21 17:59:41
DDMAL/rodan-client
https://api.github.com/repos/DDMAL/rodan-client
closed
Provide form for user to set info
Priority: SHOW STOPPER Type: FEATURE
- password reset (INCREDIBLY IMPORTANT) - name - email (VERY IMPORTANT) - get general Rodan notices (requires Rodan model update) - get project notices (requires Rodan model update)
1.0
Provide form for user to set info - - password reset (INCREDIBLY IMPORTANT) - name - email (VERY IMPORTANT) - get general Rodan notices (requires Rodan model update) - get project notices (requires Rodan model update)
non_main
provide form for user to set info password reset incredibly important name email very important get general rodan notices requires rodan model update get project notices requires rodan model update
0
216,550
24,281,564,666
IssuesEvent
2022-09-28 17:53:31
liorzilberg/swagger-parser
https://api.github.com/repos/liorzilberg/swagger-parser
closed
CVE-2020-35728 (High) detected in jackson-databind-2.9.5.jar - autoclosed
security vulnerability
## CVE-2020-35728 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/target/lib/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/liorzilberg/swagger-parser/commits/299682f5b4a2ec420c0c3f91a170670051db10d0">299682f5b4a2ec420c0c3f91a170670051db10d0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.oracle.wls.shaded.org.apache.xalan.lib.sql.JNDIConnectionPool (aka embedded Xalan in org.glassfish.web/javax.servlet.jsp.jstl). <p>Publish Date: 2020-12-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35728>CVE-2020-35728</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728</a></p> <p>Release Date: 2020-12-27</p> <p>Fix Resolution: 2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-35728 (High) detected in jackson-databind-2.9.5.jar - autoclosed - ## CVE-2020-35728 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/target/lib/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/liorzilberg/swagger-parser/commits/299682f5b4a2ec420c0c3f91a170670051db10d0">299682f5b4a2ec420c0c3f91a170670051db10d0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.oracle.wls.shaded.org.apache.xalan.lib.sql.JNDIConnectionPool (aka embedded Xalan in org.glassfish.web/javax.servlet.jsp.jstl). <p>Publish Date: 2020-12-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35728>CVE-2020-35728</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728</a></p> <p>Release Date: 2020-12-27</p> <p>Fix Resolution: 2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar target lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com oracle wls shaded org apache xalan lib sql jndiconnectionpool aka embedded xalan in org glassfish web javax servlet jsp jstl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
2,028
6,778,579,808
IssuesEvent
2017-10-28 12:49:56
chocolatey/chocolatey-package-requests
https://api.github.com/repos/chocolatey/chocolatey-package-requests
closed
RFP sdkman
Status: Available For Maintainer(s)
http://sdkman.io is a package for technology built on JVM(java virtual machine) such as grails, groovy, gradel etc
True
RFP sdkman - http://sdkman.io is a package for technology built on JVM(java virtual machine) such as grails, groovy, gradel etc
main
rfp sdkman is a package for technology built on jvm java virtual machine such as grails groovy gradel etc
1
4,857
24,996,288,577
IssuesEvent
2022-11-03 00:47:17
usefulmove/comp
https://api.github.com/repos/usefulmove/comp
closed
Command response function code reuse
maintainability
There is a significant amount of code duplication in the command response functions (`c_...`). It may be possible to simplify the code in meaningful ways ( poss. using function composition ).
True
Command response function code reuse - There is a significant amount of code duplication in the command response functions (`c_...`). It may be possible to simplify the code in meaningful ways ( poss. using function composition ).
main
command response function code reuse there is a significant amount of code duplication in the command response functions c it may be possible to simplify the code in meaningful ways poss using function composition
1
315,479
23,581,438,225
IssuesEvent
2022-08-23 08:01:24
hyperledger/iroha-2-docs
https://api.github.com/repos/hyperledger/iroha-2-docs
closed
Document FFI bindgen name mangling semantics
documentation enhancement help-wanted iroha2
``` Aleksandr Petrosyan, [25.07.22 19:41] That being said, we should document that convention too. rust module_name__method_name StructName__inherent_fn TraitName__assoc_fn_name … Marin, [25.07.22 19:56] [In reply to Aleksandr Petrosyan] 1. module__name::__method_name -> we do this for functions that have the same API but are implemented on different types. Namely this is for trait impls, i.e. poor man's generics like Clone 2. yes 3. no, it's actually StructName__trait_method_name at the moment, but, in my opinion, should be reduced to 1. after 2488 (https://github.com/hyperledger/iroha/issues/2488) Marin, [25.07.22 19:58] 1. for Clone in iroha_crypto gen_ffi_impl will create iroha_crypto::__clone in that library equivalent of Clone::clone Aleksandr Petrosyan, [25.07.22 19:59] This looks like we're mangling by hand. Marin, [25.07.22 19:59] which one? Aleksandr Petrosyan, [25.07.22 20:00] [In reply to Marin] This one Aleksandr Petrosyan, [25.07.22 20:00] Under the circumstances I doubt we can do much better Marin, [25.07.22 20:01] I edited the comment Aleksandr Petrosyan, [25.07.22 20:02] [In reply to Marin] Got it. ```
1.0
Document FFI bindgen name mangling semantics - ``` Aleksandr Petrosyan, [25.07.22 19:41] That being said, we should document that convention too. rust module_name__method_name StructName__inherent_fn TraitName__assoc_fn_name … Marin, [25.07.22 19:56] [In reply to Aleksandr Petrosyan] 1. module__name::__method_name -> we do this for functions that have the same API but are implemented on different types. Namely this is for trait impls, i.e. poor man's generics like Clone 2. yes 3. no, it's actually StructName__trait_method_name at the moment, but, in my opinion, should be reduced to 1. after 2488 (https://github.com/hyperledger/iroha/issues/2488) Marin, [25.07.22 19:58] 1. for Clone in iroha_crypto gen_ffi_impl will create iroha_crypto::__clone in that library equivalent of Clone::clone Aleksandr Petrosyan, [25.07.22 19:59] This looks like we're mangling by hand. Marin, [25.07.22 19:59] which one? Aleksandr Petrosyan, [25.07.22 20:00] [In reply to Marin] This one Aleksandr Petrosyan, [25.07.22 20:00] Under the circumstances I doubt we can do much better Marin, [25.07.22 20:01] I edited the comment Aleksandr Petrosyan, [25.07.22 20:02] [In reply to Marin] Got it. ```
non_main
document ffi bindgen name mangling semantics aleksandr petrosyan that being said we should document that convention too rust module name method name structname inherent fn traitname assoc fn name … marin module name method name we do this for functions that have the same api but are implemented on different types namely this is for trait impls i e poor man s generics like clone yes no it s actually structname trait method name at the moment but in my opinion should be reduced to after marin for clone in iroha crypto gen ffi impl will create iroha crypto clone in that library equivalent of clone clone aleksandr petrosyan this looks like we re mangling by hand marin which one aleksandr petrosyan this one aleksandr petrosyan under the circumstances i doubt we can do much better marin i edited the comment aleksandr petrosyan got it
0
4,932
25,350,832,636
IssuesEvent
2022-11-19 18:56:37
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Partial summarization does not consider partial grouping_expressions and aggregation_expressions
type: bug work: backend status: ready restricted: maintainers
## Description In the [partial summarizations PR](https://github.com/centerofci/mathesar/pull/1879), an important requirement that was discussed in earlier meetings have been possibly missed. It was decided that all `initial_columns` will be automatically identified as whether they are grouped/aggregated and included within either `grouping_expressions` or `aggregation_expressions`, and that the user will be able to partially specify both `grouping_expressions` and `aggregation_expressions`. Currently, the API only performs "guessing" when 'grouping_expressions' and 'aggregation_expressions' are both not present in the request. When they are present (partially or even empty), it does not do anything. cc: @dmos62 @mathemancer
True
Partial summarization does not consider partial grouping_expressions and aggregation_expressions - ## Description In the [partial summarizations PR](https://github.com/centerofci/mathesar/pull/1879), an important requirement that was discussed in earlier meetings have been possibly missed. It was decided that all `initial_columns` will be automatically identified as whether they are grouped/aggregated and included within either `grouping_expressions` or `aggregation_expressions`, and that the user will be able to partially specify both `grouping_expressions` and `aggregation_expressions`. Currently, the API only performs "guessing" when 'grouping_expressions' and 'aggregation_expressions' are both not present in the request. When they are present (partially or even empty), it does not do anything. cc: @dmos62 @mathemancer
main
partial summarization does not consider partial grouping expressions and aggregation expressions description in the an important requirement that was discussed in earlier meetings have been possibly missed it was decided that all initial columns will be automatically identified as whether they are grouped aggregated and included within either grouping expressions or aggregation expressions and that the user will be able to partially specify both grouping expressions and aggregation expressions currently the api only performs guessing when grouping expressions and aggregation expressions are both not present in the request when they are present partially or even empty it does not do anything cc mathemancer
1
210,050
16,332,333,168
IssuesEvent
2021-05-12 10:46:31
oss-review-toolkit/ort
https://api.github.com/repos/oss-review-toolkit/ort
closed
package config: Placeholders in referenced version?
documentation question
Hello, I'm just getting started with ORT and I'm wondering about one part of the docs. Referring to [this section of the package exclude documentation](https://github.com/oss-review-toolkit/ort/blob/master/docs/config-file-package-configuration-yml.md#defining-path-excludes-and-license-finding-curations): I want to exclude a dependency from the evaluation _regardless of its version_ because I own it and it is covered by another OSS analysis already. The documentation tells: > Each package configuration applies exactly to one package id and provenance which must be specified. The ability to omit the version or use placeholders would be very helpful. If I understand the documentation correctly, one needs to update the package configuration whenever one bumps the version of any dependency that is mentioned therein. Isn't that causing needless maintenance toil? Desired solution: * possibility to declare an upwards-compatible version – as offered by the caret notation of npm: e.g. `^1.2.10` or completely omit the version * If this is not currently offered due to implementation constraints, then the docs should say so. * documentation refined in regards to this use-case Pointer: https://github.com/oss-review-toolkit/ort/commit/0aaff6b04c6121cf511d9438596f2b177d2fba2d by @fviernau
1.0
package config: Placeholders in referenced version? - Hello, I'm just getting started with ORT and I'm wondering about one part of the docs. Referring to [this section of the package exclude documentation](https://github.com/oss-review-toolkit/ort/blob/master/docs/config-file-package-configuration-yml.md#defining-path-excludes-and-license-finding-curations): I want to exclude a dependency from the evaluation _regardless of its version_ because I own it and it is covered by another OSS analysis already. The documentation tells: > Each package configuration applies exactly to one package id and provenance which must be specified. The ability to omit the version or use placeholders would be very helpful. If I understand the documentation correctly, one needs to update the package configuration whenever one bumps the version of any dependency that is mentioned therein. Isn't that causing needless maintenance toil? Desired solution: * possibility to declare an upwards-compatible version – as offered by the caret notation of npm: e.g. `^1.2.10` or completely omit the version * If this is not currently offered due to implementation constraints, then the docs should say so. * documentation refined in regards to this use-case Pointer: https://github.com/oss-review-toolkit/ort/commit/0aaff6b04c6121cf511d9438596f2b177d2fba2d by @fviernau
non_main
package config placeholders in referenced version hello i m just getting started with ort and i m wondering about one part of the docs referring to i want to exclude a dependency from the evaluation regardless of its version because i own it and it is covered by another oss analysis already the documentation tells each package configuration applies exactly to one package id and provenance which must be specified the ability to omit the version or use placeholders would be very helpful if i understand the documentation correctly one needs to update the package configuration whenever one bumps the version of any dependency that is mentioned therein isn t that causing needless maintenance toil desired solution possibility to declare an upwards compatible version – as offered by the caret notation of npm e g or completely omit the version if this is not currently offered due to implementation constraints then the docs should say so documentation refined in regards to this use case pointer by fviernau
0
162,630
20,235,413,251
IssuesEvent
2022-02-14 01:06:39
johnnymythology/material-blog-jp
https://api.github.com/repos/johnnymythology/material-blog-jp
closed
CVE-2010-3299 (Medium) detected in rails-5.0.7.2.gem - autoclosed
security vulnerability
## CVE-2010-3299 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rails-5.0.7.2.gem</b></p></summary> <p>Ruby on Rails is a full-stack web framework optimized for programmer happiness and sustainable productivity. It encourages beautiful code by favoring convention over configuration.</p> <p>Library home page: <a href="https://rubygems.org/gems/rails-5.0.7.2.gem">https://rubygems.org/gems/rails-5.0.7.2.gem</a></p> <p>Path to dependency file: /material-blog-jp/Gemfile.lock</p> <p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/rails-5.0.7.2.gem</p> <p> Dependency Hierarchy: - rogue-0.1.1.gem (Root Library) - :x: **rails-5.0.7.2.gem** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The encrypt/decrypt functions in Ruby on Rails 2.3 are vulnerable to padding oracle attacks. <p>Publish Date: 2019-11-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2010-3299>CVE-2010-3299</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2010-3299">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2010-3299</a></p> <p>Release Date: 2019-11-12</p> <p>Fix Resolution: rails - 5.2.0.beta1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2010-3299 (Medium) detected in rails-5.0.7.2.gem - autoclosed - ## CVE-2010-3299 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rails-5.0.7.2.gem</b></p></summary> <p>Ruby on Rails is a full-stack web framework optimized for programmer happiness and sustainable productivity. It encourages beautiful code by favoring convention over configuration.</p> <p>Library home page: <a href="https://rubygems.org/gems/rails-5.0.7.2.gem">https://rubygems.org/gems/rails-5.0.7.2.gem</a></p> <p>Path to dependency file: /material-blog-jp/Gemfile.lock</p> <p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/rails-5.0.7.2.gem</p> <p> Dependency Hierarchy: - rogue-0.1.1.gem (Root Library) - :x: **rails-5.0.7.2.gem** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The encrypt/decrypt functions in Ruby on Rails 2.3 are vulnerable to padding oracle attacks. <p>Publish Date: 2019-11-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2010-3299>CVE-2010-3299</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2010-3299">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2010-3299</a></p> <p>Release Date: 2019-11-12</p> <p>Fix Resolution: rails - 5.2.0.beta1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in rails gem autoclosed cve medium severity vulnerability vulnerable library rails gem ruby on rails is a full stack web framework optimized for programmer happiness and sustainable productivity it encourages beautiful code by favoring convention over configuration library home page a href path to dependency file material blog jp gemfile lock path to vulnerable library var lib gems cache rails gem dependency hierarchy rogue gem root library x rails gem vulnerable library vulnerability details the encrypt decrypt functions in ruby on rails are vulnerable to padding oracle attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rails step up your open source security game with whitesource
0
2,002
6,717,616,077
IssuesEvent
2017-10-14 23:49:09
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Shuttle subsystem recovery needs a do over
Bug Maintainability/Hinders improvements
It's very broken and you can tell by looking at the code.
True
Shuttle subsystem recovery needs a do over - It's very broken and you can tell by looking at the code.
main
shuttle subsystem recovery needs a do over it s very broken and you can tell by looking at the code
1
143,718
13,081,329,393
IssuesEvent
2020-08-01 10:40:39
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Documentation issue: Build Tutorial - Android: Example does not build
team-Android type: documentation untriaged
Documentation URL: https://docs.bazel.build/versions/3.3.0/tutorial/android-app.html Installed bazel following the Ubuntu Linux using the apt repository on 18.04.4 LTS. Stacktrace: ``` SEVERE: Unexpected com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098 at com.google.common.base.Verify.verify(Verify.java:301) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668) at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108) at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258) at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231) at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81) at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233) at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177) Jul 10, 2020 12:52:58 PM com.google.devtools.build.android.ResourceProcessorBusyBox processRequest SEVERE: Error during processing com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098 at com.google.common.base.Verify.verify(Verify.java:301) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668) at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108) at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258) at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231) at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81) at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233) at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177) Exception in thread "main" com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098 at com.google.common.base.Verify.verify(Verify.java:301) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668) at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108) at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258) at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231) at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81) at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233) at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177) Target //src/main:app failed to build INFO: Elapsed time: 11.369s, Critical Path: 4.85s INFO: 18 processes: 11 linux-sandbox, 7 worker. FAILED: Build did NOT complete successfully ``` Patch with what I've done so far (on top of https://github.com/bazelbuild/examples/tree/e5d77821c2cd686f87310111499b11e70ca56f8e): ``` diff --git a/android/tutorial/WORKSPACE b/android/tutorial/WORKSPACE new file mode 100644 index 0000000..6f41f30 --- /dev/null +++ b/android/tutorial/WORKSPACE @@ -0,0 +1,6 @@ +android_sdk_repository( + name = "androidsdk", + path = "/home/jorge/Android/Sdk", + api_level = 25, + build_tools_version = "26.0.1" +) diff --git a/android/tutorial/src/main/BUILD b/android/tutorial/src/main/BUILD new file mode 100644 index 0000000..ce1490e --- /dev/null +++ b/android/tutorial/src/main/BUILD @@ -0,0 +1,5 @@ +android_binary( + name = "app", + manifest = "AndroidManifest.xml", + deps = ["//src/main/java/com/example/bazel:greeter_activity"], +) diff --git a/android/tutorial/src/main/java/com/example/bazel/BUILD b/android/tutorial/src/main/java/com/example/bazel/BUILD new file mode 100644 index 0000000..76f8ba7 --- /dev/null +++ b/android/tutorial/src/main/java/com/example/bazel/BUILD @@ -0,0 +1,13 @@ +package( + default_visibility = ["//src:__subpackages__"], +) + +android_library( + name = "greeter_activity", + srcs = [ + "Greeter.java", + "MainActivity.java", + ], + manifest = "AndroidManifest.xml", + resource_files = glob(["res/**"]), +) ```
1.0
Documentation issue: Build Tutorial - Android: Example does not build - Documentation URL: https://docs.bazel.build/versions/3.3.0/tutorial/android-app.html Installed bazel following the Ubuntu Linux using the apt repository on 18.04.4 LTS. Stacktrace: ``` SEVERE: Unexpected com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098 at com.google.common.base.Verify.verify(Verify.java:301) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668) at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108) at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258) at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231) at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81) at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233) at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177) Jul 10, 2020 12:52:58 PM com.google.devtools.build.android.ResourceProcessorBusyBox processRequest SEVERE: Error during processing com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098 at com.google.common.base.Verify.verify(Verify.java:301) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668) at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108) at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258) at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231) at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81) at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233) at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177) Exception in thread "main" com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098 at com.google.common.base.Verify.verify(Verify.java:301) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719) at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668) at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108) at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258) at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231) at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81) at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233) at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177) Target //src/main:app failed to build INFO: Elapsed time: 11.369s, Critical Path: 4.85s INFO: 18 processes: 11 linux-sandbox, 7 worker. FAILED: Build did NOT complete successfully ``` Patch with what I've done so far (on top of https://github.com/bazelbuild/examples/tree/e5d77821c2cd686f87310111499b11e70ca56f8e): ``` diff --git a/android/tutorial/WORKSPACE b/android/tutorial/WORKSPACE new file mode 100644 index 0000000..6f41f30 --- /dev/null +++ b/android/tutorial/WORKSPACE @@ -0,0 +1,6 @@ +android_sdk_repository( + name = "androidsdk", + path = "/home/jorge/Android/Sdk", + api_level = 25, + build_tools_version = "26.0.1" +) diff --git a/android/tutorial/src/main/BUILD b/android/tutorial/src/main/BUILD new file mode 100644 index 0000000..ce1490e --- /dev/null +++ b/android/tutorial/src/main/BUILD @@ -0,0 +1,5 @@ +android_binary( + name = "app", + manifest = "AndroidManifest.xml", + deps = ["//src/main/java/com/example/bazel:greeter_activity"], +) diff --git a/android/tutorial/src/main/java/com/example/bazel/BUILD b/android/tutorial/src/main/java/com/example/bazel/BUILD new file mode 100644 index 0000000..76f8ba7 --- /dev/null +++ b/android/tutorial/src/main/java/com/example/bazel/BUILD @@ -0,0 +1,13 @@ +package( + default_visibility = ["//src:__subpackages__"], +) + +android_library( + name = "greeter_activity", + srcs = [ + "Greeter.java", + "MainActivity.java", + ], + manifest = "AndroidManifest.xml", + resource_files = glob(["res/**"]), +) ```
non_main
documentation issue build tutorial android example does not build documentation url installed bazel following the ubuntu linux using the apt repository on lts stacktrace severe unexpected com google common base verifyexception unexpected magic number in generated src main java com example bazel res values colors arsc flat at com google common base verify verify verify java at com google devtools build android androidcompileddatadeserializer readresourcecontainer androidcompileddatadeserializer java at com google devtools build android androidcompileddatadeserializer read androidcompileddatadeserializer java at com google devtools build android serializedandroiddata deserialize serializedandroiddata java at com google devtools build android androidresourcemerger mergecompileddata androidresourcemerger java at com google devtools build android androidcompiledresourcemergingaction main androidcompiledresourcemergingaction java at com google devtools build android resourceprocessorbusybox tool call resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox processrequest resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox main resourceprocessorbusybox java jul pm com google devtools build android resourceprocessorbusybox processrequest severe error during processing com google common base verifyexception unexpected magic number in generated src main java com example bazel res values colors arsc flat at com google common base verify verify verify java at com google devtools build android androidcompileddatadeserializer readresourcecontainer androidcompileddatadeserializer java at com google devtools build android androidcompileddatadeserializer read androidcompileddatadeserializer java at com google devtools build android serializedandroiddata deserialize serializedandroiddata java at com google devtools build android androidresourcemerger mergecompileddata androidresourcemerger java at com google devtools build android androidcompiledresourcemergingaction main androidcompiledresourcemergingaction java at com google devtools build android resourceprocessorbusybox tool call resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox processrequest resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox main resourceprocessorbusybox java exception in thread main com google common base verifyexception unexpected magic number in generated src main java com example bazel res values colors arsc flat at com google common base verify verify verify java at com google devtools build android androidcompileddatadeserializer readresourcecontainer androidcompileddatadeserializer java at com google devtools build android androidcompileddatadeserializer read androidcompileddatadeserializer java at com google devtools build android serializedandroiddata deserialize serializedandroiddata java at com google devtools build android androidresourcemerger mergecompileddata androidresourcemerger java at com google devtools build android androidcompiledresourcemergingaction main androidcompiledresourcemergingaction java at com google devtools build android resourceprocessorbusybox tool call resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox processrequest resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox main resourceprocessorbusybox java target src main app failed to build info elapsed time critical path info processes linux sandbox worker failed build did not complete successfully patch with what i ve done so far on top of diff git a android tutorial workspace b android tutorial workspace new file mode index dev null b android tutorial workspace android sdk repository name androidsdk path home jorge android sdk api level build tools version diff git a android tutorial src main build b android tutorial src main build new file mode index dev null b android tutorial src main build android binary name app manifest androidmanifest xml deps diff git a android tutorial src main java com example bazel build b android tutorial src main java com example bazel build new file mode index dev null b android tutorial src main java com example bazel build package default visibility android library name greeter activity srcs greeter java mainactivity java manifest androidmanifest xml resource files glob
0
352,754
10,545,856,233
IssuesEvent
2019-10-02 20:05:59
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
closed
Edit collection page closes when the language is changed
component: collections priority: p4 state: stale type: bug
STR: 1. Load AMO dev and open one of your collections 2. Click Edit collection details, make some changes (edit the text) 3. Change the language 4. Observe the screen Actual results: Collection's page is displayed not the edit page and edited text is lost. Expected results: The Edit collection page is still displayed. Notes: This issue is reproducible on AMO dev and AMO stage with FF62(Win10). ![language3](https://user-images.githubusercontent.com/33448286/45879035-b6e02400-bdab-11e8-96b6-01850911677d.gif)
1.0
Edit collection page closes when the language is changed - STR: 1. Load AMO dev and open one of your collections 2. Click Edit collection details, make some changes (edit the text) 3. Change the language 4. Observe the screen Actual results: Collection's page is displayed not the edit page and edited text is lost. Expected results: The Edit collection page is still displayed. Notes: This issue is reproducible on AMO dev and AMO stage with FF62(Win10). ![language3](https://user-images.githubusercontent.com/33448286/45879035-b6e02400-bdab-11e8-96b6-01850911677d.gif)
non_main
edit collection page closes when the language is changed str load amo dev and open one of your collections click edit collection details make some changes edit the text change the language observe the screen actual results collection s page is displayed not the edit page and edited text is lost expected results the edit collection page is still displayed notes this issue is reproducible on amo dev and amo stage with
0
11,746
13,837,099,693
IssuesEvent
2020-10-14 02:50:45
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
reopened
Remove `NtIdent` hack for regressed crates
A-proc-macros C-future-compatibility T-compiler
## What is this issue? If you're a crate author who's been linked here, this issue tracks removing a backwards-compatibility hack in Rust. Rust has a longstanding issue https://github.com/rust-lang/rust/issues/43081, which causes procedural macros to lose location and hygiene information (known as a "Span") under certain circumstances. Recently, pull request https://github.com/rust-lang/rust/pull/73084 was merged, which makes progress towards resolving https://github.com/rust-lang/rust/issues/43081. Unfortunately, older versions of certain procedural macro crates (such as `proc-macro-hack v0.5.15` and `js-sys v0.3.39`) cannot handle the changes in input caused by the Rust bugfix. To allow these crates to continue to compile, a backward-compatibility hack was added to adjust the input passed to `proc-macro-hack` and `js-sys` specifically. Eventually, we would like to remove this backwards-compatibility hack, since the compiler should not have hard-coded exceptions for certain crates. However, removing this hack will break any crates that depend on affected versions of `proc-macro-hack` or `js-sys`. To ensure that your crate continues to work, you'll want to ensure that your `Cargo.lock` references `proc-macro-hack` v0.5.16 or above, and `js-sys` v0.3.40 or above. This can be done by running `cargo update -p proc-macro-hack` and `cargo update -p js-sys`. If you maintain a library crates (without a `Cargo.lock`, no action is needed on your part). ### Internal compiler details In https://github.com/rust-lang/rust/pull/73084#issuecomment-652613950, I added a hack to change the behavior of `NtIdent`s passed to certain proc-macros. This was done by special-casing certain identifiers, and should be eventually be removed in favor of a proper solution. If we decide to always wrap single identifiers in `None`-delimited groups, then we will need to wait until enough of the ecosystem has bumped `proc-macro-hack` and `wasm-bindgen`, to avoid breaking a large number of crates. Crater run: https://crater-reports.s3.amazonaws.com/pr-73084-1/index.html Triage: https://hackmd.io/O7icbSylRP6uVZyAQ9EDeA
True
Remove `NtIdent` hack for regressed crates - ## What is this issue? If you're a crate author who's been linked here, this issue tracks removing a backwards-compatibility hack in Rust. Rust has a longstanding issue https://github.com/rust-lang/rust/issues/43081, which causes procedural macros to lose location and hygiene information (known as a "Span") under certain circumstances. Recently, pull request https://github.com/rust-lang/rust/pull/73084 was merged, which makes progress towards resolving https://github.com/rust-lang/rust/issues/43081. Unfortunately, older versions of certain procedural macro crates (such as `proc-macro-hack v0.5.15` and `js-sys v0.3.39`) cannot handle the changes in input caused by the Rust bugfix. To allow these crates to continue to compile, a backward-compatibility hack was added to adjust the input passed to `proc-macro-hack` and `js-sys` specifically. Eventually, we would like to remove this backwards-compatibility hack, since the compiler should not have hard-coded exceptions for certain crates. However, removing this hack will break any crates that depend on affected versions of `proc-macro-hack` or `js-sys`. To ensure that your crate continues to work, you'll want to ensure that your `Cargo.lock` references `proc-macro-hack` v0.5.16 or above, and `js-sys` v0.3.40 or above. This can be done by running `cargo update -p proc-macro-hack` and `cargo update -p js-sys`. If you maintain a library crates (without a `Cargo.lock`, no action is needed on your part). ### Internal compiler details In https://github.com/rust-lang/rust/pull/73084#issuecomment-652613950, I added a hack to change the behavior of `NtIdent`s passed to certain proc-macros. This was done by special-casing certain identifiers, and should be eventually be removed in favor of a proper solution. If we decide to always wrap single identifiers in `None`-delimited groups, then we will need to wait until enough of the ecosystem has bumped `proc-macro-hack` and `wasm-bindgen`, to avoid breaking a large number of crates. Crater run: https://crater-reports.s3.amazonaws.com/pr-73084-1/index.html Triage: https://hackmd.io/O7icbSylRP6uVZyAQ9EDeA
non_main
remove ntident hack for regressed crates what is this issue if you re a crate author who s been linked here this issue tracks removing a backwards compatibility hack in rust rust has a longstanding issue which causes procedural macros to lose location and hygiene information known as a span under certain circumstances recently pull request was merged which makes progress towards resolving unfortunately older versions of certain procedural macro crates such as proc macro hack and js sys cannot handle the changes in input caused by the rust bugfix to allow these crates to continue to compile a backward compatibility hack was added to adjust the input passed to proc macro hack and js sys specifically eventually we would like to remove this backwards compatibility hack since the compiler should not have hard coded exceptions for certain crates however removing this hack will break any crates that depend on affected versions of proc macro hack or js sys to ensure that your crate continues to work you ll want to ensure that your cargo lock references proc macro hack or above and js sys or above this can be done by running cargo update p proc macro hack and cargo update p js sys if you maintain a library crates without a cargo lock no action is needed on your part internal compiler details in i added a hack to change the behavior of ntident s passed to certain proc macros this was done by special casing certain identifiers and should be eventually be removed in favor of a proper solution if we decide to always wrap single identifiers in none delimited groups then we will need to wait until enough of the ecosystem has bumped proc macro hack and wasm bindgen to avoid breaking a large number of crates crater run triage
0
5,719
30,232,146,500
IssuesEvent
2023-07-06 07:44:59
UBCSailbot/raye-local-pathfinding
https://api.github.com/repos/UBCSailbot/raye-local-pathfinding
closed
Remove references to min_voltage
bug maintain
### Purpose <!-- What is the purpose of this issue? What does this issue hope to achieve? --> The `min_voltage` ROS topic was removed in a recent PR to sailbot_msg. Local pathfinding still references `min_voltage`, which prevents the workspace from building. ### Changes <!-- What changes will be made to an existing feature? How will it be better? --> Remove/refactor all references to `min_voltage`.
True
Remove references to min_voltage - ### Purpose <!-- What is the purpose of this issue? What does this issue hope to achieve? --> The `min_voltage` ROS topic was removed in a recent PR to sailbot_msg. Local pathfinding still references `min_voltage`, which prevents the workspace from building. ### Changes <!-- What changes will be made to an existing feature? How will it be better? --> Remove/refactor all references to `min_voltage`.
main
remove references to min voltage purpose the min voltage ros topic was removed in a recent pr to sailbot msg local pathfinding still references min voltage which prevents the workspace from building changes remove refactor all references to min voltage
1
4,955
25,455,594,941
IssuesEvent
2022-11-24 13:57:17
pace/bricks
https://api.github.com/repos/pace/bricks
closed
Provide test helper for metrics
EST::Hours S::Ready T::Maintainance
### Problem It is not easy to test if the metrics are really collected. Currently, we rely on the availability `/metrics` and use that hardcoded in our tests. ### Solution Provide a small `test/testmetrics` package that supports a user to see if metrics are collected. The "/metrics" should be taken from the router instead of hard-coding it. _Originally posted by @Teelevision in https://github.com/pace/bricks/pull/92_
True
Provide test helper for metrics - ### Problem It is not easy to test if the metrics are really collected. Currently, we rely on the availability `/metrics` and use that hardcoded in our tests. ### Solution Provide a small `test/testmetrics` package that supports a user to see if metrics are collected. The "/metrics" should be taken from the router instead of hard-coding it. _Originally posted by @Teelevision in https://github.com/pace/bricks/pull/92_
main
provide test helper for metrics problem it is not easy to test if the metrics are really collected currently we rely on the availability metrics and use that hardcoded in our tests solution provide a small test testmetrics package that supports a user to see if metrics are collected the metrics should be taken from the router instead of hard coding it originally posted by teelevision in
1
2,837
10,209,489,549
IssuesEvent
2019-08-14 12:50:52
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
rax_facts needs to be renamed to _info and stop returning ansible_facts
affects_2.9 bug cloud module needs_maintainer needs_triage support:community
##### SUMMARY According to the [requirements](https://docs.ansible.com/ansible/latest/dev_guide/developing_modules_checklist.html#contributing-to-ansible-objective-requirements), `_facts` modules should return information that is specific to the host the module runs on as `ansible_facts`. (See also ansible/ansible#54280 for some discussion.) I think that this module do not satisfy this condition: it returns information that is not specific to the host machine, but bound to a cloud account. Please verify whether this is actually the case. To fix this, the core team thinks the best approach is to rename the `_facts` modules to `_info` modules, and change the return value so that if the module is called with the `_facts` name (should be deprecated, and will then be removed in Ansible 2.13), it should return `ansible_facts` as now (and print a deprecation message), and if it is called with the `_info` name, it no longer returns `ansible_facts`, so that the result needs to be registered. This informs users that they have to change something, won't break their playbooks, and allows them (by switching to the `_info` names) to verify that their roles and playbooks no longer use the facts. If that approach doesn't work, or you want to change the module's interface (like rename the return value(s)), it is better to deprecate the `_facts` module and create a new `_info` module next to it. To see both approaches implemented, you can take a look at https://github.com/ansible/ansible/commit/2d98734ad51c3de557c3a7d1f7166bdd114ceb0f (ansible/ansible#60178). If you have any input — preferred method to solve this, doubts that all modules should be renamed, ... — please write it here! ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME rax_facts ##### ANSIBLE VERSION ```paste below 2.9.0 ```
True
rax_facts needs to be renamed to _info and stop returning ansible_facts - ##### SUMMARY According to the [requirements](https://docs.ansible.com/ansible/latest/dev_guide/developing_modules_checklist.html#contributing-to-ansible-objective-requirements), `_facts` modules should return information that is specific to the host the module runs on as `ansible_facts`. (See also ansible/ansible#54280 for some discussion.) I think that this module do not satisfy this condition: it returns information that is not specific to the host machine, but bound to a cloud account. Please verify whether this is actually the case. To fix this, the core team thinks the best approach is to rename the `_facts` modules to `_info` modules, and change the return value so that if the module is called with the `_facts` name (should be deprecated, and will then be removed in Ansible 2.13), it should return `ansible_facts` as now (and print a deprecation message), and if it is called with the `_info` name, it no longer returns `ansible_facts`, so that the result needs to be registered. This informs users that they have to change something, won't break their playbooks, and allows them (by switching to the `_info` names) to verify that their roles and playbooks no longer use the facts. If that approach doesn't work, or you want to change the module's interface (like rename the return value(s)), it is better to deprecate the `_facts` module and create a new `_info` module next to it. To see both approaches implemented, you can take a look at https://github.com/ansible/ansible/commit/2d98734ad51c3de557c3a7d1f7166bdd114ceb0f (ansible/ansible#60178). If you have any input — preferred method to solve this, doubts that all modules should be renamed, ... — please write it here! ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME rax_facts ##### ANSIBLE VERSION ```paste below 2.9.0 ```
main
rax facts needs to be renamed to info and stop returning ansible facts summary according to the facts modules should return information that is specific to the host the module runs on as ansible facts see also ansible ansible for some discussion i think that this module do not satisfy this condition it returns information that is not specific to the host machine but bound to a cloud account please verify whether this is actually the case to fix this the core team thinks the best approach is to rename the facts modules to info modules and change the return value so that if the module is called with the facts name should be deprecated and will then be removed in ansible it should return ansible facts as now and print a deprecation message and if it is called with the info name it no longer returns ansible facts so that the result needs to be registered this informs users that they have to change something won t break their playbooks and allows them by switching to the info names to verify that their roles and playbooks no longer use the facts if that approach doesn t work or you want to change the module s interface like rename the return value s it is better to deprecate the facts module and create a new info module next to it to see both approaches implemented you can take a look at ansible ansible if you have any input — preferred method to solve this doubts that all modules should be renamed — please write it here issue type bug report component name rax facts ansible version paste below
1
965
4,707,894,644
IssuesEvent
2016-10-13 21:31:10
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Mutiple inputs to vsphere_guest module are silently ignored when launching from a template
affects_1.9 bug_report cloud vmware waiting_on_maintainer
[![Bountysource](https://www.bountysource.com/badge/issue?issue_id=16226901)](https://www.bountysource.com/issues/16226901-mutiple-inputs-to-vsphere_guest-module-are-silently-ignored-when-launching-from-a-template?utm_source=16226901&utm_medium=shield&utm_campaign=ISSUE_BADGE) #### Issue Type: Bug Report #### Ansible Version: 1.9.1 #### Ansible Configuration: n/a #### Environment: OS X 10.10.3, vCenter 6 #### Summary: Mutiple inputs to vsphere_guest module are ignored when launching from a template. Does not throw errors for invalid data either. vm_nic, vm_disk & vm_hardware parameters are all ignored. #### Steps To Reproduce: This command creates VM with but ignores valid options. Invalid options are ignored as well. ``` - name: Create other-app VM for testing connection: local vsphere_guest: vcenter_hostname: 110.120.113.333 username: foo password: bar guest: other-app-{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}{{ 9 |random}} vm_disk: disk1: size_gb: 32 type: thin datastore: foo_VMFS_1 vm_hardware: memory_mb: 8192 num_cpus: 2 osid: ubuntu64Guest scsi: lsi vm_nic: nic1: type: vmxnet3 network: slartibartfast network_type: dvs vm_extra_config: notes: this is a test VM cluster: Ontrack-Cluster1 resource_pool: /Resources/foo-default from_template: yes template_src: ubuntu-14.04-template esxi: datacenter: foo-DC ``` #### Expected Results: VM launched with specific NIC, disk and hardware settings. Invalid settings (such as mis-spelt NIC networks should produce an error). #### Actual Results: VM is launched but network, CPU and RAM that was originally specified in the template is retained. ``` <localhost> REMOTE_MODULE vsphere_guest template_src=ubuntu-14.04-template vcenter_hostname=110.120.113.333 cluster=Ontrack-Cluster1 guest=other-app-03770 password=VALUE_HIDDEN resource_pool=/Resources/foo-default username=foo <localhost> EXEC ['/bin/sh', '-c', 'mkdir -p $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913 && chmod a+rx $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913 && echo $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913'] <localhost> PUT /var/folders/s0/jy7xj1915_bg41nmvc3cjcdw0000gn/T/tmpDWPVuq TO /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/vsphere_guest <localhost> EXEC ['/bin/sh', '-c', u'LANG=en_US.UTF-8 LC_CTYPE=en_US.UTF-8 /usr/bin/python /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/vsphere_guest; rm -rf /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/ >/dev/null 2>&1'] changed: [localhost] => {"changed": true} ```
True
Mutiple inputs to vsphere_guest module are silently ignored when launching from a template - [![Bountysource](https://www.bountysource.com/badge/issue?issue_id=16226901)](https://www.bountysource.com/issues/16226901-mutiple-inputs-to-vsphere_guest-module-are-silently-ignored-when-launching-from-a-template?utm_source=16226901&utm_medium=shield&utm_campaign=ISSUE_BADGE) #### Issue Type: Bug Report #### Ansible Version: 1.9.1 #### Ansible Configuration: n/a #### Environment: OS X 10.10.3, vCenter 6 #### Summary: Mutiple inputs to vsphere_guest module are ignored when launching from a template. Does not throw errors for invalid data either. vm_nic, vm_disk & vm_hardware parameters are all ignored. #### Steps To Reproduce: This command creates VM with but ignores valid options. Invalid options are ignored as well. ``` - name: Create other-app VM for testing connection: local vsphere_guest: vcenter_hostname: 110.120.113.333 username: foo password: bar guest: other-app-{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}{{ 9 |random}} vm_disk: disk1: size_gb: 32 type: thin datastore: foo_VMFS_1 vm_hardware: memory_mb: 8192 num_cpus: 2 osid: ubuntu64Guest scsi: lsi vm_nic: nic1: type: vmxnet3 network: slartibartfast network_type: dvs vm_extra_config: notes: this is a test VM cluster: Ontrack-Cluster1 resource_pool: /Resources/foo-default from_template: yes template_src: ubuntu-14.04-template esxi: datacenter: foo-DC ``` #### Expected Results: VM launched with specific NIC, disk and hardware settings. Invalid settings (such as mis-spelt NIC networks should produce an error). #### Actual Results: VM is launched but network, CPU and RAM that was originally specified in the template is retained. ``` <localhost> REMOTE_MODULE vsphere_guest template_src=ubuntu-14.04-template vcenter_hostname=110.120.113.333 cluster=Ontrack-Cluster1 guest=other-app-03770 password=VALUE_HIDDEN resource_pool=/Resources/foo-default username=foo <localhost> EXEC ['/bin/sh', '-c', 'mkdir -p $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913 && chmod a+rx $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913 && echo $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913'] <localhost> PUT /var/folders/s0/jy7xj1915_bg41nmvc3cjcdw0000gn/T/tmpDWPVuq TO /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/vsphere_guest <localhost> EXEC ['/bin/sh', '-c', u'LANG=en_US.UTF-8 LC_CTYPE=en_US.UTF-8 /usr/bin/python /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/vsphere_guest; rm -rf /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/ >/dev/null 2>&1'] changed: [localhost] => {"changed": true} ```
main
mutiple inputs to vsphere guest module are silently ignored when launching from a template issue type bug report ansible version ansible configuration n a environment os x vcenter summary mutiple inputs to vsphere guest module are ignored when launching from a template does not throw errors for invalid data either vm nic vm disk vm hardware parameters are all ignored steps to reproduce this command creates vm with but ignores valid options invalid options are ignored as well name create other app vm for testing connection local vsphere guest vcenter hostname username foo password bar guest other app random random random random random vm disk size gb type thin datastore foo vmfs vm hardware memory mb num cpus osid scsi lsi vm nic type network slartibartfast network type dvs vm extra config notes this is a test vm cluster ontrack resource pool resources foo default from template yes template src ubuntu template esxi datacenter foo dc expected results vm launched with specific nic disk and hardware settings invalid settings such as mis spelt nic networks should produce an error actual results vm is launched but network cpu and ram that was originally specified in the template is retained remote module vsphere guest template src ubuntu template vcenter hostname cluster ontrack guest other app password value hidden resource pool resources foo default username foo exec put var folders t tmpdwpvuq to users tpai ansible tmp ansible tmp vsphere guest exec changed changed true
1
11,501
14,380,343,568
IssuesEvent
2020-12-02 02:34:05
KevCor99/4a
https://api.github.com/repos/KevCor99/4a
closed
file_size_estimating_template
process-dashboard
-llenado de template de estimacion de lineas de coodigo en process dashboard - correr el PROBE wizard
1.0
file_size_estimating_template - -llenado de template de estimacion de lineas de coodigo en process dashboard - correr el PROBE wizard
non_main
file size estimating template llenado de template de estimacion de lineas de coodigo en process dashboard correr el probe wizard
0
250,733
21,332,376,563
IssuesEvent
2022-04-18 10:08:36
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
[CI] CCRIT.testAutoFollowing
>test-failure :Distributed/CCR
### CI Link https://gradle-enterprise.elastic.co/s/xqkefsonks3ni/tests/:client:rest-high-level:asyncIntegTest/org.elasticsearch.client.CCRIT/testAutoFollowing?top-execution=1 ### Repro line ./gradlew ':client:rest-high-level:asyncIntegTest' --tests "org.elasticsearch.client.CCRIT.testAutoFollowing" -Dtests.seed=BD8CF40508D1FA6A -Dtests.locale=en-US -Dtests.timezone=Pacific/Saipan -Druntime.java=17 ### Does it reproduce? Yes ### Applicable branches master ### Failure history _No response_ ### Failure excerpt ``` org.elasticsearch.client.CCRIT > testAutoFollowing FAILED java.lang.AssertionError: Expected: not null but: was null at __randomizedtesting.SeedInfo.seed([BD8CF40508D1FA6A:1262D2050C785147]:0) at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) at org.junit.Assert.assertThat(Assert.java:956) at org.junit.Assert.assertThat(Assert.java:923) at org.elasticsearch.client.CCRIT.lambda$testAutoFollowing$3(CCRIT.java:293) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1069) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1042) at org.elasticsearch.client.CCRIT.testAutoFollowing(CCRIT.java:289) ``` The test fails locally too but with a different error: ``` org.elasticsearch.client.CCRIT > testAutoFollowing FAILED org.elasticsearch.ElasticsearchStatusException: Elasticsearch exception [type=status_exception, reason=can not fetch remote cluster state as the license state of the remote cluster [local_cluster] could not be determined] at __randomizedtesting.SeedInfo.seed([BD8CF40508D1FA6A:1262D2050C785147]:0) at app//org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:178) at app//org.elasticsearch.client.RestHighLevelClient.parseEntity(RestHighLevelClient.java:2484) at app//org.elasticsearch.client.RestHighLevelClient.parseResponseException(RestHighLevelClient.java:2461) at app//org.elasticsearch.client.RestHighLevelClient$1.onFailure(RestHighLevelClient.java:2373) at app//org.elasticsearch.client.RestClient$FailureTrackingResponseListener.onDefinitiveFailure(RestClient.java:672) at app//org.elasticsearch.client.RestClient$1.completed(RestClient.java:408) at app//org.elasticsearch.client.RestClient$1.completed(RestClient.java:392) at app//org.apache.http.concurrent.BasicFuture.completed(BasicFuture.java:122) at app//org.apache.http.impl.nio.client.DefaultClientExchangeHandlerImpl.responseCompleted(DefaultClientExchangeHandlerImpl.java:181) at app//org.apache.http.nio.protocol.HttpAsyncRequestExecutor.processResponse(HttpAsyncRequestExecutor.java:448) at app//org.apache.http.nio.protocol.HttpAsyncRequestExecutor.inputReady(HttpAsyncRequestExecutor.java:338) at app//org.apache.http.impl.nio.DefaultNHttpClientConnection.consumeInput(DefaultNHttpClientConnection.java:265) at app//org.apache.http.impl.nio.client.InternalIODispatch.onInputReady(InternalIODispatch.java:81) at app//org.apache.http.impl.nio.client.InternalIODispatch.onInputReady(InternalIODispatch.java:39) at app//org.apache.http.impl.nio.reactor.AbstractIODispatch.inputReady(AbstractIODispatch.java:114) at app//org.apache.http.impl.nio.reactor.BaseIOReactor.readable(BaseIOReactor.java:162) at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.processEvent(AbstractIOReactor.java:337) at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.processEvents(AbstractIOReactor.java:315) at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.execute(AbstractIOReactor.java:276) at app//org.apache.http.impl.nio.reactor.BaseIOReactor.execute(BaseIOReactor.java:104) at app//org.apache.http.impl.nio.reactor.AbstractMultiworkerIOReactor$Worker.run(AbstractMultiworkerIOReactor.java:591) at java.base@18/java.lang.Thread.run(Thread.java:833) Caused by: org.elasticsearch.ElasticsearchException: Elasticsearch exception [type=resource_not_found_exception, reason=license info is missing for cluster [local_cluster]] at app//org.elasticsearch.ElasticsearchException.innerFromXContent(ElasticsearchException.java:490) at app//org.elasticsearch.ElasticsearchException.fromXContent(ElasticsearchException.java:401) at app//org.elasticsearch.ElasticsearchException.innerFromXContent(ElasticsearchException.java:431) at app//org.elasticsearch.ElasticsearchException.failureFromXContent(ElasticsearchException.java:597) at app//org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:170) ... 21 more ```
1.0
[CI] CCRIT.testAutoFollowing - ### CI Link https://gradle-enterprise.elastic.co/s/xqkefsonks3ni/tests/:client:rest-high-level:asyncIntegTest/org.elasticsearch.client.CCRIT/testAutoFollowing?top-execution=1 ### Repro line ./gradlew ':client:rest-high-level:asyncIntegTest' --tests "org.elasticsearch.client.CCRIT.testAutoFollowing" -Dtests.seed=BD8CF40508D1FA6A -Dtests.locale=en-US -Dtests.timezone=Pacific/Saipan -Druntime.java=17 ### Does it reproduce? Yes ### Applicable branches master ### Failure history _No response_ ### Failure excerpt ``` org.elasticsearch.client.CCRIT > testAutoFollowing FAILED java.lang.AssertionError: Expected: not null but: was null at __randomizedtesting.SeedInfo.seed([BD8CF40508D1FA6A:1262D2050C785147]:0) at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) at org.junit.Assert.assertThat(Assert.java:956) at org.junit.Assert.assertThat(Assert.java:923) at org.elasticsearch.client.CCRIT.lambda$testAutoFollowing$3(CCRIT.java:293) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1069) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1042) at org.elasticsearch.client.CCRIT.testAutoFollowing(CCRIT.java:289) ``` The test fails locally too but with a different error: ``` org.elasticsearch.client.CCRIT > testAutoFollowing FAILED org.elasticsearch.ElasticsearchStatusException: Elasticsearch exception [type=status_exception, reason=can not fetch remote cluster state as the license state of the remote cluster [local_cluster] could not be determined] at __randomizedtesting.SeedInfo.seed([BD8CF40508D1FA6A:1262D2050C785147]:0) at app//org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:178) at app//org.elasticsearch.client.RestHighLevelClient.parseEntity(RestHighLevelClient.java:2484) at app//org.elasticsearch.client.RestHighLevelClient.parseResponseException(RestHighLevelClient.java:2461) at app//org.elasticsearch.client.RestHighLevelClient$1.onFailure(RestHighLevelClient.java:2373) at app//org.elasticsearch.client.RestClient$FailureTrackingResponseListener.onDefinitiveFailure(RestClient.java:672) at app//org.elasticsearch.client.RestClient$1.completed(RestClient.java:408) at app//org.elasticsearch.client.RestClient$1.completed(RestClient.java:392) at app//org.apache.http.concurrent.BasicFuture.completed(BasicFuture.java:122) at app//org.apache.http.impl.nio.client.DefaultClientExchangeHandlerImpl.responseCompleted(DefaultClientExchangeHandlerImpl.java:181) at app//org.apache.http.nio.protocol.HttpAsyncRequestExecutor.processResponse(HttpAsyncRequestExecutor.java:448) at app//org.apache.http.nio.protocol.HttpAsyncRequestExecutor.inputReady(HttpAsyncRequestExecutor.java:338) at app//org.apache.http.impl.nio.DefaultNHttpClientConnection.consumeInput(DefaultNHttpClientConnection.java:265) at app//org.apache.http.impl.nio.client.InternalIODispatch.onInputReady(InternalIODispatch.java:81) at app//org.apache.http.impl.nio.client.InternalIODispatch.onInputReady(InternalIODispatch.java:39) at app//org.apache.http.impl.nio.reactor.AbstractIODispatch.inputReady(AbstractIODispatch.java:114) at app//org.apache.http.impl.nio.reactor.BaseIOReactor.readable(BaseIOReactor.java:162) at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.processEvent(AbstractIOReactor.java:337) at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.processEvents(AbstractIOReactor.java:315) at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.execute(AbstractIOReactor.java:276) at app//org.apache.http.impl.nio.reactor.BaseIOReactor.execute(BaseIOReactor.java:104) at app//org.apache.http.impl.nio.reactor.AbstractMultiworkerIOReactor$Worker.run(AbstractMultiworkerIOReactor.java:591) at java.base@18/java.lang.Thread.run(Thread.java:833) Caused by: org.elasticsearch.ElasticsearchException: Elasticsearch exception [type=resource_not_found_exception, reason=license info is missing for cluster [local_cluster]] at app//org.elasticsearch.ElasticsearchException.innerFromXContent(ElasticsearchException.java:490) at app//org.elasticsearch.ElasticsearchException.fromXContent(ElasticsearchException.java:401) at app//org.elasticsearch.ElasticsearchException.innerFromXContent(ElasticsearchException.java:431) at app//org.elasticsearch.ElasticsearchException.failureFromXContent(ElasticsearchException.java:597) at app//org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:170) ... 21 more ```
non_main
ccrit testautofollowing ci link repro line gradlew client rest high level asyncintegtest tests org elasticsearch client ccrit testautofollowing dtests seed dtests locale en us dtests timezone pacific saipan druntime java does it reproduce yes applicable branches master failure history no response failure excerpt org elasticsearch client ccrit testautofollowing failed java lang assertionerror expected not null but was null at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch client ccrit lambda testautofollowing ccrit java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch client ccrit testautofollowing ccrit java the test fails locally too but with a different error org elasticsearch client ccrit testautofollowing failed org elasticsearch elasticsearchstatusexception elasticsearch exception could not be determined at randomizedtesting seedinfo seed at app org elasticsearch rest bytesrestresponse errorfromxcontent bytesrestresponse java at app org elasticsearch client resthighlevelclient parseentity resthighlevelclient java at app org elasticsearch client resthighlevelclient parseresponseexception resthighlevelclient java at app org elasticsearch client resthighlevelclient onfailure resthighlevelclient java at app org elasticsearch client restclient failuretrackingresponselistener ondefinitivefailure restclient java at app org elasticsearch client restclient completed restclient java at app org elasticsearch client restclient completed restclient java at app org apache http concurrent basicfuture completed basicfuture java at app org apache http impl nio client defaultclientexchangehandlerimpl responsecompleted defaultclientexchangehandlerimpl java at app org apache http nio protocol httpasyncrequestexecutor processresponse httpasyncrequestexecutor java at app org apache http nio protocol httpasyncrequestexecutor inputready httpasyncrequestexecutor java at app org apache http impl nio defaultnhttpclientconnection consumeinput defaultnhttpclientconnection java at app org apache http impl nio client internaliodispatch oninputready internaliodispatch java at app org apache http impl nio client internaliodispatch oninputready internaliodispatch java at app org apache http impl nio reactor abstractiodispatch inputready abstractiodispatch java at app org apache http impl nio reactor baseioreactor readable baseioreactor java at app org apache http impl nio reactor abstractioreactor processevent abstractioreactor java at app org apache http impl nio reactor abstractioreactor processevents abstractioreactor java at app org apache http impl nio reactor abstractioreactor execute abstractioreactor java at app org apache http impl nio reactor baseioreactor execute baseioreactor java at app org apache http impl nio reactor abstractmultiworkerioreactor worker run abstractmultiworkerioreactor java at java base java lang thread run thread java caused by org elasticsearch elasticsearchexception elasticsearch exception at app org elasticsearch elasticsearchexception innerfromxcontent elasticsearchexception java at app org elasticsearch elasticsearchexception fromxcontent elasticsearchexception java at app org elasticsearch elasticsearchexception innerfromxcontent elasticsearchexception java at app org elasticsearch elasticsearchexception failurefromxcontent elasticsearchexception java at app org elasticsearch rest bytesrestresponse errorfromxcontent bytesrestresponse java more
0
111,623
17,030,479,286
IssuesEvent
2021-07-04 13:07:13
turkdevops/next-auth
https://api.github.com/repos/turkdevops/next-auth
closed
CVE-2021-28092 (High) detected in is-svg-3.0.0.tgz - autoclosed
security vulnerability
## CVE-2021-28092 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary> <p>Check if a string or buffer is SVG</p> <p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p> <p>Path to dependency file: next-auth/package.json</p> <p>Path to vulnerable library: next-auth/node_modules/is-svg/package.json,next-auth/www/node_modules/is-svg/package.json</p> <p> Dependency Hierarchy: - cssnano-4.1.10.tgz (Root Library) - cssnano-preset-default-4.0.7.tgz - postcss-svgo-4.0.2.tgz - :x: **is-svg-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/next-auth/commit/d96c3a73550059033f41882dbe71f9064e527688">d96c3a73550059033f41882dbe71f9064e527688</a></p> <p>Found in base branch: <b>canary</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The is-svg package 2.1.0 through 4.2.1 for Node.js uses a regular expression that is vulnerable to Regular Expression Denial of Service (ReDoS). If an attacker provides a malicious string, is-svg will get stuck processing the input for a very long time. <p>Publish Date: 2021-03-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092>CVE-2021-28092</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092</a></p> <p>Release Date: 2021-03-12</p> <p>Fix Resolution: v4.2.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-28092 (High) detected in is-svg-3.0.0.tgz - autoclosed - ## CVE-2021-28092 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary> <p>Check if a string or buffer is SVG</p> <p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p> <p>Path to dependency file: next-auth/package.json</p> <p>Path to vulnerable library: next-auth/node_modules/is-svg/package.json,next-auth/www/node_modules/is-svg/package.json</p> <p> Dependency Hierarchy: - cssnano-4.1.10.tgz (Root Library) - cssnano-preset-default-4.0.7.tgz - postcss-svgo-4.0.2.tgz - :x: **is-svg-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/next-auth/commit/d96c3a73550059033f41882dbe71f9064e527688">d96c3a73550059033f41882dbe71f9064e527688</a></p> <p>Found in base branch: <b>canary</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The is-svg package 2.1.0 through 4.2.1 for Node.js uses a regular expression that is vulnerable to Regular Expression Denial of Service (ReDoS). If an attacker provides a malicious string, is-svg will get stuck processing the input for a very long time. <p>Publish Date: 2021-03-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092>CVE-2021-28092</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092</a></p> <p>Release Date: 2021-03-12</p> <p>Fix Resolution: v4.2.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in is svg tgz autoclosed cve high severity vulnerability vulnerable library is svg tgz check if a string or buffer is svg library home page a href path to dependency file next auth package json path to vulnerable library next auth node modules is svg package json next auth www node modules is svg package json dependency hierarchy cssnano tgz root library cssnano preset default tgz postcss svgo tgz x is svg tgz vulnerable library found in head commit a href found in base branch canary vulnerability details the is svg package through for node js uses a regular expression that is vulnerable to regular expression denial of service redos if an attacker provides a malicious string is svg will get stuck processing the input for a very long time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
4,180
20,112,696,646
IssuesEvent
2022-02-07 16:25:36
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[a11y]: <AVT1> The 'id' "downshift-8-label" specified for the ARIA property 'aria-labelledby' value is not valid
type: a11y ♿ component: dropdown status: waiting for maintainer response 💬
### Package carbon-components-react ### Browser Chrome ### Operating System MacOS ### Package version v7.50.0 ### React version v16.2.0 ### Automated testing tool and ruleset IBM Equal Access Accessibility Checker ### Assistive technology _No response_ ### Description ![image](https://user-images.githubusercontent.com/8787607/152258599-a3440451-5eb7-4fac-9775-caa927a94af3.png) Issue can be reproduced in https://react.carbondesignsystem.com/?path=/docs/components-dropdown--default Issue reproduced in code sandbox ![image](https://user-images.githubusercontent.com/8787607/152259189-45eb9f20-bd0f-41f9-b17d-017dcdfabd89.png) Originally issue was opened in dns-ui https://github.ibm.com/NetworkTribe/dns-ui/issues/1521 ### WCAG 2.1 Violation _No response_ ### CodeSandbox example https://codesandbox.io/s/5xrmq?file=/index.js ### Steps to reproduce 1. Render a dropdown with few items 2. Inspect on dropdown menu 3. You should see the "'id' specified for the ARIA property 'aria-labelledby' value is not valid" ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
True
[a11y]: <AVT1> The 'id' "downshift-8-label" specified for the ARIA property 'aria-labelledby' value is not valid - ### Package carbon-components-react ### Browser Chrome ### Operating System MacOS ### Package version v7.50.0 ### React version v16.2.0 ### Automated testing tool and ruleset IBM Equal Access Accessibility Checker ### Assistive technology _No response_ ### Description ![image](https://user-images.githubusercontent.com/8787607/152258599-a3440451-5eb7-4fac-9775-caa927a94af3.png) Issue can be reproduced in https://react.carbondesignsystem.com/?path=/docs/components-dropdown--default Issue reproduced in code sandbox ![image](https://user-images.githubusercontent.com/8787607/152259189-45eb9f20-bd0f-41f9-b17d-017dcdfabd89.png) Originally issue was opened in dns-ui https://github.ibm.com/NetworkTribe/dns-ui/issues/1521 ### WCAG 2.1 Violation _No response_ ### CodeSandbox example https://codesandbox.io/s/5xrmq?file=/index.js ### Steps to reproduce 1. Render a dropdown with few items 2. Inspect on dropdown menu 3. You should see the "'id' specified for the ARIA property 'aria-labelledby' value is not valid" ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
main
the id downshift label specified for the aria property aria labelledby value is not valid package carbon components react browser chrome operating system macos package version react version automated testing tool and ruleset ibm equal access accessibility checker assistive technology no response description issue can be reproduced in issue reproduced in code sandbox originally issue was opened in dns ui wcag violation no response codesandbox example steps to reproduce render a dropdown with few items inspect on dropdown menu you should see the id specified for the aria property aria labelledby value is not valid code of conduct i agree to follow this project s i checked the for duplicate problems
1
203,906
15,890,724,780
IssuesEvent
2021-04-10 16:28:27
veg-share/frontend-ui
https://api.github.com/repos/veg-share/frontend-ui
closed
Dev Issue : Initial SetUp
Developer Phase 1 documentation set-up
**What is the issue?** > Set up initial build of React app install updated dependencies establish local/remote repository connections for all members -[ ] https://www.npmjs.com/package/react-modal
1.0
Dev Issue : Initial SetUp - **What is the issue?** > Set up initial build of React app install updated dependencies establish local/remote repository connections for all members -[ ] https://www.npmjs.com/package/react-modal
non_main
dev issue initial setup what is the issue set up initial build of react app install updated dependencies establish local remote repository connections for all members
0
17,763
10,765,919,647
IssuesEvent
2019-11-01 12:26:48
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Can we change the service principal id/secret on agent nodes at runtime?
Pri1 container-service/svc cxp product-question triaged
Hi, let's say I specified a service principal "A" when creating my AKS cluster, and as described above the service principal info will be stored in file "/etc/kubernetes/azure.json" on agent node VMs. Due to some security concern, can I change the "azure.json" file to set another service principal "B" on agent node VMs only without changing the whole service principal used by AKS (which means not use "az aks update-credentials" command). We did some experiment that it works, but we'd also like to get confirmation from AKS production team to see whether this is an official supported approach and whether there is any side-effect of doing so. Thanks. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 25a0d502-ce71-76d4-bf8a-060e63913b52 * Version Independent ID: 69d7fa26-1a37-e477-0a2c-db58ce19a84f * Content: [Service principals for Azure Kubernetes Services (AKS)](https://docs.microsoft.com/en-us/azure/aks/kubernetes-service-principal) * Content Source: [articles/aks/kubernetes-service-principal.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/kubernetes-service-principal.md) * Service: **container-service** * GitHub Login: @mlearned * Microsoft Alias: **mlearned**
1.0
Can we change the service principal id/secret on agent nodes at runtime? - Hi, let's say I specified a service principal "A" when creating my AKS cluster, and as described above the service principal info will be stored in file "/etc/kubernetes/azure.json" on agent node VMs. Due to some security concern, can I change the "azure.json" file to set another service principal "B" on agent node VMs only without changing the whole service principal used by AKS (which means not use "az aks update-credentials" command). We did some experiment that it works, but we'd also like to get confirmation from AKS production team to see whether this is an official supported approach and whether there is any side-effect of doing so. Thanks. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 25a0d502-ce71-76d4-bf8a-060e63913b52 * Version Independent ID: 69d7fa26-1a37-e477-0a2c-db58ce19a84f * Content: [Service principals for Azure Kubernetes Services (AKS)](https://docs.microsoft.com/en-us/azure/aks/kubernetes-service-principal) * Content Source: [articles/aks/kubernetes-service-principal.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/kubernetes-service-principal.md) * Service: **container-service** * GitHub Login: @mlearned * Microsoft Alias: **mlearned**
non_main
can we change the service principal id secret on agent nodes at runtime hi let s say i specified a service principal a when creating my aks cluster and as described above the service principal info will be stored in file etc kubernetes azure json on agent node vms due to some security concern can i change the azure json file to set another service principal b on agent node vms only without changing the whole service principal used by aks which means not use az aks update credentials command we did some experiment that it works but we d also like to get confirmation from aks production team to see whether this is an official supported approach and whether there is any side effect of doing so thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned
0
144,545
19,287,858,868
IssuesEvent
2021-12-11 08:53:39
turkdevops/electron
https://api.github.com/repos/turkdevops/electron
closed
CVE-2012-6708 (Medium) detected in jquery-1.7.2.min.js - autoclosed
security vulnerability
## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: electron/node_modules/jmespath/index.html</p> <p>Path to vulnerable library: /node_modules/jmespath/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/electron/commit/5da3bcb84e69be57f16313e78dd435a9af067d0d">5da3bcb84e69be57f16313e78dd435a9af067d0d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2012-6708 (Medium) detected in jquery-1.7.2.min.js - autoclosed - ## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: electron/node_modules/jmespath/index.html</p> <p>Path to vulnerable library: /node_modules/jmespath/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/electron/commit/5da3bcb84e69be57f16313e78dd435a9af067d0d">5da3bcb84e69be57f16313e78dd435a9af067d0d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in jquery min js autoclosed cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file electron node modules jmespath index html path to vulnerable library node modules jmespath index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
78,822
10,090,549,621
IssuesEvent
2019-07-26 12:00:12
XanaduAI/pennylane
https://api.github.com/repos/XanaduAI/pennylane
closed
QubitUnitary does not act on all wires
documentation
#### Issue description Description of the issue - In the documentation, QubitUnitary operator: 1. is stated to act on `all wires` in the system. I tested it and it can act on subsytem of wires. 2. is said to support finite difference as its gradient recipe. This can not be right as we can only input a fixed, non-differentiable matrix to this operator. So the gradient recipe should be `None`. I am making a PR for this issue.
1.0
QubitUnitary does not act on all wires - #### Issue description Description of the issue - In the documentation, QubitUnitary operator: 1. is stated to act on `all wires` in the system. I tested it and it can act on subsytem of wires. 2. is said to support finite difference as its gradient recipe. This can not be right as we can only input a fixed, non-differentiable matrix to this operator. So the gradient recipe should be `None`. I am making a PR for this issue.
non_main
qubitunitary does not act on all wires issue description description of the issue in the documentation qubitunitary operator is stated to act on all wires in the system i tested it and it can act on subsytem of wires is said to support finite difference as its gradient recipe this can not be right as we can only input a fixed non differentiable matrix to this operator so the gradient recipe should be none i am making a pr for this issue
0
220,760
17,259,335,804
IssuesEvent
2021-07-22 04:08:26
DeFiCh/jellyfish
https://api.github.com/repos/DeFiCh/jellyfish
closed
`MasterNodeRegTestContainer` block gen feature should be controlled by testcontainers
area/testcontainers kind/refactor priority/important-soon triage/accepted
<!-- Please only use this template for submitting enhancement/feature requests --> #### What would you like to be added: `MasterNodeRegTestContainer` block generation should be automated by testcontainers. `@defichain/testcontainers` should be created such that block generation can only be done by code within testcontainers. `gen=1` was set to automated the generation process on ain side. #### Why is this needed: Having automated block gen on the ain side is unpredictable and creates too many conditions for "side-effect" when testing on jellyfish. This effect is cascaded down to whale, playground and potentially wallet implementation. #### Additional comments: https://github.com/DeFiCh/playground uses its own block generation interval, we can follow that implementation reference. https://github.com/DeFiCh/playground/blob/09b29fe1d47f48ac75283ecdda85c234c4bf94c2/src/module.playground/playground.block.ts#L13-L17
1.0
`MasterNodeRegTestContainer` block gen feature should be controlled by testcontainers - <!-- Please only use this template for submitting enhancement/feature requests --> #### What would you like to be added: `MasterNodeRegTestContainer` block generation should be automated by testcontainers. `@defichain/testcontainers` should be created such that block generation can only be done by code within testcontainers. `gen=1` was set to automated the generation process on ain side. #### Why is this needed: Having automated block gen on the ain side is unpredictable and creates too many conditions for "side-effect" when testing on jellyfish. This effect is cascaded down to whale, playground and potentially wallet implementation. #### Additional comments: https://github.com/DeFiCh/playground uses its own block generation interval, we can follow that implementation reference. https://github.com/DeFiCh/playground/blob/09b29fe1d47f48ac75283ecdda85c234c4bf94c2/src/module.playground/playground.block.ts#L13-L17
non_main
masternoderegtestcontainer block gen feature should be controlled by testcontainers what would you like to be added masternoderegtestcontainer block generation should be automated by testcontainers defichain testcontainers should be created such that block generation can only be done by code within testcontainers gen was set to automated the generation process on ain side why is this needed having automated block gen on the ain side is unpredictable and creates too many conditions for side effect when testing on jellyfish this effect is cascaded down to whale playground and potentially wallet implementation additional comments uses its own block generation interval we can follow that implementation reference
0
769,335
27,002,020,038
IssuesEvent
2023-02-10 08:40:16
shaka-project/shaka-player
https://api.github.com/repos/shaka-project/shaka-player
closed
DASH unprotected content stuck on the first frame on Samsung Tizen 5.5
type: bug flag: seeking PR status: waiting on response priority: P2 component: DASH platform: Tizen
@joeyparrish I have a potentially similar issue, both to #3046 and to #2620. Basically in a few of several attempts, the DASH stream fails to start properly, it's just stuck on the first frame, but can be initiated if the user seeks forward. From the logs, I can see that the Shaka player has detected a stall, and tries to unstall by pause/play, but with no success. Only reproducible on Tizen 5.5 (2020). I've been testing on model: UE43TU7072UXXH, but it's reproducible on other 5.5 models as well. Shaka player version: 3.1.2 The stream where this can be reproduced (not consistently): https://trailers.redbox.com/assets/211568/dash/211568.mpd And it's not specific to this DASH stream, can be reproduced with others as well, I think that it's happening a bit more often with it. Logs: <details> <summary>Details</summary> ![Screenshot 2021-08-18 at 15 14 00](https://user-images.githubusercontent.com/25148427/129904456-1e7b2752-8488-4298-b0a3-7031b1106f0b.png) ![Screenshot 2021-08-18 at 15 14 06](https://user-images.githubusercontent.com/25148427/129904468-d9f59a2c-6395-4200-a60f-64db32e2c82f.png) </details> Shaka player config: ```js { streaming: { ignoreTextStreamFailures: true, // https://shaka-player-demo.appspot.com/docs/api/tutorial-network-and-buffering-config.html bufferingGoal: 60, // how much to buffer while the video is playing, to have enough content in case of network hiccups rebufferingGoal: 3, // how much to buffer before resuming playing, aka on video start and after seeking video bufferBehind: 30, jumpLargeGaps: true, stallSkip: 0.1, retryParameters: { maxAttempts: 10, baseDelay: 1000, backoffFactor: 1.75, fuzzFactor: 0.5, timeout: 10000, }, }, manifest: { dash: { ignoreMinBufferTime: true, }, retryParameters: { maxAttempts: 5, baseDelay: 1000, backoffFactor: 2, fuzzFactor: 0.5, timeout: 25000, }, }, drm: { servers: getDrmLicenseServers(), advanced: { 'com.widevine.alpha': { videoRobustness: 'SW_SECURE_CRYPTO', audioRobustness: 'SW_SECURE_CRYPTO', }, 'com.microsoft.playready': withAltTVODDrmConfig ? { distinctiveIdentifierRequired: true, sessionType: 'persistent-license', } : {}, }, retryParameters: { maxAttempts: 5, baseDelay: 1000, backoffFactor: 2, fuzzFactor: 0.5, timeout: 10000, }, }, restrictions: { minHeight: 360, }, preferredAudioChannelCount: 6, } ``` I've also tried playing with the default configuration instead of the one above, and the issue persists. The content sometimes starts playing after some time, but I could not reproduce this consistently. Also, looks like there is a lot of calls to switch after the shaka player loads the content: ```log 13:01:41.638 mathRound.install 13:01:41.642 MediaSource.install 13:01:41.645 Using native MSE as-is. 13:01:41.649 VideoPlayPromise.install 13:01:41.652 Using native VTTCue. 13:01:41.656 MediaCapabilities: install 13:01:41.659 MediaCapabilities: Native mediaCapabilities support found. 13:01:42.060 mathRound.install 13:01:42.073 MediaSource.install 13:01:42.076 Using native MSE as-is. 13:01:42.079 VideoPlayPromise.install 13:01:42.081 Using native VTTCue. 13:01:42.084 MediaCapabilities: install 13:01:42.087 MediaCapabilities: Native mediaCapabilities support found. 13:01:43.912 Starting attach... 13:01:45.306 Starting load of https://trailers.redbox.com/assets/212464/dash/212464.mpd... 13:01:45.433 Found variant with audio and video content, so filtering out audio-only content. 13:01:45.482 codecs avc1-mp4a avg bandwidth 8596000 13:01:45.543 init: completed initial Stream setup 13:01:45.567 After load: https://trailers.redbox.com/assets/212464/dash/212464.mpd 13:01:52.201 Calling switch_(), bandwidth=1048 kbps 13:01:52.210 switch_ 13:01:52.218 switch: switching to Stream (video:6) 13:01:52.233 switch: Stream (audio:8) already active 13:02:00.656 Calling switch_(), bandwidth=826 kbps 13:02:00.667 switch_ 13:02:09.045 Calling switch_(), bandwidth=987 kbps 13:02:09.057 switch_ 13:02:17.317 Calling switch_(), bandwidth=951 kbps 13:02:17.329 switch_ 13:02:25.496 Calling switch_(), bandwidth=934 kbps 13:02:25.507 switch_ 13:02:34.148 Calling switch_(), bandwidth=891 kbps 13:02:34.158 switch_ 13:02:42.549 Calling switch_(), bandwidth=898 kbps 13:02:42.560 switch_ 13:02:51.073 Calling switch_(), bandwidth=891 kbps 13:02:51.086 switch_ 13:02:59.432 Calling switch_(), bandwidth=856 kbps 13:02:59.443 switch_ 13:03:07.906 Calling switch_(), bandwidth=844 kbps 13:03:07.917 switch_ 13:03:16.035 Calling switch_(), bandwidth=819 kbps 13:03:16.047 switch_ 13:03:24.335 Calling switch_(), bandwidth=850 kbps 13:03:24.348 switch_ 13:03:33.144 Calling switch_(), bandwidth=822 kbps 13:03:33.157 switch_ 13:03:41.333 Calling switch_(), bandwidth=807 kbps 13:03:41.348 switch_ 13:03:49.476 Calling switch_(), bandwidth=808 kbps 13:03:49.489 switch_ 13:03:57.587 Calling switch_(), bandwidth=808 kbps 13:03:57.599 switch_ 13:04:05.866 Calling switch_(), bandwidth=796 kbps 13:04:05.876 switch_ 13:04:14.165 Calling switch_(), bandwidth=747 kbps 13:04:14.176 switch_ 13:04:22.345 Calling switch_(), bandwidth=731 kbps 13:04:22.357 switch_ 13:04:30.521 Calling switch_(), bandwidth=732 kbps 13:04:30.531 switch_ 13:04:39.179 Calling switch_(), bandwidth=706 kbps 13:04:39.196 switch_ 13:04:47.848 Calling switch_(), bandwidth=659 kbps 13:04:47.863 switch_ 13:04:56.334 Calling switch_(), bandwidth=699 kbps 13:04:56.344 switch_ 13:05:04.836 Calling switch_(), bandwidth=700 kbps 13:05:04.846 switch_ 13:05:13.258 Calling switch_(), bandwidth=715 kbps 13:05:13.267 switch_ 13:05:21.740 Calling switch_(), bandwidth=662 kbps 13:05:21.750 switch_ 13:05:30.741 Calling switch_(), bandwidth=704 kbps 13:05:30.749 switch_ 13:05:39.415 Calling switch_(), bandwidth=694 kbps 13:05:39.424 switch_ 13:05:47.876 Calling switch_(), bandwidth=655 kbps 13:05:47.885 switch_ 13:05:56.757 Calling switch_(), bandwidth=631 kbps 13:05:56.768 switch_ 13:06:05.725 Calling switch_(), bandwidth=656 kbps 13:06:05.734 switch_ 13:06:14.687 Calling switch_(), bandwidth=664 kbps 13:06:14.697 switch_ 13:06:23.384 Calling switch_(), bandwidth=645 kbps 13:06:23.394 switch_ 13:06:31.451 Calling switch_(), bandwidth=620 kbps 13:06:31.459 switch_ 13:06:40.143 Calling switch_(), bandwidth=625 kbps 13:06:40.152 switch_ 13:06:48.708 Calling switch_(), bandwidth=638 kbps 13:06:48.719 switch_ 13:06:57.602 Calling switch_(), bandwidth=625 kbps 13:06:57.613 switch_ 13:07:05.908 Calling switch_(), bandwidth=600 kbps 13:07:05.918 switch_ 13:07:14.040 Calling switch_(), bandwidth=623 kbps 13:07:14.053 switch_ 13:07:22.273 Calling switch_(), bandwidth=622 kbps 13:07:22.282 switch_ 13:07:30.844 Calling switch_(), bandwidth=547 kbps 13:07:30.855 switch_ 13:07:39.932 Calling switch_(), bandwidth=550 kbps 13:07:39.944 switch_ 13:07:48.012 Calling switch_(), bandwidth=543 kbps 13:07:48.024 switch_ 13:07:56.416 Calling switch_(), bandwidth=584 kbps 13:07:56.428 switch_ 13:08:05.005 Calling switch_(), bandwidth=595 kbps 13:08:05.017 switch_ 13:08:13.812 Calling switch_(), bandwidth=534 kbps 13:08:13.829 switch_ 13:08:20.219 Jumping forward 0.033 seconds because of gap before start time of 0.033 13:08:21.137 (all): seeked: buffered seek: presentationTime=0.033 13:08:25.570 Jumping forward 0.033 seconds because of gap before start time of 0.033 13:08:25.596 (all): seeked: buffered seek: presentationTime=0.033 13:08:27.965 Calling switch_(), bandwidth=534 kbps 13:08:27.972 switch_ ``` In the log above, the playback started after 6 minutes... Because of this, I would say that this is also related to #3076. Any suggestion on how to debug/fix this would be much appreciated. _Originally posted by @Puritanic in https://github.com/google/shaka-player/issues/3046#issuecomment-901113523_
1.0
DASH unprotected content stuck on the first frame on Samsung Tizen 5.5 - @joeyparrish I have a potentially similar issue, both to #3046 and to #2620. Basically in a few of several attempts, the DASH stream fails to start properly, it's just stuck on the first frame, but can be initiated if the user seeks forward. From the logs, I can see that the Shaka player has detected a stall, and tries to unstall by pause/play, but with no success. Only reproducible on Tizen 5.5 (2020). I've been testing on model: UE43TU7072UXXH, but it's reproducible on other 5.5 models as well. Shaka player version: 3.1.2 The stream where this can be reproduced (not consistently): https://trailers.redbox.com/assets/211568/dash/211568.mpd And it's not specific to this DASH stream, can be reproduced with others as well, I think that it's happening a bit more often with it. Logs: <details> <summary>Details</summary> ![Screenshot 2021-08-18 at 15 14 00](https://user-images.githubusercontent.com/25148427/129904456-1e7b2752-8488-4298-b0a3-7031b1106f0b.png) ![Screenshot 2021-08-18 at 15 14 06](https://user-images.githubusercontent.com/25148427/129904468-d9f59a2c-6395-4200-a60f-64db32e2c82f.png) </details> Shaka player config: ```js { streaming: { ignoreTextStreamFailures: true, // https://shaka-player-demo.appspot.com/docs/api/tutorial-network-and-buffering-config.html bufferingGoal: 60, // how much to buffer while the video is playing, to have enough content in case of network hiccups rebufferingGoal: 3, // how much to buffer before resuming playing, aka on video start and after seeking video bufferBehind: 30, jumpLargeGaps: true, stallSkip: 0.1, retryParameters: { maxAttempts: 10, baseDelay: 1000, backoffFactor: 1.75, fuzzFactor: 0.5, timeout: 10000, }, }, manifest: { dash: { ignoreMinBufferTime: true, }, retryParameters: { maxAttempts: 5, baseDelay: 1000, backoffFactor: 2, fuzzFactor: 0.5, timeout: 25000, }, }, drm: { servers: getDrmLicenseServers(), advanced: { 'com.widevine.alpha': { videoRobustness: 'SW_SECURE_CRYPTO', audioRobustness: 'SW_SECURE_CRYPTO', }, 'com.microsoft.playready': withAltTVODDrmConfig ? { distinctiveIdentifierRequired: true, sessionType: 'persistent-license', } : {}, }, retryParameters: { maxAttempts: 5, baseDelay: 1000, backoffFactor: 2, fuzzFactor: 0.5, timeout: 10000, }, }, restrictions: { minHeight: 360, }, preferredAudioChannelCount: 6, } ``` I've also tried playing with the default configuration instead of the one above, and the issue persists. The content sometimes starts playing after some time, but I could not reproduce this consistently. Also, looks like there is a lot of calls to switch after the shaka player loads the content: ```log 13:01:41.638 mathRound.install 13:01:41.642 MediaSource.install 13:01:41.645 Using native MSE as-is. 13:01:41.649 VideoPlayPromise.install 13:01:41.652 Using native VTTCue. 13:01:41.656 MediaCapabilities: install 13:01:41.659 MediaCapabilities: Native mediaCapabilities support found. 13:01:42.060 mathRound.install 13:01:42.073 MediaSource.install 13:01:42.076 Using native MSE as-is. 13:01:42.079 VideoPlayPromise.install 13:01:42.081 Using native VTTCue. 13:01:42.084 MediaCapabilities: install 13:01:42.087 MediaCapabilities: Native mediaCapabilities support found. 13:01:43.912 Starting attach... 13:01:45.306 Starting load of https://trailers.redbox.com/assets/212464/dash/212464.mpd... 13:01:45.433 Found variant with audio and video content, so filtering out audio-only content. 13:01:45.482 codecs avc1-mp4a avg bandwidth 8596000 13:01:45.543 init: completed initial Stream setup 13:01:45.567 After load: https://trailers.redbox.com/assets/212464/dash/212464.mpd 13:01:52.201 Calling switch_(), bandwidth=1048 kbps 13:01:52.210 switch_ 13:01:52.218 switch: switching to Stream (video:6) 13:01:52.233 switch: Stream (audio:8) already active 13:02:00.656 Calling switch_(), bandwidth=826 kbps 13:02:00.667 switch_ 13:02:09.045 Calling switch_(), bandwidth=987 kbps 13:02:09.057 switch_ 13:02:17.317 Calling switch_(), bandwidth=951 kbps 13:02:17.329 switch_ 13:02:25.496 Calling switch_(), bandwidth=934 kbps 13:02:25.507 switch_ 13:02:34.148 Calling switch_(), bandwidth=891 kbps 13:02:34.158 switch_ 13:02:42.549 Calling switch_(), bandwidth=898 kbps 13:02:42.560 switch_ 13:02:51.073 Calling switch_(), bandwidth=891 kbps 13:02:51.086 switch_ 13:02:59.432 Calling switch_(), bandwidth=856 kbps 13:02:59.443 switch_ 13:03:07.906 Calling switch_(), bandwidth=844 kbps 13:03:07.917 switch_ 13:03:16.035 Calling switch_(), bandwidth=819 kbps 13:03:16.047 switch_ 13:03:24.335 Calling switch_(), bandwidth=850 kbps 13:03:24.348 switch_ 13:03:33.144 Calling switch_(), bandwidth=822 kbps 13:03:33.157 switch_ 13:03:41.333 Calling switch_(), bandwidth=807 kbps 13:03:41.348 switch_ 13:03:49.476 Calling switch_(), bandwidth=808 kbps 13:03:49.489 switch_ 13:03:57.587 Calling switch_(), bandwidth=808 kbps 13:03:57.599 switch_ 13:04:05.866 Calling switch_(), bandwidth=796 kbps 13:04:05.876 switch_ 13:04:14.165 Calling switch_(), bandwidth=747 kbps 13:04:14.176 switch_ 13:04:22.345 Calling switch_(), bandwidth=731 kbps 13:04:22.357 switch_ 13:04:30.521 Calling switch_(), bandwidth=732 kbps 13:04:30.531 switch_ 13:04:39.179 Calling switch_(), bandwidth=706 kbps 13:04:39.196 switch_ 13:04:47.848 Calling switch_(), bandwidth=659 kbps 13:04:47.863 switch_ 13:04:56.334 Calling switch_(), bandwidth=699 kbps 13:04:56.344 switch_ 13:05:04.836 Calling switch_(), bandwidth=700 kbps 13:05:04.846 switch_ 13:05:13.258 Calling switch_(), bandwidth=715 kbps 13:05:13.267 switch_ 13:05:21.740 Calling switch_(), bandwidth=662 kbps 13:05:21.750 switch_ 13:05:30.741 Calling switch_(), bandwidth=704 kbps 13:05:30.749 switch_ 13:05:39.415 Calling switch_(), bandwidth=694 kbps 13:05:39.424 switch_ 13:05:47.876 Calling switch_(), bandwidth=655 kbps 13:05:47.885 switch_ 13:05:56.757 Calling switch_(), bandwidth=631 kbps 13:05:56.768 switch_ 13:06:05.725 Calling switch_(), bandwidth=656 kbps 13:06:05.734 switch_ 13:06:14.687 Calling switch_(), bandwidth=664 kbps 13:06:14.697 switch_ 13:06:23.384 Calling switch_(), bandwidth=645 kbps 13:06:23.394 switch_ 13:06:31.451 Calling switch_(), bandwidth=620 kbps 13:06:31.459 switch_ 13:06:40.143 Calling switch_(), bandwidth=625 kbps 13:06:40.152 switch_ 13:06:48.708 Calling switch_(), bandwidth=638 kbps 13:06:48.719 switch_ 13:06:57.602 Calling switch_(), bandwidth=625 kbps 13:06:57.613 switch_ 13:07:05.908 Calling switch_(), bandwidth=600 kbps 13:07:05.918 switch_ 13:07:14.040 Calling switch_(), bandwidth=623 kbps 13:07:14.053 switch_ 13:07:22.273 Calling switch_(), bandwidth=622 kbps 13:07:22.282 switch_ 13:07:30.844 Calling switch_(), bandwidth=547 kbps 13:07:30.855 switch_ 13:07:39.932 Calling switch_(), bandwidth=550 kbps 13:07:39.944 switch_ 13:07:48.012 Calling switch_(), bandwidth=543 kbps 13:07:48.024 switch_ 13:07:56.416 Calling switch_(), bandwidth=584 kbps 13:07:56.428 switch_ 13:08:05.005 Calling switch_(), bandwidth=595 kbps 13:08:05.017 switch_ 13:08:13.812 Calling switch_(), bandwidth=534 kbps 13:08:13.829 switch_ 13:08:20.219 Jumping forward 0.033 seconds because of gap before start time of 0.033 13:08:21.137 (all): seeked: buffered seek: presentationTime=0.033 13:08:25.570 Jumping forward 0.033 seconds because of gap before start time of 0.033 13:08:25.596 (all): seeked: buffered seek: presentationTime=0.033 13:08:27.965 Calling switch_(), bandwidth=534 kbps 13:08:27.972 switch_ ``` In the log above, the playback started after 6 minutes... Because of this, I would say that this is also related to #3076. Any suggestion on how to debug/fix this would be much appreciated. _Originally posted by @Puritanic in https://github.com/google/shaka-player/issues/3046#issuecomment-901113523_
non_main
dash unprotected content stuck on the first frame on samsung tizen joeyparrish i have a potentially similar issue both to and to basically in a few of several attempts the dash stream fails to start properly it s just stuck on the first frame but can be initiated if the user seeks forward from the logs i can see that the shaka player has detected a stall and tries to unstall by pause play but with no success only reproducible on tizen i ve been testing on model but it s reproducible on other models as well shaka player version the stream where this can be reproduced not consistently and it s not specific to this dash stream can be reproduced with others as well i think that it s happening a bit more often with it logs details shaka player config js streaming ignoretextstreamfailures true bufferinggoal how much to buffer while the video is playing to have enough content in case of network hiccups rebufferinggoal how much to buffer before resuming playing aka on video start and after seeking video bufferbehind jumplargegaps true stallskip retryparameters maxattempts basedelay backofffactor fuzzfactor timeout manifest dash ignoreminbuffertime true retryparameters maxattempts basedelay backofffactor fuzzfactor timeout drm servers getdrmlicenseservers advanced com widevine alpha videorobustness sw secure crypto audiorobustness sw secure crypto com microsoft playready withalttvoddrmconfig distinctiveidentifierrequired true sessiontype persistent license retryparameters maxattempts basedelay backofffactor fuzzfactor timeout restrictions minheight preferredaudiochannelcount i ve also tried playing with the default configuration instead of the one above and the issue persists the content sometimes starts playing after some time but i could not reproduce this consistently also looks like there is a lot of calls to switch after the shaka player loads the content log mathround install mediasource install using native mse as is videoplaypromise install using native vttcue mediacapabilities install mediacapabilities native mediacapabilities support found mathround install mediasource install using native mse as is videoplaypromise install using native vttcue mediacapabilities install mediacapabilities native mediacapabilities support found starting attach starting load of found variant with audio and video content so filtering out audio only content codecs avg bandwidth init completed initial stream setup after load calling switch bandwidth kbps switch switch switching to stream video switch stream audio already active calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch jumping forward seconds because of gap before start time of all seeked buffered seek presentationtime jumping forward seconds because of gap before start time of all seeked buffered seek presentationtime calling switch bandwidth kbps switch in the log above the playback started after minutes because of this i would say that this is also related to any suggestion on how to debug fix this would be much appreciated originally posted by puritanic in
0
652,736
21,560,106,163
IssuesEvent
2022-05-01 03:03:55
devs-immortal/Paradise-Lost
https://api.github.com/repos/devs-immortal/Paradise-Lost
closed
Portal not working!!!
Bug duplicate Priority More Info Required
### What happened? he animation goes along with the sound, but the portal doesn't work. I tried to play without other mods, it did not help (Maybe the portal does not work well with worlds that were created before installing). https://www.youtube.com/watch?v=U3udsIqGWI4 ### Mod Version 1.6.5 - 1.18 ### Fabric API Version last ### Relevant log output _No response_ ### Other mods _No response_ ### Additional Information https://www.youtube.com/watch?v=U3udsIqGWI4
1.0
Portal not working!!! - ### What happened? he animation goes along with the sound, but the portal doesn't work. I tried to play without other mods, it did not help (Maybe the portal does not work well with worlds that were created before installing). https://www.youtube.com/watch?v=U3udsIqGWI4 ### Mod Version 1.6.5 - 1.18 ### Fabric API Version last ### Relevant log output _No response_ ### Other mods _No response_ ### Additional Information https://www.youtube.com/watch?v=U3udsIqGWI4
non_main
portal not working what happened he animation goes along with the sound but the portal doesn t work i tried to play without other mods it did not help maybe the portal does not work well with worlds that were created before installing mod version fabric api version last relevant log output no response other mods no response additional information
0
737
4,330,673,379
IssuesEvent
2016-07-26 20:47:13
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
opened
Untappd: Update triggering? Example query not working
Maintainer Input Requested
One of the supplied example queries doesn't show the Untappd IA: https://duckduckgo.com/?q=bommen%20%26%20granaten%20barleywine&ia=food&drink ------ IA Page: http://duck.co/ia/view/untappd [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @urbanserj
True
Untappd: Update triggering? Example query not working - One of the supplied example queries doesn't show the Untappd IA: https://duckduckgo.com/?q=bommen%20%26%20granaten%20barleywine&ia=food&drink ------ IA Page: http://duck.co/ia/view/untappd [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @urbanserj
main
untappd update triggering example query not working one of the supplied example queries doesn t show the untappd ia ia page urbanserj
1
2,592
8,817,635,485
IssuesEvent
2018-12-31 03:07:10
mlabbe/nativefiledialog
https://api.github.com/repos/mlabbe/nativefiledialog
closed
ShellItem not released on error in Windows implementation
bug maintainer-confirmed windows
Hi, The following `ShellItem` is not released if an error is returned while using it (the function jumps to the end): https://github.com/mlabbe/nativefiledialog/blob/5cfe5002eb0fac1e49777a17dec70134147931e2/src/nfd_win.cpp#L403 Similarly, the file dialog itself is never released. Compare it to [the official example](https://msdn.microsoft.com/en-us/library/windows/desktop/bb776913(v=vs.85).aspx#usage).
True
ShellItem not released on error in Windows implementation - Hi, The following `ShellItem` is not released if an error is returned while using it (the function jumps to the end): https://github.com/mlabbe/nativefiledialog/blob/5cfe5002eb0fac1e49777a17dec70134147931e2/src/nfd_win.cpp#L403 Similarly, the file dialog itself is never released. Compare it to [the official example](https://msdn.microsoft.com/en-us/library/windows/desktop/bb776913(v=vs.85).aspx#usage).
main
shellitem not released on error in windows implementation hi the following shellitem is not released if an error is returned while using it the function jumps to the end similarly the file dialog itself is never released compare it to
1
121,747
16,019,876,350
IssuesEvent
2021-04-20 21:08:24
equinor/design-system
https://api.github.com/repos/equinor/design-system
opened
Create color scheme for oil and gas products
design feature request
Unwritten standards exist in Equinor for visualizing products. Oil = green gas = red NGL = yellow Condensate and/or water = blue Some info about this [here from NPD:](https://statoilsrm.sharepoint.com/sites/Designsystem/_layouts/OneNote.aspx?id=%2Fsites%2FDesignsystem%2FShared%20Documents%2FGeneral%2FPower%20BI%2FExamples%2FPower%20BI%20examples&wd=target%28Color%20schemes.one%7CC461E3B1-30B9-4CD3-AE59-C6112EFAADA9%2FNPD%7CE603561B-EE51-49CE-BF9F-40F19141BBDF%2F%29 onenote:https://statoilsrm.sharepoint.com/sites/Designsystem/Shared%20Documents/General/Power%20BI/Examples/Power%20BI%20examples/Color%20schemes.one#NPD&section-id={C461E3B1-30B9-4CD3-AE59-C6112EFAADA9}&page-id={E603561B-EE51-49CE-BF9F-40F19141BBDF}&end)
1.0
Create color scheme for oil and gas products - Unwritten standards exist in Equinor for visualizing products. Oil = green gas = red NGL = yellow Condensate and/or water = blue Some info about this [here from NPD:](https://statoilsrm.sharepoint.com/sites/Designsystem/_layouts/OneNote.aspx?id=%2Fsites%2FDesignsystem%2FShared%20Documents%2FGeneral%2FPower%20BI%2FExamples%2FPower%20BI%20examples&wd=target%28Color%20schemes.one%7CC461E3B1-30B9-4CD3-AE59-C6112EFAADA9%2FNPD%7CE603561B-EE51-49CE-BF9F-40F19141BBDF%2F%29 onenote:https://statoilsrm.sharepoint.com/sites/Designsystem/Shared%20Documents/General/Power%20BI/Examples/Power%20BI%20examples/Color%20schemes.one#NPD&section-id={C461E3B1-30B9-4CD3-AE59-C6112EFAADA9}&page-id={E603561B-EE51-49CE-BF9F-40F19141BBDF}&end)
non_main
create color scheme for oil and gas products unwritten standards exist in equinor for visualizing products oil green gas red ngl yellow condensate and or water blue some info about this onenote
0
39,048
10,282,202,527
IssuesEvent
2019-08-26 10:27:34
letscontrolit/ESPEasy
https://api.github.com/repos/letscontrolit/ESPEasy
closed
Puya image - where to find or what to use
Category: Build Type: Question
ESP-01S to be flashed with puya image. Read a lot, but what image to be used? In the releases there is no image name with "puya" in it? https://github.com/letscontrolit/ESPEasy/releases
1.0
Puya image - where to find or what to use - ESP-01S to be flashed with puya image. Read a lot, but what image to be used? In the releases there is no image name with "puya" in it? https://github.com/letscontrolit/ESPEasy/releases
non_main
puya image where to find or what to use esp to be flashed with puya image read a lot but what image to be used in the releases there is no image name with puya in it
0
4,957
25,457,718,309
IssuesEvent
2022-11-24 15:31:22
permon/permon
https://api.github.com/repos/permon/permon
opened
Review petscfix functions
maintainability
These functions should be contributed to PETSc or removed if already fixed
True
Review petscfix functions - These functions should be contributed to PETSc or removed if already fixed
main
review petscfix functions these functions should be contributed to petsc or removed if already fixed
1
4,938
25,376,858,620
IssuesEvent
2022-11-21 14:43:15
ClubInfoInsaT/application-amicale
https://api.github.com/repos/ClubInfoInsaT/application-amicale
closed
Mettre à jour le logo de proximo
enhancement maintainers
Le nouveau logo est sur le discord. Se fait directement sur le site etud.
True
Mettre à jour le logo de proximo - Le nouveau logo est sur le discord. Se fait directement sur le site etud.
main
mettre à jour le logo de proximo le nouveau logo est sur le discord se fait directement sur le site etud
1
244,881
18,768,825,426
IssuesEvent
2021-11-06 12:48:43
Team-Hydra-Discord/Feedback
https://api.github.com/repos/Team-Hydra-Discord/Feedback
closed
[Documentation] Create AppBot Docs
Documentation
### Describe The Issue With This Content Create AppBot docs. ### Where Does This Issue Reside? ```bash Team Hydra Docs ``` ### Expected Content Full AppBot Docs ### Additional Context _No response_
1.0
[Documentation] Create AppBot Docs - ### Describe The Issue With This Content Create AppBot docs. ### Where Does This Issue Reside? ```bash Team Hydra Docs ``` ### Expected Content Full AppBot Docs ### Additional Context _No response_
non_main
create appbot docs describe the issue with this content create appbot docs where does this issue reside bash team hydra docs expected content full appbot docs additional context no response
0
3,219
12,337,823,476
IssuesEvent
2020-05-14 15:32:13
chaoss/website
https://api.github.com/repos/chaoss/website
closed
Request to get my article featured on CHAOSS Website Blog
Maintainer Task
Good day all, I got accepted as an Outreachy intern for CHAOSS, we are expected to blog every 2 weeks. The articles are meant to be featured in the community website blog or any blog currently maintained by the community. I wrote my first article on "How i got accepted into the Outreachy Internship". Please read here https://docs.google.com/document/d/1vF69kwXMYtuzl9S0jXyOyS0AeSDga9SJet054ooIIWU/edit?usp=sharing
True
Request to get my article featured on CHAOSS Website Blog - Good day all, I got accepted as an Outreachy intern for CHAOSS, we are expected to blog every 2 weeks. The articles are meant to be featured in the community website blog or any blog currently maintained by the community. I wrote my first article on "How i got accepted into the Outreachy Internship". Please read here https://docs.google.com/document/d/1vF69kwXMYtuzl9S0jXyOyS0AeSDga9SJet054ooIIWU/edit?usp=sharing
main
request to get my article featured on chaoss website blog good day all i got accepted as an outreachy intern for chaoss we are expected to blog every weeks the articles are meant to be featured in the community website blog or any blog currently maintained by the community i wrote my first article on how i got accepted into the outreachy internship please read here
1
5,701
30,022,705,657
IssuesEvent
2023-06-27 01:50:40
darekkay/todoist-export
https://api.github.com/repos/darekkay/todoist-export
closed
Improve state passing
Type: Maintainance
Form options (CSV/JSON) are currently passed through the OAuth `state` string parameter. This complicates passing state and is nonstandard. Instead, we should be storing this state locally with the `state` token as a key (and ideally validating the request): > If state is used for carrying application state, and integrity of its contents is a concern, clients MUST protect state against tampering and swapping. This can be achieved by binding the contents of state to the browser session and/or signed/encrypted state values See: * https://tools.ietf.org/html/draft-bradley-oauth-jwt-encoded-state-09 * http://www.thread-safe.com/2014/05/the-correct-use-of-state-parameter-in.html * https://stackoverflow.com/questions/52459476/does-oauth-state-mitigate-any-genuinely-dangerous-attacks * https://pipedrive.readme.io/docs/marketplace-oauth-authorization-state-parameter
True
Improve state passing - Form options (CSV/JSON) are currently passed through the OAuth `state` string parameter. This complicates passing state and is nonstandard. Instead, we should be storing this state locally with the `state` token as a key (and ideally validating the request): > If state is used for carrying application state, and integrity of its contents is a concern, clients MUST protect state against tampering and swapping. This can be achieved by binding the contents of state to the browser session and/or signed/encrypted state values See: * https://tools.ietf.org/html/draft-bradley-oauth-jwt-encoded-state-09 * http://www.thread-safe.com/2014/05/the-correct-use-of-state-parameter-in.html * https://stackoverflow.com/questions/52459476/does-oauth-state-mitigate-any-genuinely-dangerous-attacks * https://pipedrive.readme.io/docs/marketplace-oauth-authorization-state-parameter
main
improve state passing form options csv json are currently passed through the oauth state string parameter this complicates passing state and is nonstandard instead we should be storing this state locally with the state token as a key and ideally validating the request if state is used for carrying application state and integrity of its contents is a concern clients must protect state against tampering and swapping this can be achieved by binding the contents of state to the browser session and or signed encrypted state values see
1
226,849
25,008,692,955
IssuesEvent
2022-11-03 13:49:11
temporalio/sdk-go
https://api.github.com/repos/temporalio/sdk-go
closed
golang.org/x/tools-v0.1.10: 3 vulnerabilities (highest severity is: 6.1) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p></summary> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p> <p> <p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (golang.org/x/tools-v0.1.10 version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.1.10 | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | golang.org/x/tools-v0.1.10 | Direct | jQuery - 3.0.0 | &#10060; | | [CVE-2012-6708](https://www.mend.io/vulnerability-database/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | golang.org/x/tools-v0.1.10 | Direct | jQuery - v1.9.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p> <p> Dependency Hierarchy: - :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2015-9251</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p> <p> Dependency Hierarchy: - :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>3.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - 3.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2012-6708</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p> <p> Dependency Hierarchy: - :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-6708>CVE-2012-6708</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>3.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> <p></p> </details>
True
golang.org/x/tools-v0.1.10: 3 vulnerabilities (highest severity is: 6.1) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p></summary> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p> <p> <p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (golang.org/x/tools-v0.1.10 version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.1.10 | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | golang.org/x/tools-v0.1.10 | Direct | jQuery - 3.0.0 | &#10060; | | [CVE-2012-6708](https://www.mend.io/vulnerability-database/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | golang.org/x/tools-v0.1.10 | Direct | jQuery - v1.9.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p> <p> Dependency Hierarchy: - :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2015-9251</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p> <p> Dependency Hierarchy: - :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>3.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - 3.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2012-6708</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p> <p> Dependency Hierarchy: - :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-6708>CVE-2012-6708</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>3.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> <p></p> </details>
non_main
golang org x tools vulnerabilities highest severity is autoclosed vulnerable library golang org x tools library home page a href found in head commit a href vulnerabilities cve severity cvss dependency type fixed in golang org x tools version remediation available medium golang org x tools direct jquery low golang org x tools direct jquery low golang org x tools direct jquery details cve vulnerable library golang org x tools library home page a href dependency hierarchy x golang org x tools vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library golang org x tools library home page a href dependency hierarchy x golang org x tools vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library golang org x tools library home page a href dependency hierarchy x golang org x tools vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery
0
234,916
18,022,591,209
IssuesEvent
2021-09-16 21:36:12
AlmaLinux/build-system
https://api.github.com/repos/AlmaLinux/build-system
opened
Add a link to the Build System project board to README
documentation
We are using a GitHub's kanban board https://github.com/orgs/AlmaLinux/projects/1 for the Build System project. It would be nice to mention that fact in the README file so that people know where progress can be tracked.
1.0
Add a link to the Build System project board to README - We are using a GitHub's kanban board https://github.com/orgs/AlmaLinux/projects/1 for the Build System project. It would be nice to mention that fact in the README file so that people know where progress can be tracked.
non_main
add a link to the build system project board to readme we are using a github s kanban board for the build system project it would be nice to mention that fact in the readme file so that people know where progress can be tracked
0
283,858
21,335,390,265
IssuesEvent
2022-04-18 14:01:08
MetLife/TlsTestGate
https://api.github.com/repos/MetLife/TlsTestGate
closed
Wrong screenshot?
documentation
in https://marketplace.visualstudio.com/items?itemName=gattjoe.TlsTestGate i see: ![image](https://user-images.githubusercontent.com/29877074/163319786-0ddddb52-8258-4a33-bb32-ba8c03baa8e5.png) in [manual](https://github.com/MetLife/TlsTestGate#yaml-sample) i see: `steps: - task: JoeGatt.TlsTestGate.custom-build-release-task.TlsTestGate@1 displayName: 'github.com SSL/TLS Test Gate' inputs: baseURL: github.com port: 443 dnsserver: 8.8.8.8 **failBuild**: true`
1.0
Wrong screenshot? - in https://marketplace.visualstudio.com/items?itemName=gattjoe.TlsTestGate i see: ![image](https://user-images.githubusercontent.com/29877074/163319786-0ddddb52-8258-4a33-bb32-ba8c03baa8e5.png) in [manual](https://github.com/MetLife/TlsTestGate#yaml-sample) i see: `steps: - task: JoeGatt.TlsTestGate.custom-build-release-task.TlsTestGate@1 displayName: 'github.com SSL/TLS Test Gate' inputs: baseURL: github.com port: 443 dnsserver: 8.8.8.8 **failBuild**: true`
non_main
wrong screenshot in i see in i see steps task joegatt tlstestgate custom build release task tlstestgate displayname github com ssl tls test gate inputs baseurl github com port dnsserver failbuild true
0
351,659
10,521,765,214
IssuesEvent
2019-09-30 07:04:52
AY1920S1-CS2103T-T09-2/main
https://api.github.com/repos/AY1920S1-CS2103T-T09-2/main
opened
As a student who wants to get stronger i want to be advised on how much increment I should make for each exercise
priority.Low type.Story
so that i can train progressively
1.0
As a student who wants to get stronger i want to be advised on how much increment I should make for each exercise - so that i can train progressively
non_main
as a student who wants to get stronger i want to be advised on how much increment i should make for each exercise so that i can train progressively
0
5,039
25,841,221,827
IssuesEvent
2022-12-13 00:37:36
ElasticPerch/websocket
https://api.github.com/repos/ElasticPerch/websocket
opened
Add Upgrader.NegotiateSubprotocols
waiting on new maintainer feature request
From websocket created by [garyburd](https://github.com/garyburd): gorilla/websocket#480 See discussion [in this comment](https://github.com/gorilla/websocket/issues/399#issuecomment-413661967).
True
Add Upgrader.NegotiateSubprotocols - From websocket created by [garyburd](https://github.com/garyburd): gorilla/websocket#480 See discussion [in this comment](https://github.com/gorilla/websocket/issues/399#issuecomment-413661967).
main
add upgrader negotiatesubprotocols from websocket created by gorilla websocket see discussion
1
300,993
22,708,219,232
IssuesEvent
2022-07-05 16:29:14
Issafalcon/lsp-overloads.nvim
https://api.github.com/repos/Issafalcon/lsp-overloads.nvim
closed
Plugin documentation
documentation
Generate vim docs and README: - Write README to include: - Summary of the plugin - Gif of signature helper - Setup and configuration - Acknowledgements - Use github action to generate and commit Pandoc conversions of README to vim docs
1.0
Plugin documentation - Generate vim docs and README: - Write README to include: - Summary of the plugin - Gif of signature helper - Setup and configuration - Acknowledgements - Use github action to generate and commit Pandoc conversions of README to vim docs
non_main
plugin documentation generate vim docs and readme write readme to include summary of the plugin gif of signature helper setup and configuration acknowledgements use github action to generate and commit pandoc conversions of readme to vim docs
0
54,458
6,388,330,830
IssuesEvent
2017-08-03 15:22:13
WormBase/website
https://api.github.com/repos/WormBase/website
closed
Broken expression rest endpoint live/staging
Bug Report HelpDesk Under testing Webteam
*Submitted from the feedback form on the WormBase website.* <p>Server Error: 500 at <a href="http://www.wormbase.org/rest/widget/gene/WBGene00000412/expression">rest/widget/gene/WBGene00000412/expression</a></p><p> </p><p>Error Message:<br> Caught exception in WormBase::Web::Controller::REST-&gt;widget "http://rest.wormbase.org/rest/widget/gene/WBGene00000412/expression failed with 500 at /var/lib/jenkins/jobs/staging_build/workspace/lib/WormBase/Web/Controller/REST.pm line 915." </p> **Reported by:** abio******************* **Submitted from:** <a target="_blank" href="http://www.wormbase.org//rest/widget/gene/WBGene00000412/expression">/rest/widget/gene/WBGene00000412/expression</a> **Browser:** Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.115 Safari/537.36
1.0
Broken expression rest endpoint live/staging - *Submitted from the feedback form on the WormBase website.* <p>Server Error: 500 at <a href="http://www.wormbase.org/rest/widget/gene/WBGene00000412/expression">rest/widget/gene/WBGene00000412/expression</a></p><p> </p><p>Error Message:<br> Caught exception in WormBase::Web::Controller::REST-&gt;widget "http://rest.wormbase.org/rest/widget/gene/WBGene00000412/expression failed with 500 at /var/lib/jenkins/jobs/staging_build/workspace/lib/WormBase/Web/Controller/REST.pm line 915." </p> **Reported by:** abio******************* **Submitted from:** <a target="_blank" href="http://www.wormbase.org//rest/widget/gene/WBGene00000412/expression">/rest/widget/gene/WBGene00000412/expression</a> **Browser:** Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.115 Safari/537.36
non_main
broken expression rest endpoint live staging submitted from the feedback form on the wormbase website server error at a href error message caught exception in wormbase web controller rest gt widget failed with at var lib jenkins jobs staging build workspace lib wormbase web controller rest pm line reported by abio submitted from a target blank href browser mozilla windows nt applewebkit khtml like gecko chrome safari
0
95,551
12,004,320,581
IssuesEvent
2020-04-09 11:18:55
Western-Health-Covid19-Collaboration/wh_covid19_app
https://api.github.com/repos/Western-Health-Covid19-Collaboration/wh_covid19_app
closed
Change intubation checklist infographic
#1 Priority Design
The intubation checklist infographic (accessible through tertiary PDF button) requires updating to reflect changes in guidelines form 9th April. Attached is new infographic. @lukesleeman @Western-Health-Covid19-Collaboration/designers This might have to go through design first of all prior to devs? [9thAprSAS COVID 19 Intubaiton Checklist WH version 1 (1).pdf](https://github.com/Western-Health-Covid19-Collaboration/wh_covid19_app/files/4454181/9thAprSAS.COVID.19.Intubaiton.Checklist.WH.version.1.1.pdf)
1.0
Change intubation checklist infographic - The intubation checklist infographic (accessible through tertiary PDF button) requires updating to reflect changes in guidelines form 9th April. Attached is new infographic. @lukesleeman @Western-Health-Covid19-Collaboration/designers This might have to go through design first of all prior to devs? [9thAprSAS COVID 19 Intubaiton Checklist WH version 1 (1).pdf](https://github.com/Western-Health-Covid19-Collaboration/wh_covid19_app/files/4454181/9thAprSAS.COVID.19.Intubaiton.Checklist.WH.version.1.1.pdf)
non_main
change intubation checklist infographic the intubation checklist infographic accessible through tertiary pdf button requires updating to reflect changes in guidelines form april attached is new infographic lukesleeman western health collaboration designers this might have to go through design first of all prior to devs
0
768,426
26,962,837,021
IssuesEvent
2023-02-08 19:37:51
nanovms/nanos
https://api.github.com/repos/nanovms/nanos
closed
tracking tkt for elasticsearch - low priority
jvm low-priority
get a few variations of this ``` warning: overwriting existing file /etc/resolv.conf hostpath old: /home/eyberg/.ops/.staging/java_1.8.0_191/sysroot/etc/resolv.conf new: /home/eyberg/.ops/java_temp/resolv.conf warning: overwriting existing file /proc/sys/kernel/hostname hostpath old: /home/eyberg/.ops/.staging/java_1.8.0_191/sysroot/proc/sys/kernel/hostname new: /home/eyberg/.ops/java_temp/hostname [java -Xms1g -Xmx1g -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=75 -XX:+UseCMSInitiatingOccupancyOnly -Des.networkaddress.cache.ttl=60 -Des.networkaddress.cache.negative.ttl=10 -XX:+AlwaysPreTouch -Xss1m -Djava.awt.headless=true -Dfile.encoding=UTF-8 -Djna.nosys=true -XX:-OmitStackTraceInFastThrow -Dio.netty.noUnsafe=true -Dio.netty.noKeySetOptimization=true -Dio.netty.recycler.maxCapacityPerThread=0 -Dlog4j.shutdownHookEnabled=false -Dlog4j2.disable.jmx=true -Dos.name=Linux -Djava.io.tmpdir=/elasticsearch-9822391376708845111 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=data -XX:ErrorFile=logs/hs_err_pid%p.log -Djava.locale.providers=COMPAT -Dio.netty.allocator.type=unpooled -Des.path.home=/ -Des.path.conf=/home/eyberg/el/elasticsearch-7.0.0/config -Des.distribution.flavor=default -Des.distribution.type=tar -Des.bundled_jdk=true -cp /lib/* org.elasticsearch.bootstrap.Elasticsearch] booting /home/eyberg/.ops/images/java.img ... qemu-system-x86_64: warning: TCG doesn't support requested feature: CPUID.01H:ECX.vmx [bit 5] assigned: 10.0.2.15 OpenJDK 64-Bit Server VM warning: Can't detect initial thread stack location - find_vma failed unix_fault_page error: no vmap found for vaddr 0x1008b18 Unhandled: 000000000000000e Page fault interrupt: 000000000000000e frame: 0000000100e01500 error code: 0000000000000000 address: 0000000001008b18 rax: 0000000000000028 rbx: 0000000001008b10 rcx: 0000000000000000 rdx: 0000000100207400 rsi: 00000070280b3d28 rdi: 00000070280b3d28 rbp: 00000070bf8ff9d0 rsp: 00000070bf8ff9b0 r8: 0000000000000000 r9: 0000000000000000 r10: 0000000000000000 r11: 0000000000000006 r12: 0000000100a3a9c0 r13: 00000070280b3d28 r14: 00000070280b3d28 r15: 0000000000000000 rip: 000000007f00d60a (table_find + 0000000000000074/000000000000009d) flags: 0000000000000002 frame trace: 000000007f01f767 (soft_create_futex + 0000000000000028/0000000000000073) 000000007f01fe81 (futex + 0000000000000077/0000000000000566) 000000007f01b38f (syscall_debug + 000000000000028f/000000000000029c) 000000007f0002b0 (syscall_enter + 0000000000000062/000000000000006c) 00000070a98a8a55 00000070a98a9c78 00000070a9996228 00000070a9ab79de 00000070a9ab7dc7 00000070a98ea992 00000070938076db stack trace: 0000000000000081 000000007f043608 (bootstrap_region + 0000000000000b28/0000000000200000) 00000070280b3d28 0000000100e01700 00000070bf8ffa00 000000007f01f767 (soft_create_futex + 0000000000000028/0000000000000073) 0000000000000081 00000070280b3d28 0000000000000000 0000000000000000 00000070bf8ffa90 000000007f01fe81 (futex + 0000000000000077/0000000000000566) 00000000000000ff 0000007039132169 0000000000000000 00000001fb01beec 0000000000000000 0000000100203400 ffffffffffffffda 0000000100203400 00000070280b3d28 1bd4100efb01beec 00000070bf8ffa90 0000000100203400 halt exit status 255 ``` ``` eyberg@s1:~/el/elasticsearch-7.0.0$ cat config.json { "Args": ["-Xms1g", "-Xmx1g", "-XX:+UseConcMarkSweepGC", "-XX:CMSInitiatingOccupancyFraction=75", "-XX:+UseCMSInitiatingOccupancyOnly", "-Des.networkaddress.cache.ttl=60", "-Des.networkaddress.cache.negative.ttl=10", "-XX:+AlwaysPreTouch", "-Xss1m", "-Djava.awt.headless=true", "-Dfile.encoding=UTF-8", "-Djna.nosys=true", "-XX:-OmitStackTraceInFastThrow", "-Dio.netty.noUnsafe=true", "-Dio.netty.noKeySetOptimization=true", "-Dio.netty.recycler.maxCapacityPerThread=0", "-Dlog4j.shutdownHookEnabled=false", "-Dlog4j2.disable.jmx=true", "-Dos.name=Linux", "-Djava.io.tmpdir=/elasticsearch-9822391376708845111", "-XX:+HeapDumpOnOutOfMemoryError", "-XX:HeapDumpPath=data", "-XX:ErrorFile=logs/hs_err_pid%p.log", "-Djava.locale.providers=COMPAT", "-Dio.netty.allocator.type=unpooled", "-Des.path.home=/", "-Des.path.conf=/home/eyberg/el/elasticsearch-7.0.0/config", "-Des.distribution.flavor=default", "-Des.distribution.type=tar", "-Des.bundled_jdk=true", "-cp", "/lib/*", "org.elasticsearch.bootstrap.Elasticsearch"], "Dirs":["lib", "elasticsearch-9822391376708845111"], "RunConfig": { "Memory": "3G" } } ```
1.0
tracking tkt for elasticsearch - low priority - get a few variations of this ``` warning: overwriting existing file /etc/resolv.conf hostpath old: /home/eyberg/.ops/.staging/java_1.8.0_191/sysroot/etc/resolv.conf new: /home/eyberg/.ops/java_temp/resolv.conf warning: overwriting existing file /proc/sys/kernel/hostname hostpath old: /home/eyberg/.ops/.staging/java_1.8.0_191/sysroot/proc/sys/kernel/hostname new: /home/eyberg/.ops/java_temp/hostname [java -Xms1g -Xmx1g -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=75 -XX:+UseCMSInitiatingOccupancyOnly -Des.networkaddress.cache.ttl=60 -Des.networkaddress.cache.negative.ttl=10 -XX:+AlwaysPreTouch -Xss1m -Djava.awt.headless=true -Dfile.encoding=UTF-8 -Djna.nosys=true -XX:-OmitStackTraceInFastThrow -Dio.netty.noUnsafe=true -Dio.netty.noKeySetOptimization=true -Dio.netty.recycler.maxCapacityPerThread=0 -Dlog4j.shutdownHookEnabled=false -Dlog4j2.disable.jmx=true -Dos.name=Linux -Djava.io.tmpdir=/elasticsearch-9822391376708845111 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=data -XX:ErrorFile=logs/hs_err_pid%p.log -Djava.locale.providers=COMPAT -Dio.netty.allocator.type=unpooled -Des.path.home=/ -Des.path.conf=/home/eyberg/el/elasticsearch-7.0.0/config -Des.distribution.flavor=default -Des.distribution.type=tar -Des.bundled_jdk=true -cp /lib/* org.elasticsearch.bootstrap.Elasticsearch] booting /home/eyberg/.ops/images/java.img ... qemu-system-x86_64: warning: TCG doesn't support requested feature: CPUID.01H:ECX.vmx [bit 5] assigned: 10.0.2.15 OpenJDK 64-Bit Server VM warning: Can't detect initial thread stack location - find_vma failed unix_fault_page error: no vmap found for vaddr 0x1008b18 Unhandled: 000000000000000e Page fault interrupt: 000000000000000e frame: 0000000100e01500 error code: 0000000000000000 address: 0000000001008b18 rax: 0000000000000028 rbx: 0000000001008b10 rcx: 0000000000000000 rdx: 0000000100207400 rsi: 00000070280b3d28 rdi: 00000070280b3d28 rbp: 00000070bf8ff9d0 rsp: 00000070bf8ff9b0 r8: 0000000000000000 r9: 0000000000000000 r10: 0000000000000000 r11: 0000000000000006 r12: 0000000100a3a9c0 r13: 00000070280b3d28 r14: 00000070280b3d28 r15: 0000000000000000 rip: 000000007f00d60a (table_find + 0000000000000074/000000000000009d) flags: 0000000000000002 frame trace: 000000007f01f767 (soft_create_futex + 0000000000000028/0000000000000073) 000000007f01fe81 (futex + 0000000000000077/0000000000000566) 000000007f01b38f (syscall_debug + 000000000000028f/000000000000029c) 000000007f0002b0 (syscall_enter + 0000000000000062/000000000000006c) 00000070a98a8a55 00000070a98a9c78 00000070a9996228 00000070a9ab79de 00000070a9ab7dc7 00000070a98ea992 00000070938076db stack trace: 0000000000000081 000000007f043608 (bootstrap_region + 0000000000000b28/0000000000200000) 00000070280b3d28 0000000100e01700 00000070bf8ffa00 000000007f01f767 (soft_create_futex + 0000000000000028/0000000000000073) 0000000000000081 00000070280b3d28 0000000000000000 0000000000000000 00000070bf8ffa90 000000007f01fe81 (futex + 0000000000000077/0000000000000566) 00000000000000ff 0000007039132169 0000000000000000 00000001fb01beec 0000000000000000 0000000100203400 ffffffffffffffda 0000000100203400 00000070280b3d28 1bd4100efb01beec 00000070bf8ffa90 0000000100203400 halt exit status 255 ``` ``` eyberg@s1:~/el/elasticsearch-7.0.0$ cat config.json { "Args": ["-Xms1g", "-Xmx1g", "-XX:+UseConcMarkSweepGC", "-XX:CMSInitiatingOccupancyFraction=75", "-XX:+UseCMSInitiatingOccupancyOnly", "-Des.networkaddress.cache.ttl=60", "-Des.networkaddress.cache.negative.ttl=10", "-XX:+AlwaysPreTouch", "-Xss1m", "-Djava.awt.headless=true", "-Dfile.encoding=UTF-8", "-Djna.nosys=true", "-XX:-OmitStackTraceInFastThrow", "-Dio.netty.noUnsafe=true", "-Dio.netty.noKeySetOptimization=true", "-Dio.netty.recycler.maxCapacityPerThread=0", "-Dlog4j.shutdownHookEnabled=false", "-Dlog4j2.disable.jmx=true", "-Dos.name=Linux", "-Djava.io.tmpdir=/elasticsearch-9822391376708845111", "-XX:+HeapDumpOnOutOfMemoryError", "-XX:HeapDumpPath=data", "-XX:ErrorFile=logs/hs_err_pid%p.log", "-Djava.locale.providers=COMPAT", "-Dio.netty.allocator.type=unpooled", "-Des.path.home=/", "-Des.path.conf=/home/eyberg/el/elasticsearch-7.0.0/config", "-Des.distribution.flavor=default", "-Des.distribution.type=tar", "-Des.bundled_jdk=true", "-cp", "/lib/*", "org.elasticsearch.bootstrap.Elasticsearch"], "Dirs":["lib", "elasticsearch-9822391376708845111"], "RunConfig": { "Memory": "3G" } } ```
non_main
tracking tkt for elasticsearch low priority get a few variations of this warning overwriting existing file etc resolv conf hostpath old home eyberg ops staging java sysroot etc resolv conf new home eyberg ops java temp resolv conf warning overwriting existing file proc sys kernel hostname hostpath old home eyberg ops staging java sysroot proc sys kernel hostname new home eyberg ops java temp hostname booting home eyberg ops images java img qemu system warning tcg doesn t support requested feature cpuid ecx vmx assigned openjdk bit server vm warning can t detect initial thread stack location find vma failed unix fault page error no vmap found for vaddr unhandled page fault interrupt frame error code address rax rbx rcx rdx rsi rdi rbp rsp rip table find flags frame trace soft create futex futex syscall debug syscall enter stack trace bootstrap region soft create futex futex ffffffffffffffda halt exit status eyberg el elasticsearch cat config json args xx useconcmarksweepgc xx cmsinitiatingoccupancyfraction xx usecmsinitiatingoccupancyonly des networkaddress cache ttl des networkaddress cache negative ttl xx alwayspretouch djava awt headless true dfile encoding utf djna nosys true xx omitstacktraceinfastthrow dio netty nounsafe true dio netty nokeysetoptimization true dio netty recycler maxcapacityperthread shutdownhookenabled false disable jmx true dos name linux djava io tmpdir elasticsearch xx heapdumponoutofmemoryerror xx heapdumppath data xx errorfile logs hs err pid p log djava locale providers compat dio netty allocator type unpooled des path home des path conf home eyberg el elasticsearch config des distribution flavor default des distribution type tar des bundled jdk true cp lib org elasticsearch bootstrap elasticsearch dirs runconfig memory
0
1,638
6,572,661,924
IssuesEvent
2017-09-11 04:11:13
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
firewalld: Allow Service definition and Source IP/Block specification
affects_2.2 feature_idea waiting_on_maintainer
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME firewalld ##### ANSIBLE VERSION ``` ansible 2.2.0.0 ``` ##### OS / ENVIRONMENT Redhat 7+ ##### SUMMARY The firewalld module should be improved to allow more basic functions. Right now, you can allow everything from a host, everything to a port, everything to a previously defined service, or a rich rule, which includes everything else. There should be a way to add a new service and then allow connections to that service from an IP or IP block. Match rules, tps rules, and the like, should still be within rich rules, but defining services in firewalld and slightly more granular control of source IPs should really be within the ability of this module.
True
firewalld: Allow Service definition and Source IP/Block specification - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME firewalld ##### ANSIBLE VERSION ``` ansible 2.2.0.0 ``` ##### OS / ENVIRONMENT Redhat 7+ ##### SUMMARY The firewalld module should be improved to allow more basic functions. Right now, you can allow everything from a host, everything to a port, everything to a previously defined service, or a rich rule, which includes everything else. There should be a way to add a new service and then allow connections to that service from an IP or IP block. Match rules, tps rules, and the like, should still be within rich rules, but defining services in firewalld and slightly more granular control of source IPs should really be within the ability of this module.
main
firewalld allow service definition and source ip block specification issue type feature idea component name firewalld ansible version ansible os environment redhat summary the firewalld module should be improved to allow more basic functions right now you can allow everything from a host everything to a port everything to a previously defined service or a rich rule which includes everything else there should be a way to add a new service and then allow connections to that service from an ip or ip block match rules tps rules and the like should still be within rich rules but defining services in firewalld and slightly more granular control of source ips should really be within the ability of this module
1
12,601
3,631,332,747
IssuesEvent
2016-02-11 00:49:17
systemd/systemd
https://api.github.com/repos/systemd/systemd
closed
Manpage for systemctl is incorrect, full service name isn't needed for enable/disable
documentation systemctl
The manpage (http://www.freedesktop.org/software/systemd/man/systemctl.html) says: >For unit file commands, the specified NAME should be the full name of the unit file, or the absolute path to the unit file: > > systemctl enable foo.service >or > systemctl link /path/to/foo.service but you can actually use systemctl enable foo without any problems.
1.0
Manpage for systemctl is incorrect, full service name isn't needed for enable/disable - The manpage (http://www.freedesktop.org/software/systemd/man/systemctl.html) says: >For unit file commands, the specified NAME should be the full name of the unit file, or the absolute path to the unit file: > > systemctl enable foo.service >or > systemctl link /path/to/foo.service but you can actually use systemctl enable foo without any problems.
non_main
manpage for systemctl is incorrect full service name isn t needed for enable disable the manpage says for unit file commands the specified name should be the full name of the unit file or the absolute path to the unit file systemctl enable foo service or systemctl link path to foo service but you can actually use systemctl enable foo without any problems
0
1,889
6,577,532,919
IssuesEvent
2017-09-12 01:34:29
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
iam_policy module errors out when using policy_json field
affects_2.0 aws bug_report cloud waiting_on_maintainer
##### Issue Type: - Bug Report ##### Plugin Name: iam_policy ##### Ansible Version: 2.0.1.0 ##### Ansible Configuration: n/a ##### Environment: n/a ##### Summary: iam_policy module errors out if you use the policy_json field. PR #2730 introduced this bug. ##### Steps To Reproduce: Example: ``` - iam_policy: iam_type: role iam_name: my-role-name policy_name: my-policy-name policy_json: "{{ lookup( 'template', 'policy.json.j2', convert_data=False) }}" state: present ``` ##### Expected Results: Expected to succeed. ##### Actual Results: ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/root/.ansible/tmp/ansible-tmp-1457652747.27-251976645440330/iam_policy", line 2528, in <module> main() File "/root/.ansible/tmp/ansible-tmp-1457652747.27-251976645440330/iam_policy", line 306, in main if not isinstance(pdoc, basestring): UnboundLocalError: local variable 'pdoc' referenced before assignment fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "iam_policy"}, "parsed": false} ```
True
iam_policy module errors out when using policy_json field - ##### Issue Type: - Bug Report ##### Plugin Name: iam_policy ##### Ansible Version: 2.0.1.0 ##### Ansible Configuration: n/a ##### Environment: n/a ##### Summary: iam_policy module errors out if you use the policy_json field. PR #2730 introduced this bug. ##### Steps To Reproduce: Example: ``` - iam_policy: iam_type: role iam_name: my-role-name policy_name: my-policy-name policy_json: "{{ lookup( 'template', 'policy.json.j2', convert_data=False) }}" state: present ``` ##### Expected Results: Expected to succeed. ##### Actual Results: ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/root/.ansible/tmp/ansible-tmp-1457652747.27-251976645440330/iam_policy", line 2528, in <module> main() File "/root/.ansible/tmp/ansible-tmp-1457652747.27-251976645440330/iam_policy", line 306, in main if not isinstance(pdoc, basestring): UnboundLocalError: local variable 'pdoc' referenced before assignment fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "iam_policy"}, "parsed": false} ```
main
iam policy module errors out when using policy json field issue type bug report plugin name iam policy ansible version ansible configuration n a environment n a summary iam policy module errors out if you use the policy json field pr introduced this bug steps to reproduce example iam policy iam type role iam name my role name policy name my policy name policy json lookup template policy json convert data false state present expected results expected to succeed actual results an exception occurred during task execution the full traceback is traceback most recent call last file root ansible tmp ansible tmp iam policy line in main file root ansible tmp ansible tmp iam policy line in main if not isinstance pdoc basestring unboundlocalerror local variable pdoc referenced before assignment fatal failed changed false failed true invocation module name iam policy parsed false
1
1,422
3,738,859,406
IssuesEvent
2016-03-09 00:53:53
Royedc4/caYca
https://api.github.com/repos/Royedc4/caYca
closed
Presentation 4 korea
requirement
Required by Alba Isabel. New specific presentation of caYca to Korea people.
1.0
Presentation 4 korea - Required by Alba Isabel. New specific presentation of caYca to Korea people.
non_main
presentation korea required by alba isabel new specific presentation of cayca to korea people
0
21,915
11,424,548,328
IssuesEvent
2020-02-03 17:59:09
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Keras RNN training speed significantly slower with eager execution/control flow v2
TF 2.0 comp:keras type:performance
**System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): 2.0.0 - Python version: 3.6.8 - Bazel version (if compiling from source): N/A - GCC/Compiler version (if compiling from source): N/A - CUDA/cuDNN version: 10.0/7.6 - GPU model and memory: GTX 980 Ti **Describe the current behavior** Enabling eager execution or control flow v2 causes RNN training speed to decrease significantly. **Describe the expected behavior** Enabling eager mode or control flow v2 should not affect the training time (or improve it, ideally). **Code to reproduce the issue** ``` python import tensorflow as tf import numpy as np import timeit use_eager = False use_v2 = False if not use_eager: tf.compat.v1.disable_eager_execution() if not use_v2: tf.compat.v1.disable_control_flow_v2() n_steps = 1000 n_input = 100 n_hidden = 1000 batch_size = 64 inputs = tf.keras.Input((n_steps, n_input)) outputs = tf.keras.layers.SimpleRNN(units=n_hidden, return_sequences=True)(inputs) outputs = tf.keras.layers.Dense(units=n_input)(outputs) model = tf.keras.Model(inputs=inputs, outputs=outputs) model.compile(optimizer=tf.optimizers.SGD(0.1), loss="mse") x = np.ones((batch_size, n_steps, n_input)) y = np.ones((batch_size, n_steps, n_input)) # warmup model.fit(x, y, epochs=1) start = timeit.default_timer() model.fit(x, y, epochs=10) print("Execution time:", timeit.default_timer() - start) ``` **Other info / logs** On my machine the results look like: - use_eager=False, use_v2=False: 5.90s - use_eager=False, use_v2=True: 8.08s - use_eager=True, use_v2=False: 9.81s - use_eager=True, use_v2=True: 10.10s So, overall a >60% increase in training time comparing no eager and no v2 to the current defaults.
True
Keras RNN training speed significantly slower with eager execution/control flow v2 - **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): 2.0.0 - Python version: 3.6.8 - Bazel version (if compiling from source): N/A - GCC/Compiler version (if compiling from source): N/A - CUDA/cuDNN version: 10.0/7.6 - GPU model and memory: GTX 980 Ti **Describe the current behavior** Enabling eager execution or control flow v2 causes RNN training speed to decrease significantly. **Describe the expected behavior** Enabling eager mode or control flow v2 should not affect the training time (or improve it, ideally). **Code to reproduce the issue** ``` python import tensorflow as tf import numpy as np import timeit use_eager = False use_v2 = False if not use_eager: tf.compat.v1.disable_eager_execution() if not use_v2: tf.compat.v1.disable_control_flow_v2() n_steps = 1000 n_input = 100 n_hidden = 1000 batch_size = 64 inputs = tf.keras.Input((n_steps, n_input)) outputs = tf.keras.layers.SimpleRNN(units=n_hidden, return_sequences=True)(inputs) outputs = tf.keras.layers.Dense(units=n_input)(outputs) model = tf.keras.Model(inputs=inputs, outputs=outputs) model.compile(optimizer=tf.optimizers.SGD(0.1), loss="mse") x = np.ones((batch_size, n_steps, n_input)) y = np.ones((batch_size, n_steps, n_input)) # warmup model.fit(x, y, epochs=1) start = timeit.default_timer() model.fit(x, y, epochs=10) print("Execution time:", timeit.default_timer() - start) ``` **Other info / logs** On my machine the results look like: - use_eager=False, use_v2=False: 5.90s - use_eager=False, use_v2=True: 8.08s - use_eager=True, use_v2=False: 9.81s - use_eager=True, use_v2=True: 10.10s So, overall a >60% increase in training time comparing no eager and no v2 to the current defaults.
non_main
keras rnn training speed significantly slower with eager execution control flow system information have i written custom code as opposed to using a stock example script provided in tensorflow yes os platform and distribution e g linux ubuntu windows mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device n a tensorflow installed from source or binary source tensorflow version use command below python version bazel version if compiling from source n a gcc compiler version if compiling from source n a cuda cudnn version gpu model and memory gtx ti describe the current behavior enabling eager execution or control flow causes rnn training speed to decrease significantly describe the expected behavior enabling eager mode or control flow should not affect the training time or improve it ideally code to reproduce the issue python import tensorflow as tf import numpy as np import timeit use eager false use false if not use eager tf compat disable eager execution if not use tf compat disable control flow n steps n input n hidden batch size inputs tf keras input n steps n input outputs tf keras layers simplernn units n hidden return sequences true inputs outputs tf keras layers dense units n input outputs model tf keras model inputs inputs outputs outputs model compile optimizer tf optimizers sgd loss mse x np ones batch size n steps n input y np ones batch size n steps n input warmup model fit x y epochs start timeit default timer model fit x y epochs print execution time timeit default timer start other info logs on my machine the results look like use eager false use false use eager false use true use eager true use false use eager true use true so overall a increase in training time comparing no eager and no to the current defaults
0
14,985
9,635,603,485
IssuesEvent
2019-05-16 01:58:25
PuzzleServer/mainpuzzleserver
https://api.github.com/repos/PuzzleServer/mainpuzzleserver
closed
[P0] Add team blurbs
enhancement usability
Allow teams to fill in a blurb that will be shown during matchmaking to help them find players that are good fits for their teams
True
[P0] Add team blurbs - Allow teams to fill in a blurb that will be shown during matchmaking to help them find players that are good fits for their teams
non_main
add team blurbs allow teams to fill in a blurb that will be shown during matchmaking to help them find players that are good fits for their teams
0
557,441
16,509,691,989
IssuesEvent
2021-05-26 01:22:28
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
appclient -user xxx option is ignored unless -passwordfile is given
Component: security ERR: Assignee Priority: Major Stale Type: Bug
Steps to reproduce: * appclient -name myname -client MyClient.jar Expected result: * Login dialog should default user name to "myname". Actual result: * Login dialog defaults user name to Windows Account. #### Environment GlassFish 3.1.1, Win 7 Pro SP 1 (64 Bit), JDK 1.7.0_21
1.0
appclient -user xxx option is ignored unless -passwordfile is given - Steps to reproduce: * appclient -name myname -client MyClient.jar Expected result: * Login dialog should default user name to "myname". Actual result: * Login dialog defaults user name to Windows Account. #### Environment GlassFish 3.1.1, Win 7 Pro SP 1 (64 Bit), JDK 1.7.0_21
non_main
appclient user xxx option is ignored unless passwordfile is given steps to reproduce appclient name myname client myclient jar expected result login dialog should default user name to myname actual result login dialog defaults user name to windows account environment glassfish win pro sp bit jdk
0
1,305
5,545,189,143
IssuesEvent
2017-03-22 20:58:11
WhitestormJS/whitestorm.js
https://api.github.com/repos/WhitestormJS/whitestorm.js
opened
Spec/tests closer to src code
MAINTAINANCE
Ideally the tests should be closer to the src code. Why? Incentive to test the code. Not having to search for where existing specs are etc. If not possible/difficult with our current setup, let's at least bring structure to the specs we have in the `test` folder. Or discuss. what we have right now is far from ideal. ###### Version: - [x] v2.x.x - [ ] v1.x.x ###### Issue type: - [ ] Bug - [ ] Proposal/Enhancement - [ ] Question - [ ] Discussion ------ <details> <summary> <b>Tested on: </b> </summary> ###### Desktop - [ ] Chrome - [ ] Chrome Canary - [ ] Chrome dev-channel - [ ] Firefox - [ ] Opera - [ ] Microsoft IE - [ ] Microsoft Edge ###### Android - [ ] Chrome - [ ] Firefox - [ ] Opera ###### IOS - [ ] Chrome - [ ] Firefox - [ ] Opera </details>
True
Spec/tests closer to src code - Ideally the tests should be closer to the src code. Why? Incentive to test the code. Not having to search for where existing specs are etc. If not possible/difficult with our current setup, let's at least bring structure to the specs we have in the `test` folder. Or discuss. what we have right now is far from ideal. ###### Version: - [x] v2.x.x - [ ] v1.x.x ###### Issue type: - [ ] Bug - [ ] Proposal/Enhancement - [ ] Question - [ ] Discussion ------ <details> <summary> <b>Tested on: </b> </summary> ###### Desktop - [ ] Chrome - [ ] Chrome Canary - [ ] Chrome dev-channel - [ ] Firefox - [ ] Opera - [ ] Microsoft IE - [ ] Microsoft Edge ###### Android - [ ] Chrome - [ ] Firefox - [ ] Opera ###### IOS - [ ] Chrome - [ ] Firefox - [ ] Opera </details>
main
spec tests closer to src code ideally the tests should be closer to the src code why incentive to test the code not having to search for where existing specs are etc if not possible difficult with our current setup let s at least bring structure to the specs we have in the test folder or discuss what we have right now is far from ideal version x x x x issue type bug proposal enhancement question discussion tested on desktop chrome chrome canary chrome dev channel firefox opera microsoft ie microsoft edge android chrome firefox opera ios chrome firefox opera
1
3,116
11,904,991,493
IssuesEvent
2020-03-30 17:47:30
diofant/diofant
https://api.github.com/repos/diofant/diofant
opened
Port remaining polys/dense*.py methods to sparse polys
maintainability polys
Ex., implement PolyElement.decompose(), per [this TODO](https://github.com/diofant/diofant/blob/3d08f9ab8cd77359f97411382ad754b5dc09b96e/diofant/polys/rings.py#L2058-L2059).
True
Port remaining polys/dense*.py methods to sparse polys - Ex., implement PolyElement.decompose(), per [this TODO](https://github.com/diofant/diofant/blob/3d08f9ab8cd77359f97411382ad754b5dc09b96e/diofant/polys/rings.py#L2058-L2059).
main
port remaining polys dense py methods to sparse polys ex implement polyelement decompose per
1
3,581
14,367,640,018
IssuesEvent
2020-12-01 07:03:55
adda-team/adda
https://api.github.com/repos/adda-team/adda
closed
Upgrade to clFFT and clBLAS
OpenCL comp-Library enhancement maintainability pri-High
``` AMD has open-sourced its projects clAmdFft and clAmdBlas some time ago. Now they are called clFFT and clBLAS respectively. And these projects have new binary releases: https://github.com/clMathLibraries/clFFT/releases https://github.com/clMathLibraries/clBLAS/releases The changelog is probably not that impressive yet, but it make sense to upgrade to it before the next release. Then the user will be able to benefit from further development of clFFT and clBLAS simply by downloading new DLLs from this projects. The upgrade should be a matter of removing "Amd" from all names of headers, libraries, and functions. But also wiki-s with installation instructions for these packages should be updated. ``` Original issue reported on code.google.com by `yurkin` on 3 Aug 2014 at 4:49 - Blocking: #157
True
Upgrade to clFFT and clBLAS - ``` AMD has open-sourced its projects clAmdFft and clAmdBlas some time ago. Now they are called clFFT and clBLAS respectively. And these projects have new binary releases: https://github.com/clMathLibraries/clFFT/releases https://github.com/clMathLibraries/clBLAS/releases The changelog is probably not that impressive yet, but it make sense to upgrade to it before the next release. Then the user will be able to benefit from further development of clFFT and clBLAS simply by downloading new DLLs from this projects. The upgrade should be a matter of removing "Amd" from all names of headers, libraries, and functions. But also wiki-s with installation instructions for these packages should be updated. ``` Original issue reported on code.google.com by `yurkin` on 3 Aug 2014 at 4:49 - Blocking: #157
main
upgrade to clfft and clblas amd has open sourced its projects clamdfft and clamdblas some time ago now they are called clfft and clblas respectively and these projects have new binary releases the changelog is probably not that impressive yet but it make sense to upgrade to it before the next release then the user will be able to benefit from further development of clfft and clblas simply by downloading new dlls from this projects the upgrade should be a matter of removing amd from all names of headers libraries and functions but also wiki s with installation instructions for these packages should be updated original issue reported on code google com by yurkin on aug at blocking
1
60,289
8,414,787,861
IssuesEvent
2018-10-13 07:08:41
deeplearning4j/deeplearning4j
https://api.github.com/repos/deeplearning4j/deeplearning4j
closed
ND4J: Add instructions for AVX artifacts
Documentation ND4J
The instructions to use AVX binaries on the following page isn't available in the new docs: https://github.com/deeplearning4j/nd4j/blob/gh-pages/getstarted.md
1.0
ND4J: Add instructions for AVX artifacts - The instructions to use AVX binaries on the following page isn't available in the new docs: https://github.com/deeplearning4j/nd4j/blob/gh-pages/getstarted.md
non_main
add instructions for avx artifacts the instructions to use avx binaries on the following page isn t available in the new docs
0
104,009
11,387,474,641
IssuesEvent
2020-01-29 15:05:06
vtex-apps/io-documentation
https://api.github.com/repos/vtex-apps/io-documentation
closed
vtex-apps/helloworld has no documentation yet
no-documentation
[vtex-apps/helloworld](https://github.com/vtex-apps/helloworld) hasn't created any README file yet or is not using Docs Builder
1.0
vtex-apps/helloworld has no documentation yet - [vtex-apps/helloworld](https://github.com/vtex-apps/helloworld) hasn't created any README file yet or is not using Docs Builder
non_main
vtex apps helloworld has no documentation yet hasn t created any readme file yet or is not using docs builder
0
815,533
30,560,275,691
IssuesEvent
2023-07-20 14:13:09
intersystems-community/vscode-objectscript
https://api.github.com/repos/intersystems-community/vscode-objectscript
closed
revisit support export/import xml
enhancement upstream priority/important-soon
Hi All, with the deprecation of InterSystems Studio, can we look into getting export/import of xml packaged source code implemented? As a developer i often have the issue that i need to transport code from e.g. a non CCR controlled scratch environment to the BASE environment which then is under CCR source control. This usually is done via ISC Studio export/import functionality. ISC Studio import functionality also triggers the question on source controlled environment to add all imported code to source control (here CCR). Using UDL as the export format can be prohibitive if the code consists of multiple classes/routines in multiple packages. (e.g. InterSystems Integration projects) Also the current export functionality is rather clunky in large namespaces (i.e. thousands of classes) and also i found a bit hit and miss in regards to it working (got it working 2 out of 10 tries, usually not working when i try to present the functionality to others)
1.0
revisit support export/import xml - Hi All, with the deprecation of InterSystems Studio, can we look into getting export/import of xml packaged source code implemented? As a developer i often have the issue that i need to transport code from e.g. a non CCR controlled scratch environment to the BASE environment which then is under CCR source control. This usually is done via ISC Studio export/import functionality. ISC Studio import functionality also triggers the question on source controlled environment to add all imported code to source control (here CCR). Using UDL as the export format can be prohibitive if the code consists of multiple classes/routines in multiple packages. (e.g. InterSystems Integration projects) Also the current export functionality is rather clunky in large namespaces (i.e. thousands of classes) and also i found a bit hit and miss in regards to it working (got it working 2 out of 10 tries, usually not working when i try to present the functionality to others)
non_main
revisit support export import xml hi all with the deprecation of intersystems studio can we look into getting export import of xml packaged source code implemented as a developer i often have the issue that i need to transport code from e g a non ccr controlled scratch environment to the base environment which then is under ccr source control this usually is done via isc studio export import functionality isc studio import functionality also triggers the question on source controlled environment to add all imported code to source control here ccr using udl as the export format can be prohibitive if the code consists of multiple classes routines in multiple packages e g intersystems integration projects also the current export functionality is rather clunky in large namespaces i e thousands of classes and also i found a bit hit and miss in regards to it working got it working out of tries usually not working when i try to present the functionality to others
0
1,262
5,353,855,504
IssuesEvent
2017-02-20 07:54:14
espeak-ng/espeak-ng
https://api.github.com/repos/espeak-ng/espeak-ng
closed
Merge the android branch into master.
maintainability portability resolved/fixed
Now that espeak-ng has diverged from espeak, it makes sense to have the android branch merged into the main development line. This will make it easier to maintain the Android support in the future and keep it up-to-date. - [x] Merge the android code into the master branch. - [x] Fix building the JNI and libespeak-ng code for Android. - [x] Incorporate the Android README text into the README file. - [x] Add ucd-tools as a git subtree. - [x] Use ucd-tools to create a compatibility shim for the wide-character APIs.
True
Merge the android branch into master. - Now that espeak-ng has diverged from espeak, it makes sense to have the android branch merged into the main development line. This will make it easier to maintain the Android support in the future and keep it up-to-date. - [x] Merge the android code into the master branch. - [x] Fix building the JNI and libespeak-ng code for Android. - [x] Incorporate the Android README text into the README file. - [x] Add ucd-tools as a git subtree. - [x] Use ucd-tools to create a compatibility shim for the wide-character APIs.
main
merge the android branch into master now that espeak ng has diverged from espeak it makes sense to have the android branch merged into the main development line this will make it easier to maintain the android support in the future and keep it up to date merge the android code into the master branch fix building the jni and libespeak ng code for android incorporate the android readme text into the readme file add ucd tools as a git subtree use ucd tools to create a compatibility shim for the wide character apis
1
7,797
8,056,695,841
IssuesEvent
2018-08-02 13:29:07
spring-io/initializr
https://api.github.com/repos/spring-io/initializr
closed
Remove JTA entries
type: service-upgrade
We currently have three entries for JTA and the traffic on them is quite low. While supporting JTA is important for Spring Boot, I am not entirely sure we should promote its usage for new projects. In our effort to focus our proposition, I am wondering if we shouldn't remove them.
1.0
Remove JTA entries - We currently have three entries for JTA and the traffic on them is quite low. While supporting JTA is important for Spring Boot, I am not entirely sure we should promote its usage for new projects. In our effort to focus our proposition, I am wondering if we shouldn't remove them.
non_main
remove jta entries we currently have three entries for jta and the traffic on them is quite low while supporting jta is important for spring boot i am not entirely sure we should promote its usage for new projects in our effort to focus our proposition i am wondering if we shouldn t remove them
0
975
4,726,024,203
IssuesEvent
2016-10-18 08:54:17
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_owner fails with recurse option specified
affects_2.1 bug_report waiting_on_maintainer windows
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME win_owner ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 ``` ##### CONFIGURATION <!--- --> N/A ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Windows ##### SUMMARY <!--- Explain the problem briefly --> Ansible converts yes/true/no/false to booleans before passing them to win_owner but win_owner's ``recurse`` parameter seems is expecting yes/no string instead of a boolean. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - win_owner: path: "{{ app_dir }}" user: user@domain recurse: yes ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Set ownership of specified directory recursively without errors. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` fatal: [hostname]: FAILED! => {"changed": false, "failed": true, "msg": "Argument recurse needs to be one of no,yes but was True."} ```
True
win_owner fails with recurse option specified - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME win_owner ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 ``` ##### CONFIGURATION <!--- --> N/A ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Windows ##### SUMMARY <!--- Explain the problem briefly --> Ansible converts yes/true/no/false to booleans before passing them to win_owner but win_owner's ``recurse`` parameter seems is expecting yes/no string instead of a boolean. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - win_owner: path: "{{ app_dir }}" user: user@domain recurse: yes ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Set ownership of specified directory recursively without errors. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` fatal: [hostname]: FAILED! => {"changed": false, "failed": true, "msg": "Argument recurse needs to be one of no,yes but was True."} ```
main
win owner fails with recurse option specified issue type bug report component name win owner ansible version ansible configuration n a os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific windows summary ansible converts yes true no false to booleans before passing them to win owner but win owner s recurse parameter seems is expecting yes no string instead of a boolean steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used win owner path app dir user user domain recurse yes expected results set ownership of specified directory recursively without errors actual results fatal failed changed false failed true msg argument recurse needs to be one of no yes but was true
1
515
3,883,049,194
IssuesEvent
2016-04-13 12:31:37
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Quack
Improvement Maintainer Input Requested
The "Need a Slack Invitation?" message needs updating after the Slackbot is deployed. ------ IA Page: http://duck.co/ia/view/quackhack_cheat_sheet [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @zekiel
True
Quack - The "Need a Slack Invitation?" message needs updating after the Slackbot is deployed. ------ IA Page: http://duck.co/ia/view/quackhack_cheat_sheet [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @zekiel
main
quack the need a slack invitation message needs updating after the slackbot is deployed ia page zekiel
1
2,170
7,603,104,890
IssuesEvent
2018-04-29 10:50:00
lansuite/lansuite
https://api.github.com/repos/lansuite/lansuite
opened
Scheduling a release / TODOs for the next release
pending-maintainer-response question
## Context The last release of LanSuite was v4.2 from Mar 15, 2015. Since then we did a couple of changes here and there. What changes do we want/need to do before we ship a new release? - [ ] Add a `CHANGELOG.md` with the listed changes - [ ] Fix #72 - [ ] Fix #9 - [ ] Fix #89 - [ ] Fix #145 - [ ] Fix #185 - [ ] Fix #222 What else comes into your mind @M4LuZ ? What about #183 ? I know that we talked about to test the new code at a party before we release it. The thing is: Not so many people run this code at a party on master version (maybe you and me, but that's it). And if we wait and keep changing, we pile up changes and changes and many people are "afraid" to upgrade. So the coin has two sides. What do you think?
True
Scheduling a release / TODOs for the next release - ## Context The last release of LanSuite was v4.2 from Mar 15, 2015. Since then we did a couple of changes here and there. What changes do we want/need to do before we ship a new release? - [ ] Add a `CHANGELOG.md` with the listed changes - [ ] Fix #72 - [ ] Fix #9 - [ ] Fix #89 - [ ] Fix #145 - [ ] Fix #185 - [ ] Fix #222 What else comes into your mind @M4LuZ ? What about #183 ? I know that we talked about to test the new code at a party before we release it. The thing is: Not so many people run this code at a party on master version (maybe you and me, but that's it). And if we wait and keep changing, we pile up changes and changes and many people are "afraid" to upgrade. So the coin has two sides. What do you think?
main
scheduling a release todos for the next release context the last release of lansuite was from mar since then we did a couple of changes here and there what changes do we want need to do before we ship a new release add a changelog md with the listed changes fix fix fix fix fix fix what else comes into your mind what about i know that we talked about to test the new code at a party before we release it the thing is not so many people run this code at a party on master version maybe you and me but that s it and if we wait and keep changing we pile up changes and changes and many people are afraid to upgrade so the coin has two sides what do you think
1
2,714
9,544,095,952
IssuesEvent
2019-05-01 13:06:38
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
closed
EventArgs should not implement interfaces
Area: analyzer Area: maintainability backlog feature
`EventArgs` are by themselves contracts. Therefore, they should not implement any additional interface.
True
EventArgs should not implement interfaces - `EventArgs` are by themselves contracts. Therefore, they should not implement any additional interface.
main
eventargs should not implement interfaces eventargs are by themselves contracts therefore they should not implement any additional interface
1
4,438
23,057,650,776
IssuesEvent
2022-07-25 06:56:32
samkatakouzinosracing/nspace.hu-soundshift
https://api.github.com/repos/samkatakouzinosracing/nspace.hu-soundshift
closed
Edit Global Mazda MX-5 Cup shift point to 7200 rpm
vehicle maintain
It's set at 6400 rpm now, which is hit almost immediately after changing up a gear. I have set it at 7200 rpm now, which is also the exact moment the red light comes up in game and gives you just enough time to change gears before hitting the limiter. In the MX-5, afaik, it's important to stay in a gear as long as possible to get the best acceleration. So, can you please update it to 7200 rpm?
True
Edit Global Mazda MX-5 Cup shift point to 7200 rpm - It's set at 6400 rpm now, which is hit almost immediately after changing up a gear. I have set it at 7200 rpm now, which is also the exact moment the red light comes up in game and gives you just enough time to change gears before hitting the limiter. In the MX-5, afaik, it's important to stay in a gear as long as possible to get the best acceleration. So, can you please update it to 7200 rpm?
main
edit global mazda mx cup shift point to rpm it s set at rpm now which is hit almost immediately after changing up a gear i have set it at rpm now which is also the exact moment the red light comes up in game and gives you just enough time to change gears before hitting the limiter in the mx afaik it s important to stay in a gear as long as possible to get the best acceleration so can you please update it to rpm
1
35,237
16,992,924,901
IssuesEvent
2021-07-01 00:02:53
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
kvserver: limit MVCC scan in Raft Entries retrieval
A-kv-replication C-bug C-performance N-followup O-postmortem T-kv
In a recent customer escalation, we saw a newly elected Raft leader spend so long to send 100 log entries to catch up a follower that it missed its heartbeat. The log entries were found to be about 132 kb each, and we limit these messages to be a maximum of 32 kb (but at least 1 entry), so responses were being sent with a single entry each. It seemed to be taking 60 ms to retrieve a single entry, totalling 6s for 100 log entries. The log was also huge, at ~900,000 entries, with the current commit index at 1500. Digging into the Raft code, we find that `raft.maybeSendAppend()` calls int `raftLog.entries()` to fetch entries to send to the follower. Notice how it limits this solely by `maxsize` (32 kb), otherwise using the entire Raft log (`lastIndex()+1`) as an upper bound: https://github.com/cockroachdb/vendored/blob/310ea3cd09b55055b290f0a4cf9991a91725797e/go.etcd.io/etcd/raft/v3/log.go#L287-L292 This in turn gets passed into `Storage.Entries()`: https://github.com/cockroachdb/vendored/blob/310ea3cd09b55055b290f0a4cf9991a91725797e/go.etcd.io/etcd/raft/v3/log.go#L350 Our implementation of `Entries()`, however, does not pass the max size bound to the MVCC iterator, only the key bounds, instead relying on the passed scan function to error when the size limit is exceeded: https://github.com/cockroachdb/cockroach/blob/6b9168de33628257f3f1fdb17af1f8205ee21d32/pkg/kv/kvserver/replica_raftstorage.go#L236-L243 The problem is that `MVCCIterate` will actually fetch 1000 entries before calling scanFunc: https://github.com/cockroachdb/cockroach/blob/8f5231d3443f372c88632c8b77c7d954890cc334/pkg/storage/mvcc.go#L2662-L2666 https://github.com/cockroachdb/cockroach/blob/8f5231d3443f372c88632c8b77c7d954890cc334/pkg/storage/mvcc.go#L2679-L2686 Since we're using the entire Raft log as an upper bound, not passing in a `MVCCScanOptions.MaxSize` for the iterator, and only doing size limiting in the passed `scanFunc`, this means that every time we fetch a single log entry we're actually fetching and decoding 1000 entries, each 132 kb large. The simplest fix here is probably to just pass in the max size from Raft in `MVCCScanOptions`. /cc @cockroachdb/kv
True
kvserver: limit MVCC scan in Raft Entries retrieval - In a recent customer escalation, we saw a newly elected Raft leader spend so long to send 100 log entries to catch up a follower that it missed its heartbeat. The log entries were found to be about 132 kb each, and we limit these messages to be a maximum of 32 kb (but at least 1 entry), so responses were being sent with a single entry each. It seemed to be taking 60 ms to retrieve a single entry, totalling 6s for 100 log entries. The log was also huge, at ~900,000 entries, with the current commit index at 1500. Digging into the Raft code, we find that `raft.maybeSendAppend()` calls int `raftLog.entries()` to fetch entries to send to the follower. Notice how it limits this solely by `maxsize` (32 kb), otherwise using the entire Raft log (`lastIndex()+1`) as an upper bound: https://github.com/cockroachdb/vendored/blob/310ea3cd09b55055b290f0a4cf9991a91725797e/go.etcd.io/etcd/raft/v3/log.go#L287-L292 This in turn gets passed into `Storage.Entries()`: https://github.com/cockroachdb/vendored/blob/310ea3cd09b55055b290f0a4cf9991a91725797e/go.etcd.io/etcd/raft/v3/log.go#L350 Our implementation of `Entries()`, however, does not pass the max size bound to the MVCC iterator, only the key bounds, instead relying on the passed scan function to error when the size limit is exceeded: https://github.com/cockroachdb/cockroach/blob/6b9168de33628257f3f1fdb17af1f8205ee21d32/pkg/kv/kvserver/replica_raftstorage.go#L236-L243 The problem is that `MVCCIterate` will actually fetch 1000 entries before calling scanFunc: https://github.com/cockroachdb/cockroach/blob/8f5231d3443f372c88632c8b77c7d954890cc334/pkg/storage/mvcc.go#L2662-L2666 https://github.com/cockroachdb/cockroach/blob/8f5231d3443f372c88632c8b77c7d954890cc334/pkg/storage/mvcc.go#L2679-L2686 Since we're using the entire Raft log as an upper bound, not passing in a `MVCCScanOptions.MaxSize` for the iterator, and only doing size limiting in the passed `scanFunc`, this means that every time we fetch a single log entry we're actually fetching and decoding 1000 entries, each 132 kb large. The simplest fix here is probably to just pass in the max size from Raft in `MVCCScanOptions`. /cc @cockroachdb/kv
non_main
kvserver limit mvcc scan in raft entries retrieval in a recent customer escalation we saw a newly elected raft leader spend so long to send log entries to catch up a follower that it missed its heartbeat the log entries were found to be about kb each and we limit these messages to be a maximum of kb but at least entry so responses were being sent with a single entry each it seemed to be taking ms to retrieve a single entry totalling for log entries the log was also huge at entries with the current commit index at digging into the raft code we find that raft maybesendappend calls int raftlog entries to fetch entries to send to the follower notice how it limits this solely by maxsize kb otherwise using the entire raft log lastindex as an upper bound this in turn gets passed into storage entries our implementation of entries however does not pass the max size bound to the mvcc iterator only the key bounds instead relying on the passed scan function to error when the size limit is exceeded the problem is that mvcciterate will actually fetch entries before calling scanfunc since we re using the entire raft log as an upper bound not passing in a mvccscanoptions maxsize for the iterator and only doing size limiting in the passed scanfunc this means that every time we fetch a single log entry we re actually fetching and decoding entries each kb large the simplest fix here is probably to just pass in the max size from raft in mvccscanoptions cc cockroachdb kv
0
1,338
5,721,482,647
IssuesEvent
2017-04-20 06:46:16
tomchentw/react-google-maps
https://api.github.com/repos/tomchentw/react-google-maps
closed
triggerEvent('resize') does...what?
CALL_FOR_MAINTAINERS
http://react-google-maps.tomchentw.com/#/?_k=vkf06f If you load this in Chrome nightly and open the console, you can see the logs dump on the window resize event, but the map is not being updated to reflect the changing screen size. The example code just logs the component instance on resize. But how does one actually trigger the resize _on the map_, with access only to the _component_? Looking at the source code, it looks like `triggerEvent` just spreads the args, but my understanding of the gMaps API is that I need to run something like `map.setCenter` or `setBounds` or something on the event? What's the way to do this on the given implementation? I wouldn't ask, but it just doesn't seem to be working in the demo example. Thanks!
True
triggerEvent('resize') does...what? - http://react-google-maps.tomchentw.com/#/?_k=vkf06f If you load this in Chrome nightly and open the console, you can see the logs dump on the window resize event, but the map is not being updated to reflect the changing screen size. The example code just logs the component instance on resize. But how does one actually trigger the resize _on the map_, with access only to the _component_? Looking at the source code, it looks like `triggerEvent` just spreads the args, but my understanding of the gMaps API is that I need to run something like `map.setCenter` or `setBounds` or something on the event? What's the way to do this on the given implementation? I wouldn't ask, but it just doesn't seem to be working in the demo example. Thanks!
main
triggerevent resize does what if you load this in chrome nightly and open the console you can see the logs dump on the window resize event but the map is not being updated to reflect the changing screen size the example code just logs the component instance on resize but how does one actually trigger the resize on the map with access only to the component looking at the source code it looks like triggerevent just spreads the args but my understanding of the gmaps api is that i need to run something like map setcenter or setbounds or something on the event what s the way to do this on the given implementation i wouldn t ask but it just doesn t seem to be working in the demo example thanks
1
1,030
4,827,515,754
IssuesEvent
2016-11-07 13:52:44
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
temporary AWS Access Keys results in auth failures
affects_1.9 aws bug_report cloud waiting_on_maintainer
#### Issue Type: Bug Report #### Component Name: ec2 module #### Ansible Version: ansible 1.9.4 #### Ansible Configuration: none #### Environment: Mac OSX 10.11 / Not applicable #### Summary: Using temporary AWS Access Keys results in auth failures #### Steps to reproduce: generate temporary access keys (eg. via STS or SAML provider) attempt to create ec2 resource #### Expected results: ec2 resource is created #### Actual results: AWS AuthFailure exception I originally lodged this against ansible/ansible as https://github.com/ansible/ansible/issues/12959 but I think maybe this should be resolved in the core modules code. There's a comment in the code that suggests, that perhaps the modules just need to be modified to use `connect_to_aws()` ``` def get_ec2_creds(module): ''' for compatibility mode with old modules that don't/can't yet use ec2_connect method ''' ``` ``` failed: [localhost] => {"failed": true, "parsed": false} Traceback (most recent call last): File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 3070, in <module> main() File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 1249, in main (instance_dict_array, new_instance_ids, changed) = create_instances(module, ec2, vpc) File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 792, in create_instances vpc_id = vpc.get_all_subnets(subnet_ids=[vpc_subnet_id])[0].vpc_id File "/Library/Python/2.7/site-packages/boto/vpc/__init__.py", line 1153, in get_all_subnets return self.get_list('DescribeSubnets', params, [('item', Subnet)]) File "/Library/Python/2.7/site-packages/boto/connection.py", line 1186, in get_list raise self.ResponseError(response.status, response.reason, body) boto.exception.EC2ResponseError: EC2ResponseError: 401 Unauthorized <?xml version="1.0" encoding="UTF-8"?> <Response><Errors><Error><Code>AuthFailure</Code><Message>AWS was not able to validate the provided access credentials</Message></Error></Errors><RequestID>bcce0f14-b8d4-46e0-a582-17993365b18b</RequestID></Response> from my investigation the issue appears in the module_utils/ec2.py get_ec2_creds( ) which returns ec2_url, boto_params['aws_access_key_id'], boto_params['aws_secret_access_key'], region ``` since the aws_access_key_id in this example will only work with a security_token, the method is effectively broken. I think the function should at least warn when it detects a security token or an access key that starts with ASIA instead of AKIA.
True
temporary AWS Access Keys results in auth failures - #### Issue Type: Bug Report #### Component Name: ec2 module #### Ansible Version: ansible 1.9.4 #### Ansible Configuration: none #### Environment: Mac OSX 10.11 / Not applicable #### Summary: Using temporary AWS Access Keys results in auth failures #### Steps to reproduce: generate temporary access keys (eg. via STS or SAML provider) attempt to create ec2 resource #### Expected results: ec2 resource is created #### Actual results: AWS AuthFailure exception I originally lodged this against ansible/ansible as https://github.com/ansible/ansible/issues/12959 but I think maybe this should be resolved in the core modules code. There's a comment in the code that suggests, that perhaps the modules just need to be modified to use `connect_to_aws()` ``` def get_ec2_creds(module): ''' for compatibility mode with old modules that don't/can't yet use ec2_connect method ''' ``` ``` failed: [localhost] => {"failed": true, "parsed": false} Traceback (most recent call last): File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 3070, in <module> main() File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 1249, in main (instance_dict_array, new_instance_ids, changed) = create_instances(module, ec2, vpc) File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 792, in create_instances vpc_id = vpc.get_all_subnets(subnet_ids=[vpc_subnet_id])[0].vpc_id File "/Library/Python/2.7/site-packages/boto/vpc/__init__.py", line 1153, in get_all_subnets return self.get_list('DescribeSubnets', params, [('item', Subnet)]) File "/Library/Python/2.7/site-packages/boto/connection.py", line 1186, in get_list raise self.ResponseError(response.status, response.reason, body) boto.exception.EC2ResponseError: EC2ResponseError: 401 Unauthorized <?xml version="1.0" encoding="UTF-8"?> <Response><Errors><Error><Code>AuthFailure</Code><Message>AWS was not able to validate the provided access credentials</Message></Error></Errors><RequestID>bcce0f14-b8d4-46e0-a582-17993365b18b</RequestID></Response> from my investigation the issue appears in the module_utils/ec2.py get_ec2_creds( ) which returns ec2_url, boto_params['aws_access_key_id'], boto_params['aws_secret_access_key'], region ``` since the aws_access_key_id in this example will only work with a security_token, the method is effectively broken. I think the function should at least warn when it detects a security token or an access key that starts with ASIA instead of AKIA.
main
temporary aws access keys results in auth failures issue type bug report component name module ansible version ansible ansible configuration none environment mac osx not applicable summary using temporary aws access keys results in auth failures steps to reproduce generate temporary access keys eg via sts or saml provider attempt to create resource expected results resource is created actual results aws authfailure exception i originally lodged this against ansible ansible as but i think maybe this should be resolved in the core modules code there s a comment in the code that suggests that perhaps the modules just need to be modified to use connect to aws def get creds module for compatibility mode with old modules that don t can t yet use connect method failed failed true parsed false traceback most recent call last file users secole ansible tmp ansible tmp line in main file users secole ansible tmp ansible tmp line in main instance dict array new instance ids changed create instances module vpc file users secole ansible tmp ansible tmp line in create instances vpc id vpc get all subnets subnet ids vpc id file library python site packages boto vpc init py line in get all subnets return self get list describesubnets params file library python site packages boto connection py line in get list raise self responseerror response status response reason body boto exception unauthorized authfailure aws was not able to validate the provided access credentials from my investigation the issue appears in the module utils py get creds which returns url boto params boto params region since the aws access key id in this example will only work with a security token the method is effectively broken i think the function should at least warn when it detects a security token or an access key that starts with asia instead of akia
1
285,789
31,155,588,778
IssuesEvent
2023-08-16 12:56:53
nidhi7598/linux-4.1.15_CVE-2018-5873
https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2018-5873
opened
CVE-2022-22764 (High) detected in linuxlinux-4.1.52
Mend: dependency security vulnerability
## CVE-2022-22764 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Mozilla developers Paul Adenot and the Mozilla Fuzzing Team reported memory safety bugs present in Firefox 96 and Firefox ESR 91.5. Some of these bugs showed evidence of memory corruption and we presume that with enough effort some of these could have been exploited to run arbitrary code. This vulnerability affects Firefox < 97, Thunderbird < 91.6, and Firefox ESR < 91.6. <p>Publish Date: 2022-12-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22764>CVE-2022-22764</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-22764">https://nvd.nist.gov/vuln/detail/CVE-2022-22764</a></p> <p>Release Date: 2022-01-07</p> <p>Fix Resolution: linux-libc-headers - 5.14;linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-22764 (High) detected in linuxlinux-4.1.52 - ## CVE-2022-22764 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Mozilla developers Paul Adenot and the Mozilla Fuzzing Team reported memory safety bugs present in Firefox 96 and Firefox ESR 91.5. Some of these bugs showed evidence of memory corruption and we presume that with enough effort some of these could have been exploited to run arbitrary code. This vulnerability affects Firefox < 97, Thunderbird < 91.6, and Firefox ESR < 91.6. <p>Publish Date: 2022-12-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22764>CVE-2022-22764</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-22764">https://nvd.nist.gov/vuln/detail/CVE-2022-22764</a></p> <p>Release Date: 2022-01-07</p> <p>Fix Resolution: linux-libc-headers - 5.14;linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files vulnerability details mozilla developers paul adenot and the mozilla fuzzing team reported memory safety bugs present in firefox and firefox esr some of these bugs showed evidence of memory corruption and we presume that with enough effort some of these could have been exploited to run arbitrary code this vulnerability affects firefox thunderbird and firefox esr publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux libc headers linux yocto gitautoinc gitautoinc step up your open source security game with mend
0
32,686
8,921,078,995
IssuesEvent
2019-01-21 09:08:32
neovim/neovim
https://api.github.com/repos/neovim/neovim
closed
(clang >= 6.0 bug) Annoying warnings for isnan(), fpclassify(), et al
blocked:external build help wanted
- `nvim --version`: ef4feab0e75be - Vim (version: 8.0.1565) behaves differently? No warnings, but I didn't check if the same functions were used - Operating system/version: arch linux - Terminal name/version: pangoterm - `$TERM`: xterm ### Steps to reproduce using `nvim -u NORC` ``` rm -rf build && CMAKE_EXTRA_FLAGS="-DCMAKE_C_COMPILER=clang -DCLANG_ASAN_UBSAN=1" make -j4 ``` ### Actual behaviour ``` [197/284] Building C object src/nvim/CMakeFiles/nvim.dir/eval/encode.c.o In file included from ../src/nvim/eval/encode.c:455: ../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv ersion] TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:455: ../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon version] TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:493: ../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv ersion] TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:493: ../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon version] TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:762: ../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv ersion] TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:762: ../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon version] TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ 6 warnings generated. [261/284] Building C object src/nvim/CMakeFiles/nvim.dir/strings.c.o ../src/nvim/strings.c:1223:23: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion] if (isinf((double)f) ~~~~~~^~~~~~~~~~ /usr/include/math.h:472:46: note: expanded from macro 'isinf' # define isinf(x) __MATH_TG ((x), __isinf, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ ../src/nvim/strings.c:1230:30: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion] } else if (isnan(f)) { ~~~~~~^~ /usr/include/math.h:455:46: note: expanded from macro 'isnan' # define isnan(x) __MATH_TG ((x), __isnan, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ 2 warnings generated. ``` ### Expected behaviour No warnings. `maths.h` uses gcc magic to avoid this warning when compiling with gcc 4.4+, which doesn't work with clang (unless C11 is active, but using a different C version than the project's chosen one is probably a bad idea). Anyone thinking of a better workaround than moving float code to a special `-Wno-conversion` c file?
1.0
(clang >= 6.0 bug) Annoying warnings for isnan(), fpclassify(), et al - - `nvim --version`: ef4feab0e75be - Vim (version: 8.0.1565) behaves differently? No warnings, but I didn't check if the same functions were used - Operating system/version: arch linux - Terminal name/version: pangoterm - `$TERM`: xterm ### Steps to reproduce using `nvim -u NORC` ``` rm -rf build && CMAKE_EXTRA_FLAGS="-DCMAKE_C_COMPILER=clang -DCLANG_ASAN_UBSAN=1" make -j4 ``` ### Actual behaviour ``` [197/284] Building C object src/nvim/CMakeFiles/nvim.dir/eval/encode.c.o In file included from ../src/nvim/eval/encode.c:455: ../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv ersion] TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:455: ../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon version] TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:493: ../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv ersion] TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:493: ../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon version] TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:762: ../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv ersion] TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ In file included from ../src/nvim/eval/encode.c:762: ../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon version] TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float); ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT' switch (fpclassify(flt_)) { \ ~~~~~~~~~~~^~~~~ /usr/include/math.h:415:56: note: expanded from macro 'fpclassify' # define fpclassify(x) __MATH_TG ((x), __fpclassify, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ 6 warnings generated. [261/284] Building C object src/nvim/CMakeFiles/nvim.dir/strings.c.o ../src/nvim/strings.c:1223:23: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion] if (isinf((double)f) ~~~~~~^~~~~~~~~~ /usr/include/math.h:472:46: note: expanded from macro 'isinf' # define isinf(x) __MATH_TG ((x), __isinf, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ ../src/nvim/strings.c:1230:30: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion] } else if (isnan(f)) { ~~~~~~^~ /usr/include/math.h:455:46: note: expanded from macro 'isnan' # define isnan(x) __MATH_TG ((x), __isnan, (x)) ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~ /usr/include/math.h:370:16: note: expanded from macro '__MATH_TG' ? FUNC ## f ARGS \ ~~~~~~~~~ ^~~~ 2 warnings generated. ``` ### Expected behaviour No warnings. `maths.h` uses gcc magic to avoid this warning when compiling with gcc 4.4+, which doesn't work with clang (unless C11 is active, but using a different C version than the project's chosen one is probably a bad idea). Anyone thinking of a better workaround than moving float code to a special `-Wno-conversion` c file?
non_main
clang bug annoying warnings for isnan fpclassify et al nvim version vim version behaves differently no warnings but i didn t check if the same functions were used operating system version arch linux terminal name version pangoterm term xterm steps to reproduce using nvim u norc rm rf build cmake extra flags dcmake c compiler clang dclang asan ubsan make actual behaviour building c object src nvim cmakefiles nvim dir eval encode c o in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args warnings generated building c object src nvim cmakefiles nvim dir strings c o src nvim strings c warning implicit conversion loses floating point precision double to float if isinf double f usr include math h note expanded from macro isinf define isinf x math tg x isinf x usr include math h note expanded from macro math tg func f args src nvim strings c warning implicit conversion loses floating point precision double to float else if isnan f usr include math h note expanded from macro isnan define isnan x math tg x isnan x usr include math h note expanded from macro math tg func f args warnings generated expected behaviour no warnings maths h uses gcc magic to avoid this warning when compiling with gcc which doesn t work with clang unless is active but using a different c version than the project s chosen one is probably a bad idea anyone thinking of a better workaround than moving float code to a special wno conversion c file
0
2,654
9,083,455,570
IssuesEvent
2019-02-17 20:37:25
pound-python/infobob
https://api.github.com/repos/pound-python/infobob
closed
Remove unused commands
maintainability
`infobob_exec`, `infobob_codepad`, and `infobob_print` can be removed, they haven't been active in a while, and it'd doubtful they even work still.
True
Remove unused commands - `infobob_exec`, `infobob_codepad`, and `infobob_print` can be removed, they haven't been active in a while, and it'd doubtful they even work still.
main
remove unused commands infobob exec infobob codepad and infobob print can be removed they haven t been active in a while and it d doubtful they even work still
1
428
3,516,654,646
IssuesEvent
2016-01-12 01:03:29
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
Java cask fails to install
awaiting maintainer feedback bug cask
### Description of issue Java cask will not install. ### Output of `brew cask install java --verbose` ``` $ brew cask install java --verbose ==> Caveats This Cask makes minor modifications to the JRE to prevent issues with packaged applications, as discussed here: https://bugs.eclipse.org/bugs/show_bug.cgi?id=411361 If your Java application still asks for JRE installation, you might need to reboot or logout/login. Installing this Cask means you have AGREED to the Oracle Binary Code License Agreement for Java SE at http://www.oracle.com/technetwork/java/javase/terms/license/index.html ==> Downloading http://download.oracle.com/otn-pub/java/jdk/8u66-b17/jdk-8u66-macosx-x64.dmg Already downloaded: /Library/Caches/Homebrew/java-1.8.0_66-b17.dmg Error: Permission denied - (/Library/Caches/Homebrew/java-1.8.0_66-b17.dmg, /Library/Caches/Homebrew/Casks/java-1.8.0_66-b17.dmg) Most likely, this means you have an outdated version of homebrew-cask. Please run: brew uninstall --force brew-cask; brew untap phinze/cask; brew update; brew cleanup; brew cask cleanup If this doesn’t fix the problem, please report this bug: https://github.com/caskroom/homebrew-cask#reporting-bugs /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:349:in `symlink' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:349:in `block in ln_s' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:1574:in `fu_each_src_dest0' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:347:in `ln_s' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:370:in `ln_sf' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/download.rb:45:in `create_cache_symlink' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/download.rb:15:in `perform' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/installer.rb:94:in `download' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/installer.rb:68:in `install' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:20:in `block in install_casks' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:17:in `each' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:17:in `install_casks' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:6:in `run' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli.rb:78:in `run_command' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli.rb:116:in `process' /usr/local/Library/Taps/caskroom/homebrew-cask/cmd/brew-cask.rb:26:in `<top (required)>' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:55:in `require' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:55:in `require' /usr/local/Library/brew.rb:69:in `require?' /usr/local/Library/brew.rb:149:in `<main>' Error: Kernel.exit ``` ### Output of `brew doctor` ``` $ brew doctor Your system is ready to brew. ``` ### Output of `brew cask doctor` ``` $ brew cask doctor ==> OS X Release: 10.11 ==> OS X Release with Patchlevel: 10.11.1 ==> Hardware Architecture: intel-64 ==> Ruby Version: 2.0.0-p645 ==> Ruby Path: /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/bin/ruby ==> Homebrew Version: Homebrew 0.9.5 (git revision 3f86; last commit 2016-01-07) ==> Homebrew Executable Path: /usr/local/bin/brew ==> Homebrew Cellar Path: /usr/local/Cellar ==> Homebrew Repository Path: /usr/local ==> Homebrew Origin: https://github.com/Homebrew/homebrew ==> Homebrew-cask Version: 0.60.0 (git revision a595; last commit 41 minutes ago) ==> Homebrew-cask Install Location: <NONE> ==> Homebrew-cask Staging Location: /opt/homebrew-cask/Caskroom ==> Homebrew-cask Cached Downloads: /Library/Caches/Homebrew /Library/Caches/Homebrew/Casks 0 files 0.00 megs ==> Homebrew-cask Default Tap Path: /usr/local/Library/Taps/caskroom/homebrew-cask ==> Homebrew-cask Alternate Cask Taps: <NONE> ==> Homebrew-cask Default Tap Cask Count: 2941 ==> Contents of $LOAD_PATH: /usr/local/Library/Taps/caskroom/homebrew-cask/lib /usr/local/Library/Homebrew /Library/Ruby/Site/2.0.0 /Library/Ruby/Site/2.0.0/x86_64-darwin15 /Library/Ruby/Site/2.0.0/universal-darwin15 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin15 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin15 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin15 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin15 ==> Contents of $RUBYLIB Environment Variable: <NONE> ==> Contents of $RUBYOPT Environment Variable: <NONE> ==> Contents of $RUBYPATH Environment Variable: <NONE> ==> Contents of $RBENV_VERSION Environment Variable: <NONE> ==> Contents of $CHRUBY_VERSION Environment Variable: <NONE> ==> Contents of $GEM_HOME Environment Variable: <NONE> ==> Contents of $GEM_PATH Environment Variable: <NONE> ==> Contents of $BUNDLE_PATH Environment Variable: <NONE> ==> Contents of $PATH Environment Variable: PATH="/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/Library/Taps/caskroom/homebrew-cask/cmd:/usr/local/Library/ENV/scm" ==> Contents of $SHELL Environment Variable: SHELL="/bin/bash" ==> Contents of Locale Environment Variables: LANG="en_US.UTF-8" ==> Running As Privileged User: No ```
True
Java cask fails to install - ### Description of issue Java cask will not install. ### Output of `brew cask install java --verbose` ``` $ brew cask install java --verbose ==> Caveats This Cask makes minor modifications to the JRE to prevent issues with packaged applications, as discussed here: https://bugs.eclipse.org/bugs/show_bug.cgi?id=411361 If your Java application still asks for JRE installation, you might need to reboot or logout/login. Installing this Cask means you have AGREED to the Oracle Binary Code License Agreement for Java SE at http://www.oracle.com/technetwork/java/javase/terms/license/index.html ==> Downloading http://download.oracle.com/otn-pub/java/jdk/8u66-b17/jdk-8u66-macosx-x64.dmg Already downloaded: /Library/Caches/Homebrew/java-1.8.0_66-b17.dmg Error: Permission denied - (/Library/Caches/Homebrew/java-1.8.0_66-b17.dmg, /Library/Caches/Homebrew/Casks/java-1.8.0_66-b17.dmg) Most likely, this means you have an outdated version of homebrew-cask. Please run: brew uninstall --force brew-cask; brew untap phinze/cask; brew update; brew cleanup; brew cask cleanup If this doesn’t fix the problem, please report this bug: https://github.com/caskroom/homebrew-cask#reporting-bugs /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:349:in `symlink' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:349:in `block in ln_s' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:1574:in `fu_each_src_dest0' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:347:in `ln_s' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:370:in `ln_sf' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/download.rb:45:in `create_cache_symlink' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/download.rb:15:in `perform' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/installer.rb:94:in `download' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/installer.rb:68:in `install' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:20:in `block in install_casks' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:17:in `each' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:17:in `install_casks' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:6:in `run' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli.rb:78:in `run_command' /usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli.rb:116:in `process' /usr/local/Library/Taps/caskroom/homebrew-cask/cmd/brew-cask.rb:26:in `<top (required)>' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:55:in `require' /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:55:in `require' /usr/local/Library/brew.rb:69:in `require?' /usr/local/Library/brew.rb:149:in `<main>' Error: Kernel.exit ``` ### Output of `brew doctor` ``` $ brew doctor Your system is ready to brew. ``` ### Output of `brew cask doctor` ``` $ brew cask doctor ==> OS X Release: 10.11 ==> OS X Release with Patchlevel: 10.11.1 ==> Hardware Architecture: intel-64 ==> Ruby Version: 2.0.0-p645 ==> Ruby Path: /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/bin/ruby ==> Homebrew Version: Homebrew 0.9.5 (git revision 3f86; last commit 2016-01-07) ==> Homebrew Executable Path: /usr/local/bin/brew ==> Homebrew Cellar Path: /usr/local/Cellar ==> Homebrew Repository Path: /usr/local ==> Homebrew Origin: https://github.com/Homebrew/homebrew ==> Homebrew-cask Version: 0.60.0 (git revision a595; last commit 41 minutes ago) ==> Homebrew-cask Install Location: <NONE> ==> Homebrew-cask Staging Location: /opt/homebrew-cask/Caskroom ==> Homebrew-cask Cached Downloads: /Library/Caches/Homebrew /Library/Caches/Homebrew/Casks 0 files 0.00 megs ==> Homebrew-cask Default Tap Path: /usr/local/Library/Taps/caskroom/homebrew-cask ==> Homebrew-cask Alternate Cask Taps: <NONE> ==> Homebrew-cask Default Tap Cask Count: 2941 ==> Contents of $LOAD_PATH: /usr/local/Library/Taps/caskroom/homebrew-cask/lib /usr/local/Library/Homebrew /Library/Ruby/Site/2.0.0 /Library/Ruby/Site/2.0.0/x86_64-darwin15 /Library/Ruby/Site/2.0.0/universal-darwin15 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin15 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin15 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin15 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin15 ==> Contents of $RUBYLIB Environment Variable: <NONE> ==> Contents of $RUBYOPT Environment Variable: <NONE> ==> Contents of $RUBYPATH Environment Variable: <NONE> ==> Contents of $RBENV_VERSION Environment Variable: <NONE> ==> Contents of $CHRUBY_VERSION Environment Variable: <NONE> ==> Contents of $GEM_HOME Environment Variable: <NONE> ==> Contents of $GEM_PATH Environment Variable: <NONE> ==> Contents of $BUNDLE_PATH Environment Variable: <NONE> ==> Contents of $PATH Environment Variable: PATH="/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/Library/Taps/caskroom/homebrew-cask/cmd:/usr/local/Library/ENV/scm" ==> Contents of $SHELL Environment Variable: SHELL="/bin/bash" ==> Contents of Locale Environment Variables: LANG="en_US.UTF-8" ==> Running As Privileged User: No ```
main
java cask fails to install description of issue java cask will not install output of brew cask install java verbose brew cask install java verbose caveats this cask makes minor modifications to the jre to prevent issues with packaged applications as discussed here if your java application still asks for jre installation you might need to reboot or logout login installing this cask means you have agreed to the oracle binary code license agreement for java se at downloading already downloaded library caches homebrew java dmg error permission denied library caches homebrew java dmg library caches homebrew casks java dmg most likely this means you have an outdated version of homebrew cask please run brew uninstall force brew cask brew untap phinze cask brew update brew cleanup brew cask cleanup if this doesn’t fix the problem please report this bug system library frameworks ruby framework versions usr lib ruby fileutils rb in symlink system library frameworks ruby framework versions usr lib ruby fileutils rb in block in ln s system library frameworks ruby framework versions usr lib ruby fileutils rb in fu each src system library frameworks ruby framework versions usr lib ruby fileutils rb in ln s system library frameworks ruby framework versions usr lib ruby fileutils rb in ln sf usr local library taps caskroom homebrew cask lib hbc download rb in create cache symlink usr local library taps caskroom homebrew cask lib hbc download rb in perform usr local library taps caskroom homebrew cask lib hbc installer rb in download usr local library taps caskroom homebrew cask lib hbc installer rb in install usr local library taps caskroom homebrew cask lib hbc cli install rb in block in install casks usr local library taps caskroom homebrew cask lib hbc cli install rb in each usr local library taps caskroom homebrew cask lib hbc cli install rb in install casks usr local library taps caskroom homebrew cask lib hbc cli install rb in run usr local library taps caskroom homebrew cask lib hbc cli rb in run command usr local library taps caskroom homebrew cask lib hbc cli rb in process usr local library taps caskroom homebrew cask cmd brew cask rb in system library frameworks ruby framework versions usr lib ruby rubygems core ext kernel require rb in require system library frameworks ruby framework versions usr lib ruby rubygems core ext kernel require rb in require usr local library brew rb in require usr local library brew rb in error kernel exit output of brew doctor brew doctor your system is ready to brew output of brew cask doctor brew cask doctor os x release os x release with patchlevel hardware architecture intel ruby version ruby path system library frameworks ruby framework versions usr bin ruby homebrew version homebrew git revision last commit homebrew executable path usr local bin brew homebrew cellar path usr local cellar homebrew repository path usr local homebrew origin homebrew cask version git revision last commit minutes ago homebrew cask install location homebrew cask staging location opt homebrew cask caskroom homebrew cask cached downloads library caches homebrew library caches homebrew casks files megs homebrew cask default tap path usr local library taps caskroom homebrew cask homebrew cask alternate cask taps homebrew cask default tap cask count contents of load path usr local library taps caskroom homebrew cask lib usr local library homebrew library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal contents of rubylib environment variable contents of rubyopt environment variable contents of rubypath environment variable contents of rbenv version environment variable contents of chruby version environment variable contents of gem home environment variable contents of gem path environment variable contents of bundle path environment variable contents of path environment variable path usr local bin usr bin bin usr sbin sbin usr local library taps caskroom homebrew cask cmd usr local library env scm contents of shell environment variable shell bin bash contents of locale environment variables lang en us utf running as privileged user no
1
101,110
30,876,042,897
IssuesEvent
2023-08-03 14:21:31
helidon-io/helidon
https://api.github.com/repos/helidon-io/helidon
opened
[4.x] `Prototype.BuilderInterceptor` was renamed to `Prototype.BuilderDecorator`; the builder README and `@Prototype.Blueprint` attributes are now inconsistent with the change
4.x builder
## Environment Details * Helidon Version: 4.x * Helidon SE or Helidon MP * JDK version: * OS: * Docker version (if applicable): ---------- ## Problem Description A recent PR #7281 renamed `Prototype.BuilderInterceptor` to `Prototype.BuilderDecorator`, but: * the builder `README.md` was not updated * the `@Prototype.Blueprint` annotation still uses the `builderInterceptor` attribute name.
1.0
[4.x] `Prototype.BuilderInterceptor` was renamed to `Prototype.BuilderDecorator`; the builder README and `@Prototype.Blueprint` attributes are now inconsistent with the change - ## Environment Details * Helidon Version: 4.x * Helidon SE or Helidon MP * JDK version: * OS: * Docker version (if applicable): ---------- ## Problem Description A recent PR #7281 renamed `Prototype.BuilderInterceptor` to `Prototype.BuilderDecorator`, but: * the builder `README.md` was not updated * the `@Prototype.Blueprint` annotation still uses the `builderInterceptor` attribute name.
non_main
prototype builderinterceptor was renamed to prototype builderdecorator the builder readme and prototype blueprint attributes are now inconsistent with the change environment details helidon version x helidon se or helidon mp jdk version os docker version if applicable problem description a recent pr renamed prototype builderinterceptor to prototype builderdecorator but the builder readme md was not updated the prototype blueprint annotation still uses the builderinterceptor attribute name
0
45,917
13,055,821,972
IssuesEvent
2020-07-30 02:50:20
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
libarchive review (Trac #261)
Incomplete Migration Migrated from Trac combo core defect
Migrated from https://code.icecube.wisc.edu/ticket/261 ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "See: #IT282", "reporter": "nega", "cc": "", "resolution": "worksforme", "_ts": "1416713877165085", "component": "combo core", "summary": "libarchive review", "priority": "normal", "keywords": "libarchive", "time": "2011-05-11T20:39:48", "milestone": "", "owner": "olivas", "type": "defect" } ```
1.0
libarchive review (Trac #261) - Migrated from https://code.icecube.wisc.edu/ticket/261 ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "See: #IT282", "reporter": "nega", "cc": "", "resolution": "worksforme", "_ts": "1416713877165085", "component": "combo core", "summary": "libarchive review", "priority": "normal", "keywords": "libarchive", "time": "2011-05-11T20:39:48", "milestone": "", "owner": "olivas", "type": "defect" } ```
non_main
libarchive review trac migrated from json status closed changetime description see reporter nega cc resolution worksforme ts component combo core summary libarchive review priority normal keywords libarchive time milestone owner olivas type defect
0
75,271
7,467,696,132
IssuesEvent
2018-04-02 16:15:37
GoogleCloudPlatform/forseti-security
https://api.github.com/repos/GoogleCloudPlatform/forseti-security
closed
Enable the IAM and firewall notification by default.
module: notifier priority: p1 release-testing: 2.0 RC2
This is currently disabled. But after removing the notification from the scanner, this needs to be enabled. https://github.com/GoogleCloudPlatform/forseti-security/blob/2.0-dev/configs/server/forseti_conf_server.yaml.in#L85
1.0
Enable the IAM and firewall notification by default. - This is currently disabled. But after removing the notification from the scanner, this needs to be enabled. https://github.com/GoogleCloudPlatform/forseti-security/blob/2.0-dev/configs/server/forseti_conf_server.yaml.in#L85
non_main
enable the iam and firewall notification by default this is currently disabled but after removing the notification from the scanner this needs to be enabled
0
296,732
22,321,314,747
IssuesEvent
2022-06-14 06:43:34
elastic/package-spec
https://api.github.com/repos/elastic/package-spec
opened
Create implementation guide for the spec
documentation
The package spec contains descriptions of resources that can be installed in the Elastic Stack. Some resources need special management that is not always trivial ([example](https://github.com/elastic/package-spec/pull/307/files#r895515041)). It may be good to maintain in the spec a reference documentation for clients that implement the package spec, as Fleet. This documentation may contain a list of general practices for the resources defined, as well as resource-specific considerations. We may need to target different clients, e.g. installation use cases like in fleet vs. inspection use cases as in elastic-package or the registries.
1.0
Create implementation guide for the spec - The package spec contains descriptions of resources that can be installed in the Elastic Stack. Some resources need special management that is not always trivial ([example](https://github.com/elastic/package-spec/pull/307/files#r895515041)). It may be good to maintain in the spec a reference documentation for clients that implement the package spec, as Fleet. This documentation may contain a list of general practices for the resources defined, as well as resource-specific considerations. We may need to target different clients, e.g. installation use cases like in fleet vs. inspection use cases as in elastic-package or the registries.
non_main
create implementation guide for the spec the package spec contains descriptions of resources that can be installed in the elastic stack some resources need special management that is not always trivial it may be good to maintain in the spec a reference documentation for clients that implement the package spec as fleet this documentation may contain a list of general practices for the resources defined as well as resource specific considerations we may need to target different clients e g installation use cases like in fleet vs inspection use cases as in elastic package or the registries
0
101,828
4,141,752,446
IssuesEvent
2016-06-14 07:20:38
qspin/qtaste
https://api.github.com/repos/qspin/qtaste
closed
QTaste doesn't generate all Test API documentation anymore
category:kernel priority:low status:To be validated type:update
In QTaste 2.1.0 and current 2.2.0-SNAPSHOT, the Test API documentation generated in /target/TestAPI-doc/ doesn't include the documentation of the other Test API on which the Test API depends anymore. E.g. at BMS level, I had the documentation of BMS, BDS, BSS, ICOMP, QTaste-Toolbox, etc..., now I only have BMS documentation. There is a maven inheritance configuration issue in qtaste-testapi-parent: the configuration of maven-dependency-plugin is not inherited anymore! Also even if I override the configuration myself, JavaGUI is not included because the package name doesn't match the includes: &lt;includes&gt;**/qtaste/testapi/api/*.java&lt;/includes&gt; must be changed to: &lt;includes&gt;**/qtaste/**/testapi/api/*.java&lt;/includes&gt;
1.0
QTaste doesn't generate all Test API documentation anymore - In QTaste 2.1.0 and current 2.2.0-SNAPSHOT, the Test API documentation generated in /target/TestAPI-doc/ doesn't include the documentation of the other Test API on which the Test API depends anymore. E.g. at BMS level, I had the documentation of BMS, BDS, BSS, ICOMP, QTaste-Toolbox, etc..., now I only have BMS documentation. There is a maven inheritance configuration issue in qtaste-testapi-parent: the configuration of maven-dependency-plugin is not inherited anymore! Also even if I override the configuration myself, JavaGUI is not included because the package name doesn't match the includes: &lt;includes&gt;**/qtaste/testapi/api/*.java&lt;/includes&gt; must be changed to: &lt;includes&gt;**/qtaste/**/testapi/api/*.java&lt;/includes&gt;
non_main
qtaste doesn t generate all test api documentation anymore in qtaste and current snapshot the test api documentation generated in target testapi doc doesn t include the documentation of the other test api on which the test api depends anymore e g at bms level i had the documentation of bms bds bss icomp qtaste toolbox etc now i only have bms documentation there is a maven inheritance configuration issue in qtaste testapi parent the configuration of maven dependency plugin is not inherited anymore also even if i override the configuration myself javagui is not included because the package name doesn t match the includes lt includes gt qtaste testapi api java lt includes gt must be changed to lt includes gt qtaste testapi api java lt includes gt
0
1,306
5,546,361,385
IssuesEvent
2017-03-23 00:35:46
OpenLightingProject/ola
https://api.github.com/repos/OpenLightingProject/ola
opened
Packing structs on Windows
Difficulty-Easy Language-C++ Maintainability OpSys-Windows
Do it the same way Jack does: https://github.com/jackaudio/jack2/blob/master/common/jack/systemdeps.h So we can then add that code to here: https://github.com/OpenLightingProject/ola/blob/master/include/ola/base/Macro.h#L139 And use it whenever we're packing structs with PRE_PACK and POST_PACK.
True
Packing structs on Windows - Do it the same way Jack does: https://github.com/jackaudio/jack2/blob/master/common/jack/systemdeps.h So we can then add that code to here: https://github.com/OpenLightingProject/ola/blob/master/include/ola/base/Macro.h#L139 And use it whenever we're packing structs with PRE_PACK and POST_PACK.
main
packing structs on windows do it the same way jack does so we can then add that code to here and use it whenever we re packing structs with pre pack and post pack
1
1,898
6,577,549,525
IssuesEvent
2017-09-12 01:41:36
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
while running a copy module task, ansible requested this info passed to the list.
affects_2.0 bug_report waiting_on_maintainer
##### Issue Type: - Bug Report ##### Plugin Name: <!-- Name of the plugin/module/task --> ##### Ansible Version: ``` 23:06 $ ansible --version ansible 2.0.0.2 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### Ansible Configuration: Dynamic inventory from dopy ``` [ssh_connection] scp_if_ssh=True ssh_args= -o ForwardAgent=yes ``` ##### Environment: OSX 10.10 host Ubuntu 14.04 remote ##### Summary: While running playbook, ansible reported back ``` [WARNING]: Calculating checksum failed unusually, please report this to the list so it can be fixed command: rc=flag; [ -r /etc/apt/apt.conf.d/10periodic ] || rc=2; [ -f /etc/apt/apt.conf.d/10periodic ] || rc=1; [ -d /etc/apt/apt.conf.d/10periodic ] && rc=3; python -V 2>/dev/null || rc=4; [ x"$rc" != "xflag" ] && echo "${rc} "/etc/apt/apt.conf.d/10periodic && exit 0; (python -c 'import hashlib; BLOCKSIZE = 65536; hasher = hashlib.sha1(); afile = open("'/etc/apt/apt.conf.d/10periodic'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0: hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (python -c 'import sha; BLOCKSIZE = 65536; hasher = sha.sha(); afile = open("'/etc/apt/apt.conf.d/10periodic'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0: hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (echo '0 '/etc/apt/apt.conf.d/10periodic) ---- output: {'stdout_lines': [], 'stdout': u'', 'stderr': u'', 'rc': 255} ---- ``` ##### Steps To Reproduce: Unable to reproduce. Reporting because ansible said to. ``` - name: Adjust APT update intervals copy: src: config/apt_periodic dest: /etc/apt/apt.conf.d/10periodic ```
True
while running a copy module task, ansible requested this info passed to the list. - ##### Issue Type: - Bug Report ##### Plugin Name: <!-- Name of the plugin/module/task --> ##### Ansible Version: ``` 23:06 $ ansible --version ansible 2.0.0.2 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### Ansible Configuration: Dynamic inventory from dopy ``` [ssh_connection] scp_if_ssh=True ssh_args= -o ForwardAgent=yes ``` ##### Environment: OSX 10.10 host Ubuntu 14.04 remote ##### Summary: While running playbook, ansible reported back ``` [WARNING]: Calculating checksum failed unusually, please report this to the list so it can be fixed command: rc=flag; [ -r /etc/apt/apt.conf.d/10periodic ] || rc=2; [ -f /etc/apt/apt.conf.d/10periodic ] || rc=1; [ -d /etc/apt/apt.conf.d/10periodic ] && rc=3; python -V 2>/dev/null || rc=4; [ x"$rc" != "xflag" ] && echo "${rc} "/etc/apt/apt.conf.d/10periodic && exit 0; (python -c 'import hashlib; BLOCKSIZE = 65536; hasher = hashlib.sha1(); afile = open("'/etc/apt/apt.conf.d/10periodic'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0: hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (python -c 'import sha; BLOCKSIZE = 65536; hasher = sha.sha(); afile = open("'/etc/apt/apt.conf.d/10periodic'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0: hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (echo '0 '/etc/apt/apt.conf.d/10periodic) ---- output: {'stdout_lines': [], 'stdout': u'', 'stderr': u'', 'rc': 255} ---- ``` ##### Steps To Reproduce: Unable to reproduce. Reporting because ansible said to. ``` - name: Adjust APT update intervals copy: src: config/apt_periodic dest: /etc/apt/apt.conf.d/10periodic ```
main
while running a copy module task ansible requested this info passed to the list issue type bug report plugin name ansible version ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides ansible configuration dynamic inventory from dopy scp if ssh true ssh args o forwardagent yes environment osx host ubuntu remote summary while running playbook ansible reported back calculating checksum failed unusually please report this to the list so it can be fixed command rc flag r etc apt apt conf d rc rc rc python v dev null rc echo rc etc apt apt conf d exit python c import hashlib blocksize hasher hashlib afile open etc apt apt conf d rb buf afile read blocksize while len buf hasher update buf buf afile read blocksize afile close print hasher hexdigest dev null python c import sha blocksize hasher sha sha afile open etc apt apt conf d rb buf afile read blocksize while len buf hasher update buf buf afile read blocksize afile close print hasher hexdigest dev null echo etc apt apt conf d output stdout lines stdout u stderr u rc steps to reproduce unable to reproduce reporting because ansible said to name adjust apt update intervals copy src config apt periodic dest etc apt apt conf d
1
82,161
3,603,658,295
IssuesEvent
2016-02-03 19:52:42
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
Make Kubernetes more bulletproof to gcloud failure
area/cluster-lifecycle priority/P1 team/control-plane team/gke
#20306 was one instance of this, but @brendandburns says there are multiple places in kube-up and elsewhere that we need to also subject to the same hygiene.
1.0
Make Kubernetes more bulletproof to gcloud failure - #20306 was one instance of this, but @brendandburns says there are multiple places in kube-up and elsewhere that we need to also subject to the same hygiene.
non_main
make kubernetes more bulletproof to gcloud failure was one instance of this but brendandburns says there are multiple places in kube up and elsewhere that we need to also subject to the same hygiene
0
426,249
12,369,822,939
IssuesEvent
2020-05-18 15:49:22
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
Add force IPV4 field to cloud view
Category: Sunstone Priority: Low Sponsored Status: Accepted Type: Feature
**Description** The force IPV4 input text doesn't appear in the cloud view. **Use case** Users can force an IPV4. **Interface Changes** Sunstone <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
Add force IPV4 field to cloud view - **Description** The force IPV4 input text doesn't appear in the cloud view. **Use case** Users can force an IPV4. **Interface Changes** Sunstone <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
non_main
add force field to cloud view description the force input text doesn t appear in the cloud view use case users can force an interface changes sunstone progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
0
51,293
7,694,084,490
IssuesEvent
2018-05-18 07:27:16
pombase/website
https://api.github.com/repos/pombase/website
closed
add a link to help section from
documentation gene page
Would it be useful to have a direct help link https://www.pombase.org/documentation/advanced-search from here https://www.pombase.org/query
1.0
add a link to help section from - Would it be useful to have a direct help link https://www.pombase.org/documentation/advanced-search from here https://www.pombase.org/query
non_main
add a link to help section from would it be useful to have a direct help link from here
0
829,414
31,878,881,502
IssuesEvent
2023-09-16 05:55:41
roq-trading/roq-issues
https://api.github.com/repos/roq-trading/roq-issues
closed
[roq-cme] TCP fragmentation could cause freeze
bug high priority support
iLink SBE parsing incorrectly assumed messages were received in full. The socket read code was based on MDP SBE parsing which is based on UDP where datagrams are received in full (or nothing at all). Fragmentation is not possible. However, iLink is based on TCP where messages are allowed to be fragmented (by the TCP stream protocol). Symptoms: * Gateway freeze Reason: Message parsing happens inside a loop that processes the read stream buffer. When the parser failed (due to message length) it would return 0 (bytes) to indicate that the read buffer did not yet have enough data. Unfortunately, the loop would only check on read buffer being fully processed by checking when the read buffer became empty. This resulted in an infinite loop. Solution: Break the loop when not enough data is ready to process the next message.
1.0
[roq-cme] TCP fragmentation could cause freeze - iLink SBE parsing incorrectly assumed messages were received in full. The socket read code was based on MDP SBE parsing which is based on UDP where datagrams are received in full (or nothing at all). Fragmentation is not possible. However, iLink is based on TCP where messages are allowed to be fragmented (by the TCP stream protocol). Symptoms: * Gateway freeze Reason: Message parsing happens inside a loop that processes the read stream buffer. When the parser failed (due to message length) it would return 0 (bytes) to indicate that the read buffer did not yet have enough data. Unfortunately, the loop would only check on read buffer being fully processed by checking when the read buffer became empty. This resulted in an infinite loop. Solution: Break the loop when not enough data is ready to process the next message.
non_main
tcp fragmentation could cause freeze ilink sbe parsing incorrectly assumed messages were received in full the socket read code was based on mdp sbe parsing which is based on udp where datagrams are received in full or nothing at all fragmentation is not possible however ilink is based on tcp where messages are allowed to be fragmented by the tcp stream protocol symptoms gateway freeze reason message parsing happens inside a loop that processes the read stream buffer when the parser failed due to message length it would return bytes to indicate that the read buffer did not yet have enough data unfortunately the loop would only check on read buffer being fully processed by checking when the read buffer became empty this resulted in an infinite loop solution break the loop when not enough data is ready to process the next message
0
778,159
27,305,560,064
IssuesEvent
2023-02-24 07:51:41
openforis/arena
https://api.github.com/repos/openforis/arena
closed
Copy/Clone records from cycle to another
Priority_1
We assume that schema in the source and target cycles are identical. If these are not identical, a direct cloning will be not possible.
1.0
Copy/Clone records from cycle to another - We assume that schema in the source and target cycles are identical. If these are not identical, a direct cloning will be not possible.
non_main
copy clone records from cycle to another we assume that schema in the source and target cycles are identical if these are not identical a direct cloning will be not possible
0
275,693
20,954,628,442
IssuesEvent
2022-03-27 00:20:15
numpy/numpy
https://api.github.com/repos/numpy/numpy
closed
DOC: Reference Gitpod as simple alternative in building from source guide
04 - Documentation sprint
### Issue with current documentation: I spent an hour trying and failing to get numpy to build from source. I followed the documentation but failed to discover that I could just use Gitpod. I only became aware of this as @mattip suggested it to me in a comment https://github.com/numpy/numpy/issues/20677#issuecomment-1002744924 ### Idea or request for content: The discoverability of the very helpful Gitpod setup could be improved by including a reference to it in this doc: https://numpy.org/doc/stable/dev/development_environment.html I overlooked it here, but spent most of the time in the other doc, that's maybe why I missed it. Since Gitpod works so well and is not necessarily something one would expect to exist, having more pointers to it could help.
1.0
DOC: Reference Gitpod as simple alternative in building from source guide - ### Issue with current documentation: I spent an hour trying and failing to get numpy to build from source. I followed the documentation but failed to discover that I could just use Gitpod. I only became aware of this as @mattip suggested it to me in a comment https://github.com/numpy/numpy/issues/20677#issuecomment-1002744924 ### Idea or request for content: The discoverability of the very helpful Gitpod setup could be improved by including a reference to it in this doc: https://numpy.org/doc/stable/dev/development_environment.html I overlooked it here, but spent most of the time in the other doc, that's maybe why I missed it. Since Gitpod works so well and is not necessarily something one would expect to exist, having more pointers to it could help.
non_main
doc reference gitpod as simple alternative in building from source guide issue with current documentation i spent an hour trying and failing to get numpy to build from source i followed the documentation but failed to discover that i could just use gitpod i only became aware of this as mattip suggested it to me in a comment idea or request for content the discoverability of the very helpful gitpod setup could be improved by including a reference to it in this doc i overlooked it here but spent most of the time in the other doc that s maybe why i missed it since gitpod works so well and is not necessarily something one would expect to exist having more pointers to it could help
0
281,658
8,697,938,874
IssuesEvent
2018-12-04 21:45:09
mplusmuseum/mplusmuseum-collections-explorer
https://api.github.com/repos/mplusmuseum/mplusmuseum-collections-explorer
closed
Search - feature: Inline Search Filter 'Pills'
priority-low question
Convert a comma-separated list into these 'pill' UI elements. Clicking the 'X' should clear them. @revdancatt do we know if this will be an "or" search or an "and" search? ![screen shot 2018-11-20 at 8 54 42 pm](https://user-images.githubusercontent.com/302309/48813856-93cbe500-ed06-11e8-8e18-430b77e54bee.png)
1.0
Search - feature: Inline Search Filter 'Pills' - Convert a comma-separated list into these 'pill' UI elements. Clicking the 'X' should clear them. @revdancatt do we know if this will be an "or" search or an "and" search? ![screen shot 2018-11-20 at 8 54 42 pm](https://user-images.githubusercontent.com/302309/48813856-93cbe500-ed06-11e8-8e18-430b77e54bee.png)
non_main
search feature inline search filter pills convert a comma separated list into these pill ui elements clicking the x should clear them revdancatt do we know if this will be an or search or an and search
0
3,860
17,025,697,920
IssuesEvent
2021-07-03 13:05:36
CyanCoding/Project-Tracker
https://api.github.com/repos/CyanCoding/Project-Tracker
closed
Change public readonly variables to resources
backend maintainability
In the code we have a lot of static resources (such as URLs, GUIDs, file locations, etc) that never change. These are used at various points around the program. It would be great if we could move from having a ton of global variables that are always passed as parameters and move to using these resources.
True
Change public readonly variables to resources - In the code we have a lot of static resources (such as URLs, GUIDs, file locations, etc) that never change. These are used at various points around the program. It would be great if we could move from having a ton of global variables that are always passed as parameters and move to using these resources.
main
change public readonly variables to resources in the code we have a lot of static resources such as urls guids file locations etc that never change these are used at various points around the program it would be great if we could move from having a ton of global variables that are always passed as parameters and move to using these resources
1
2,482
8,639,916,595
IssuesEvent
2018-11-23 22:41:10
F5OEO/rpitx
https://api.github.com/repos/F5OEO/rpitx
closed
RPITX : transmitting from TCP streaming
V1 related (not maintained)
Good morning, I'm facing a problem in transmitting using RPITX from TCP streaming (generated by gnuradio). The scenario is the following : a) Raspbian Stretch 4.14.44+ #1117 on RaspberryPI2B+ armv6l GNU/Linux b) rpitx Version 0.2 compiled Jun 4 2018 c) rpitx test script transmitting correctly on 10m : /rpitx -m IQ -i /home/pi/pat/ssbIQ.wav -f 28150 -l d) Gnuradio modulating the same sampleaudio.wav file in SSB / USB and sending it in complex format on TCP sink port (client mode). Waterfall shows modulated signal. e) data streaming received by netcat and sent to rpitx as follow : while true; do (nc -l 8011; dd if=/dev/zero bs=4096 count=30); done | sudo rpitx -i - -m IQ -f 28150 The result is that rpitx stays idle on the calibration phase and nothing is transmitted on air. rpitx Version 0.2 compiled Jun 4 2018 (F5OEO Evariste) running on Frequency=0.000000 SampleRate=48000 Master PLL = 1000000000 memory: 512 MB processor: Broadcom BCM2835 i2cDevice: /dev/i2c-1 model: Model B+ manufacturer: Sony pcb revision: 1 warranty void: no revision: 0010 peripheral base: 0x20000000 Jessie Using mbox device /dev/vcio. 3616000 Size NUM PAGES 883 PAGE_SIZE 4096 MASH 3 Freq PLL# 6 Calibrate : ppm=-10 DMA 160ns:1361ns ... and no more. What I'm doing wrong ? Any help appreciated Thank-you Best Regards, 73 Ugo Poddine IU1IPB
True
RPITX : transmitting from TCP streaming - Good morning, I'm facing a problem in transmitting using RPITX from TCP streaming (generated by gnuradio). The scenario is the following : a) Raspbian Stretch 4.14.44+ #1117 on RaspberryPI2B+ armv6l GNU/Linux b) rpitx Version 0.2 compiled Jun 4 2018 c) rpitx test script transmitting correctly on 10m : /rpitx -m IQ -i /home/pi/pat/ssbIQ.wav -f 28150 -l d) Gnuradio modulating the same sampleaudio.wav file in SSB / USB and sending it in complex format on TCP sink port (client mode). Waterfall shows modulated signal. e) data streaming received by netcat and sent to rpitx as follow : while true; do (nc -l 8011; dd if=/dev/zero bs=4096 count=30); done | sudo rpitx -i - -m IQ -f 28150 The result is that rpitx stays idle on the calibration phase and nothing is transmitted on air. rpitx Version 0.2 compiled Jun 4 2018 (F5OEO Evariste) running on Frequency=0.000000 SampleRate=48000 Master PLL = 1000000000 memory: 512 MB processor: Broadcom BCM2835 i2cDevice: /dev/i2c-1 model: Model B+ manufacturer: Sony pcb revision: 1 warranty void: no revision: 0010 peripheral base: 0x20000000 Jessie Using mbox device /dev/vcio. 3616000 Size NUM PAGES 883 PAGE_SIZE 4096 MASH 3 Freq PLL# 6 Calibrate : ppm=-10 DMA 160ns:1361ns ... and no more. What I'm doing wrong ? Any help appreciated Thank-you Best Regards, 73 Ugo Poddine IU1IPB
main
rpitx transmitting from tcp streaming good morning i m facing a problem in transmitting using rpitx from tcp streaming generated by gnuradio the scenario is the following a raspbian stretch on gnu linux b rpitx version compiled jun c rpitx test script transmitting correctly on rpitx m iq i home pi pat ssbiq wav f l d gnuradio modulating the same sampleaudio wav file in ssb usb and sending it in complex format on tcp sink port client mode waterfall shows modulated signal e data streaming received by netcat and sent to rpitx as follow while true do nc l dd if dev zero bs count done sudo rpitx i m iq f the result is that rpitx stays idle on the calibration phase and nothing is transmitted on air rpitx version compiled jun evariste running on frequency samplerate master pll memory mb processor broadcom dev model model b manufacturer sony pcb revision warranty void no revision peripheral base jessie using mbox device dev vcio size num pages page size mash freq pll calibrate ppm dma and no more what i m doing wrong any help appreciated thank you best regards ugo poddine
1
3,600
14,539,876,498
IssuesEvent
2020-12-15 12:31:49
pace/bricks
https://api.github.com/repos/pace/bricks
opened
Revisit ordering of default RoundTripper Chain
S::Ready T::Maintainance
The default RoundTripper Chain ordering might not make sense right now, because certain more "important" roundtrippers are executed later in the stack, e.g., request id roundtripper, and other potentially too early, e.g. dump roundtripper.
True
Revisit ordering of default RoundTripper Chain - The default RoundTripper Chain ordering might not make sense right now, because certain more "important" roundtrippers are executed later in the stack, e.g., request id roundtripper, and other potentially too early, e.g. dump roundtripper.
main
revisit ordering of default roundtripper chain the default roundtripper chain ordering might not make sense right now because certain more important roundtrippers are executed later in the stack e g request id roundtripper and other potentially too early e g dump roundtripper
1
4,602
23,849,803,478
IssuesEvent
2022-09-06 16:49:56
ocsf/ocsf-schema
https://api.github.com/repos/ocsf/ocsf-schema
closed
All first release objects merged into main
maintainers
First release Objects, checkbox checked when merged into main. - [ ] API Details - [ ] Attack - [ ] Authorization Information - [ ] CVSS Scores - [ ] Client - [ ] Cloud - [ ] Compliance Details - [ ] Container - [ ] DNS Answer - [ ] DNS Query - [ ] Device - [ ] Device Hardware Info - [ ] Digital Certificate - [ ] Digital Signature - [ ] Display - [ ] Distributed Computing Environment/Remote Procedure Call (DCE/RPC) - [ ] Endpoint - [ ] Enrichment - [ ] Event Origin - [ ] Event Source - [ ] Extended User - [ ] Feature - [ ] File - [ ] Finding Details - [ ] Fingerprint - [ ] Geo Location - [ ] Group - [ ] HTTP Cookie - [ ] HTTP Header - [ ] HTTP Request - [ ] HTTP Response - [ ] Identity - [ ] Identity Provider - [ ] Image - [ ] Job - [ ] Kernel Resource - [ ] Keyboard Information - [ ] Malware - [ ] Managed Entity - [ ] Metadata - [ ] Module - [ ] Network Connection Information - [ ] Network Endpoint - [ ] Network Interface - [ ] Network Proxy - [ ] Network Traffic - [ ] OS - [ ] Object - [ ] Observable - [ ] Peripheral Device - [ ] Process - [ ] Product - [ ] Registry Key - [ ] Registry Value - [ ] Related Findings - [ ] Remediation Details - [ ] Request Elements - [ ] Resource - [ ] Response Elements - [ ] Rule - [ ] Server - [ ] Service - [ ] Session - [ ] Subject Alternative Name - [ ] TLS Extension - [ ] Transport Layer Security (TLS) - [ ] Uniform Resource Locator - [ ] User - [ ] Vulnerability Details
True
All first release objects merged into main - First release Objects, checkbox checked when merged into main. - [ ] API Details - [ ] Attack - [ ] Authorization Information - [ ] CVSS Scores - [ ] Client - [ ] Cloud - [ ] Compliance Details - [ ] Container - [ ] DNS Answer - [ ] DNS Query - [ ] Device - [ ] Device Hardware Info - [ ] Digital Certificate - [ ] Digital Signature - [ ] Display - [ ] Distributed Computing Environment/Remote Procedure Call (DCE/RPC) - [ ] Endpoint - [ ] Enrichment - [ ] Event Origin - [ ] Event Source - [ ] Extended User - [ ] Feature - [ ] File - [ ] Finding Details - [ ] Fingerprint - [ ] Geo Location - [ ] Group - [ ] HTTP Cookie - [ ] HTTP Header - [ ] HTTP Request - [ ] HTTP Response - [ ] Identity - [ ] Identity Provider - [ ] Image - [ ] Job - [ ] Kernel Resource - [ ] Keyboard Information - [ ] Malware - [ ] Managed Entity - [ ] Metadata - [ ] Module - [ ] Network Connection Information - [ ] Network Endpoint - [ ] Network Interface - [ ] Network Proxy - [ ] Network Traffic - [ ] OS - [ ] Object - [ ] Observable - [ ] Peripheral Device - [ ] Process - [ ] Product - [ ] Registry Key - [ ] Registry Value - [ ] Related Findings - [ ] Remediation Details - [ ] Request Elements - [ ] Resource - [ ] Response Elements - [ ] Rule - [ ] Server - [ ] Service - [ ] Session - [ ] Subject Alternative Name - [ ] TLS Extension - [ ] Transport Layer Security (TLS) - [ ] Uniform Resource Locator - [ ] User - [ ] Vulnerability Details
main
all first release objects merged into main first release objects checkbox checked when merged into main api details attack authorization information cvss scores client cloud compliance details container dns answer dns query device device hardware info digital certificate digital signature display distributed computing environment remote procedure call dce rpc endpoint enrichment event origin event source extended user feature file finding details fingerprint geo location group http cookie http header http request http response identity identity provider image job kernel resource keyboard information malware managed entity metadata module network connection information network endpoint network interface network proxy network traffic os object observable peripheral device process product registry key registry value related findings remediation details request elements resource response elements rule server service session subject alternative name tls extension transport layer security tls uniform resource locator user vulnerability details
1
231,861
25,552,040,897
IssuesEvent
2022-11-30 01:08:59
npenin/thewheel-dotnet
https://api.github.com/repos/npenin/thewheel-dotnet
reopened
newtonsoft.json.9.0.1.nupkg: 1 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>newtonsoft.json.9.0.1.nupkg</b></p></summary> <p>Json.NET is a popular high-performance JSON framework for .NET</p> <p>Library home page: <a href="https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg">https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg</a></p> <p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg,/ckages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg</p> <p> <p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/8d3ffe3535dffc84b035cd193db429b0d8b00401">8d3ffe3535dffc84b035cd193db429b0d8b00401</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (newtonsoft.json.9.0.1.nupkg version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [WS-2022-0161](https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | newtonsoft.json.9.0.1.nupkg | Direct | Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2022-0161</summary> ### Vulnerable Library - <b>newtonsoft.json.9.0.1.nupkg</b></p> <p>Json.NET is a popular high-performance JSON framework for .NET</p> <p>Library home page: <a href="https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg">https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg</a></p> <p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg,/ckages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg</p> <p> Dependency Hierarchy: - :x: **newtonsoft.json.9.0.1.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/8d3ffe3535dffc84b035cd193db429b0d8b00401">8d3ffe3535dffc84b035cd193db429b0d8b00401</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Improper Handling of Exceptional Conditions in Newtonsoft.Json. Newtonsoft.Json prior to version 13.0.1 is vulnerable to Insecure Defaults due to improper handling of StackOverFlow exception (SOE) whenever nested expressions are being processed. Exploiting this vulnerability results in Denial Of Service (DoS), and it is exploitable when an attacker sends 5 requests that cause SOE in time frame of 5 minutes. This vulnerability affects Internet Information Services (IIS) Applications. <p>Publish Date: 2022-06-22 <p>URL: <a href=https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66>WS-2022-0161</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-06-22</p> <p>Fix Resolution: Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
newtonsoft.json.9.0.1.nupkg: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>newtonsoft.json.9.0.1.nupkg</b></p></summary> <p>Json.NET is a popular high-performance JSON framework for .NET</p> <p>Library home page: <a href="https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg">https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg</a></p> <p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg,/ckages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg</p> <p> <p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/8d3ffe3535dffc84b035cd193db429b0d8b00401">8d3ffe3535dffc84b035cd193db429b0d8b00401</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (newtonsoft.json.9.0.1.nupkg version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [WS-2022-0161](https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | newtonsoft.json.9.0.1.nupkg | Direct | Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2022-0161</summary> ### Vulnerable Library - <b>newtonsoft.json.9.0.1.nupkg</b></p> <p>Json.NET is a popular high-performance JSON framework for .NET</p> <p>Library home page: <a href="https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg">https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg</a></p> <p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg,/ckages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg</p> <p> Dependency Hierarchy: - :x: **newtonsoft.json.9.0.1.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/8d3ffe3535dffc84b035cd193db429b0d8b00401">8d3ffe3535dffc84b035cd193db429b0d8b00401</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Improper Handling of Exceptional Conditions in Newtonsoft.Json. Newtonsoft.Json prior to version 13.0.1 is vulnerable to Insecure Defaults due to improper handling of StackOverFlow exception (SOE) whenever nested expressions are being processed. Exploiting this vulnerability results in Denial Of Service (DoS), and it is exploitable when an attacker sends 5 requests that cause SOE in time frame of 5 minutes. This vulnerability affects Internet Information Services (IIS) Applications. <p>Publish Date: 2022-06-22 <p>URL: <a href=https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66>WS-2022-0161</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-06-22</p> <p>Fix Resolution: Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_main
newtonsoft json nupkg vulnerabilities highest severity is vulnerable library newtonsoft json nupkg json net is a popular high performance json framework for net library home page a href path to dependency file thewheel tests thewheel tests csproj path to vulnerable library home wss scanner nuget packages newtonsoft json newtonsoft json nupkg ckages newtonsoft json newtonsoft json nupkg found in head commit a href vulnerabilities cve severity cvss dependency type fixed in newtonsoft json nupkg version remediation available high newtonsoft json nupkg direct newtonsoft json microsoft extensions apidescription server details ws vulnerable library newtonsoft json nupkg json net is a popular high performance json framework for net library home page a href path to dependency file thewheel tests thewheel tests csproj path to vulnerable library home wss scanner nuget packages newtonsoft json newtonsoft json nupkg ckages newtonsoft json newtonsoft json nupkg dependency hierarchy x newtonsoft json nupkg vulnerable library found in head commit a href found in base branch main vulnerability details improper handling of exceptional conditions in newtonsoft json newtonsoft json prior to version is vulnerable to insecure defaults due to improper handling of stackoverflow exception soe whenever nested expressions are being processed exploiting this vulnerability results in denial of service dos and it is exploitable when an attacker sends requests that cause soe in time frame of minutes this vulnerability affects internet information services iis applications publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution newtonsoft json microsoft extensions apidescription server step up your open source security game with mend
0
56,661
15,266,678,273
IssuesEvent
2021-02-22 09:07:11
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
DataScroller with SelectBooleanCheckbox: Initial checked value is overwritten for loaded chunks during first decode
defect
**Describe the defect** If I create a datascroller with a list of items, that have a boolean value that is initially true and bound to a p:selectBooleanCheckbox within the datascroller, the checkbox is checked for entries that are initially dispayed (ok) but for additionally loaded elements due to scrolling the checkbox is not checked anymore. **Reproducer** [primefaces-test.zip](https://github.com/primefaces/primefaces/files/6003830/primefaces-test.zip) **Environment:** - PF Version: _10.0.0.RC1_ - JSF + version: Mojarra 2.2.8 - Affected browsers: ALL **To Reproduce** Steps to reproduce the behavior: 1. Go to test page in attached primefaces-test 2. Scroll down until additional entries are loaded 4. See error (additional entry index>5 have an unchecked checkbox) **Expected behavior** All checkboxes should be checked according to data in Model. Possible Reason for that: The problem is related to the fact that in **SelectBooleanCheckboxRenderer.decode** and submittedValue of null is interpreted as "false" and set on the underlying model. For lazy loaded chunk this seems to be executed during the ajax loadChunks call and since the new elements were just yet added their is not value in the requestparameter map for them. Other Input Forms (e.g inputtext) for fine because their just skip the updatemodel phase if submited value == null. **Example XHTML** ```html <h:form id="frmTest"> <p:dataScroller var="_item" value="#{testView.model}" chunkSize="5" > <div style="height:300px"> <p:inputText value="#{_item.text}" /> <p:selectBooleanCheckbox value="#{_item.check}"/> This checkbox should always be checked, but as soon as new chunk is loaded it is not. </div> </p:dataScroller> </h:form> ``` **Example Bean** ```java @Data @Named @ViewScoped public class TestView implements Serializable { private static final long serialVersionUID = 1L; private ListDataModel<Item> model = new ListDataModel<>(); @Data public static class Item implements Serializable { private static final long serialVersionUID = 1L; private String text; private boolean check=true; public Item(String text) { super(); this.text = text; } } @PostConstruct public void init() { List<Item> list = new ArrayList<>(); for (int i =0; i<50; i++) { list.add(new Item("Text "+i)); } model = new ListDataModel<>(list); } public ListDataModel<Item> getModel() { return model; } } ```
1.0
DataScroller with SelectBooleanCheckbox: Initial checked value is overwritten for loaded chunks during first decode - **Describe the defect** If I create a datascroller with a list of items, that have a boolean value that is initially true and bound to a p:selectBooleanCheckbox within the datascroller, the checkbox is checked for entries that are initially dispayed (ok) but for additionally loaded elements due to scrolling the checkbox is not checked anymore. **Reproducer** [primefaces-test.zip](https://github.com/primefaces/primefaces/files/6003830/primefaces-test.zip) **Environment:** - PF Version: _10.0.0.RC1_ - JSF + version: Mojarra 2.2.8 - Affected browsers: ALL **To Reproduce** Steps to reproduce the behavior: 1. Go to test page in attached primefaces-test 2. Scroll down until additional entries are loaded 4. See error (additional entry index>5 have an unchecked checkbox) **Expected behavior** All checkboxes should be checked according to data in Model. Possible Reason for that: The problem is related to the fact that in **SelectBooleanCheckboxRenderer.decode** and submittedValue of null is interpreted as "false" and set on the underlying model. For lazy loaded chunk this seems to be executed during the ajax loadChunks call and since the new elements were just yet added their is not value in the requestparameter map for them. Other Input Forms (e.g inputtext) for fine because their just skip the updatemodel phase if submited value == null. **Example XHTML** ```html <h:form id="frmTest"> <p:dataScroller var="_item" value="#{testView.model}" chunkSize="5" > <div style="height:300px"> <p:inputText value="#{_item.text}" /> <p:selectBooleanCheckbox value="#{_item.check}"/> This checkbox should always be checked, but as soon as new chunk is loaded it is not. </div> </p:dataScroller> </h:form> ``` **Example Bean** ```java @Data @Named @ViewScoped public class TestView implements Serializable { private static final long serialVersionUID = 1L; private ListDataModel<Item> model = new ListDataModel<>(); @Data public static class Item implements Serializable { private static final long serialVersionUID = 1L; private String text; private boolean check=true; public Item(String text) { super(); this.text = text; } } @PostConstruct public void init() { List<Item> list = new ArrayList<>(); for (int i =0; i<50; i++) { list.add(new Item("Text "+i)); } model = new ListDataModel<>(list); } public ListDataModel<Item> getModel() { return model; } } ```
non_main
datascroller with selectbooleancheckbox initial checked value is overwritten for loaded chunks during first decode describe the defect if i create a datascroller with a list of items that have a boolean value that is initially true and bound to a p selectbooleancheckbox within the datascroller the checkbox is checked for entries that are initially dispayed ok but for additionally loaded elements due to scrolling the checkbox is not checked anymore reproducer environment pf version jsf version mojarra affected browsers all to reproduce steps to reproduce the behavior go to test page in attached primefaces test scroll down until additional entries are loaded see error additional entry index have an unchecked checkbox expected behavior all checkboxes should be checked according to data in model possible reason for that the problem is related to the fact that in selectbooleancheckboxrenderer decode and submittedvalue of null is interpreted as false and set on the underlying model for lazy loaded chunk this seems to be executed during the ajax loadchunks call and since the new elements were just yet added their is not value in the requestparameter map for them other input forms e g inputtext for fine because their just skip the updatemodel phase if submited value null example xhtml html this checkbox should always be checked but as soon as new chunk is loaded it is not example bean java data named viewscoped public class testview implements serializable private static final long serialversionuid private listdatamodel model new listdatamodel data public static class item implements serializable private static final long serialversionuid private string text private boolean check true public item string text super this text text postconstruct public void init list list new arraylist for int i i i list add new item text i model new listdatamodel list public listdatamodel getmodel return model
0
117,120
11,945,718,945
IssuesEvent
2020-04-03 06:34:08
heidicrq/ped
https://api.github.com/repos/heidicrq/ped
opened
Adding a task - contradiction between the UG and the error message
severity.Medium type.DocumentationBug
Can’t add task without deadline even though deadline is said to be optional in the UG. However error message suggests to the user that deadline is compulsory. ![image.png](https://raw.githubusercontent.com/heidicrq/ped/master/files/16adff33-3684-4165-8417-ab553ef7e6f6.png)
1.0
Adding a task - contradiction between the UG and the error message - Can’t add task without deadline even though deadline is said to be optional in the UG. However error message suggests to the user that deadline is compulsory. ![image.png](https://raw.githubusercontent.com/heidicrq/ped/master/files/16adff33-3684-4165-8417-ab553ef7e6f6.png)
non_main
adding a task contradiction between the ug and the error message can’t add task without deadline even though deadline is said to be optional in the ug however error message suggests to the user that deadline is compulsory
0
679,139
23,222,209,359
IssuesEvent
2022-08-02 19:23:44
dominicm00/ham
https://api.github.com/repos/dominicm00/ham
closed
Refactor memory and lifetime management
high priority large tracker refactor
Ham data classes derive from `Referenceable` as a replacement to smart pointers, but have a critical flaw; `Referenceable` objects cannot be const if their reference count is to be modified. This greatly limits the use of const throughout the codebase despite the fact that most Ham data is immutable after parsing, and causes difficult to resolve conflicts between const and non-const interfaces. To resolve this, we should move to a more standard memory and lifetime management system, including: - Using `shared_ptr` instead of `Referenceable` - Removing manual memory management (e.g. the custom String classes) - Throwing exceptions when constructors fail PRs: - [x] #66 - [ ] #67
1.0
Refactor memory and lifetime management - Ham data classes derive from `Referenceable` as a replacement to smart pointers, but have a critical flaw; `Referenceable` objects cannot be const if their reference count is to be modified. This greatly limits the use of const throughout the codebase despite the fact that most Ham data is immutable after parsing, and causes difficult to resolve conflicts between const and non-const interfaces. To resolve this, we should move to a more standard memory and lifetime management system, including: - Using `shared_ptr` instead of `Referenceable` - Removing manual memory management (e.g. the custom String classes) - Throwing exceptions when constructors fail PRs: - [x] #66 - [ ] #67
non_main
refactor memory and lifetime management ham data classes derive from referenceable as a replacement to smart pointers but have a critical flaw referenceable objects cannot be const if their reference count is to be modified this greatly limits the use of const throughout the codebase despite the fact that most ham data is immutable after parsing and causes difficult to resolve conflicts between const and non const interfaces to resolve this we should move to a more standard memory and lifetime management system including using shared ptr instead of referenceable removing manual memory management e g the custom string classes throwing exceptions when constructors fail prs
0