Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
10,728 | 8,697,726,229 | IssuesEvent | 2018-12-04 21:07:59 | conversationai/conversationai-models | https://api.github.com/repos/conversationai/conversationai-models | closed | Write subset of predictions to HTML file | infrastructure | Reading CSV files is tough, but it's often useful to look through the test data and predictions beyond just looking at the accuracy metrics. One solution is to write a sample of the predictions in a HTML format that we can add some basic styling to so it's easy to read. That way we can go from new model -> analyzing results really quickly.
| 1.0 | Write subset of predictions to HTML file - Reading CSV files is tough, but it's often useful to look through the test data and predictions beyond just looking at the accuracy metrics. One solution is to write a sample of the predictions in a HTML format that we can add some basic styling to so it's easy to read. That way we can go from new model -> analyzing results really quickly.
| non_main | write subset of predictions to html file reading csv files is tough but it s often useful to look through the test data and predictions beyond just looking at the accuracy metrics one solution is to write a sample of the predictions in a html format that we can add some basic styling to so it s easy to read that way we can go from new model analyzing results really quickly | 0 |
5,056 | 25,892,470,280 | IssuesEvent | 2022-12-14 19:10:44 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | SortFieldNotFound error when linking to a newly-created table | type: bug work: backend status: ready restricted: maintainers | ## Steps to reproduce
1. Begin with our Library Schema and the code in my PR #2080.
1. From the schema page, choose "New Table" > "From Scratch".
1. When the Table Page loads, rename the table to "Reviews".
1. Note that we get an error when loading the records. That error is reported in #2089. I'm not sure if it's related to the error I'm reporting now.
1. Create a new "Many to One" link to the Publications table. (Note that since recording the video below, the "Many to One" option has been relocated to the first option.)
1. Attempt to add a new record to "Reviews".
1. Observe the records POST request to fail with the following error
<details>
<summary>Traceback</summary>
```
Environment:
Request Method: POST
Request URL: http://localhost:8000/api/db/v0/tables/89/records/
Django Version: 3.1.14
Python Version: 3.9.16
Installed Applications:
['django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'django_filters',
'django_property_filter',
'mathesar']
Installed Middleware:
['django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
'mathesar.middleware.CursorClosedHandlerMiddleware',
'mathesar.middleware.PasswordChangeNeededMiddleware',
'django_userforeignkey.middleware.UserForeignKeyMiddleware',
'django_request_cache.middleware.RequestCacheMiddleware']
Traceback (most recent call last):
File "/code/db/columns/utils.py", line 70, in find_column_by_name_in_relation
return relation.columns[col_name_string]
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1158, in __getitem__
return self._index[key]
During handling of the above exception (Column('439__323__col__323', INTEGER(), table=<anon_1>, primary_key=True, nullable=False)), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 826, in __getattr__
return getattr(self.comparator, key)
The above exception ('Comparator' object has no attribute 'split') was the direct cause of the following exception:
File "/code/db/columns/utils.py", line 54, in get_column_obj_from_relation
column = find_column_by_name_in_relation(relation, column)
File "/code/db/columns/utils.py", line 72, in find_column_by_name_in_relation
col_name_split = col_name_string.split(sep='.', maxsplit=1)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 828, in __getattr__
util.raise_(
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
During handling of the above exception (Neither 'Column' object nor 'Comparator' object has an attribute 'split'), another exception occurred:
File "/code/db/records/operations/sort.py", line 57, in _get_sorted_column_obj_from_spec
column = col_utils.get_column_obj_from_relation(relation, sort_spec.field)
File "/code/db/columns/utils.py", line 56, in get_column_obj_from_relation
column = relation.columns[column.name]
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1158, in __getitem__
return self._index[key]
During handling of the above exception ('439__323__col__323'), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view
return view_func(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view
return self.dispatch(request, *args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch
response = self.handle_exception(exc)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception
response = exception_handler(exc, context)
File "/code/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler
raise exc
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch
response = handler(request, *args, **kwargs)
File "/code/mathesar/api/db/viewsets/records.py", line 155, in create
records = paginator.paginate_queryset(
File "/code/mathesar/api/pagination.py", line 89, in paginate_queryset
records = query.get_records(
File "/code/mathesar/models/query.py", line 172, in get_records
return self.db_query.get_records(**kwargs)
File "/code/db/queries/base.py", line 108, in get_records
return records_select.get_records_with_default_order(
File "/code/db/records/operations/select.py", line 28, in get_records_with_default_order
return get_records(
File "/code/db/records/operations/select.py", line 70, in get_records
relation = apply_transformations_deprecated(
File "/code/db/transforms/operations/apply.py", line 55, in apply_transformations_deprecated
relation = apply_transformations(relation, transforms)
File "/code/db/transforms/operations/apply.py", line 8, in apply_transformations
relation = _apply_transform(relation, transform)
File "/code/db/transforms/operations/apply.py", line 14, in _apply_transform
relation = transform.apply_to_relation(relation)
File "/code/db/transforms/base.py", line 124, in apply_to_relation
executable = rec_sort.apply_relation_sorting(relation, order_by)
File "/code/db/records/operations/sort.py", line 44, in apply_relation_sorting
order_by_list = [
File "/code/db/records/operations/sort.py", line 45, in <listcomp>
_get_sorted_column_obj_from_spec(relation, spec) for spec in sort_spec
File "/code/db/records/operations/sort.py", line 59, in _get_sorted_column_obj_from_spec
raise SortFieldNotFound(e)
Exception Type: SortFieldNotFound at /api/db/v0/tables/89/records/
Exception Value: '439__323__col__323'
```
</details>
https://user-images.githubusercontent.com/42411/207400140-bc4fdd90-6655-4213-95ce-a5cb111f6749.mp4
| True | SortFieldNotFound error when linking to a newly-created table - ## Steps to reproduce
1. Begin with our Library Schema and the code in my PR #2080.
1. From the schema page, choose "New Table" > "From Scratch".
1. When the Table Page loads, rename the table to "Reviews".
1. Note that we get an error when loading the records. That error is reported in #2089. I'm not sure if it's related to the error I'm reporting now.
1. Create a new "Many to One" link to the Publications table. (Note that since recording the video below, the "Many to One" option has been relocated to the first option.)
1. Attempt to add a new record to "Reviews".
1. Observe the records POST request to fail with the following error
<details>
<summary>Traceback</summary>
```
Environment:
Request Method: POST
Request URL: http://localhost:8000/api/db/v0/tables/89/records/
Django Version: 3.1.14
Python Version: 3.9.16
Installed Applications:
['django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'django_filters',
'django_property_filter',
'mathesar']
Installed Middleware:
['django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
'mathesar.middleware.CursorClosedHandlerMiddleware',
'mathesar.middleware.PasswordChangeNeededMiddleware',
'django_userforeignkey.middleware.UserForeignKeyMiddleware',
'django_request_cache.middleware.RequestCacheMiddleware']
Traceback (most recent call last):
File "/code/db/columns/utils.py", line 70, in find_column_by_name_in_relation
return relation.columns[col_name_string]
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1158, in __getitem__
return self._index[key]
During handling of the above exception (Column('439__323__col__323', INTEGER(), table=<anon_1>, primary_key=True, nullable=False)), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 826, in __getattr__
return getattr(self.comparator, key)
The above exception ('Comparator' object has no attribute 'split') was the direct cause of the following exception:
File "/code/db/columns/utils.py", line 54, in get_column_obj_from_relation
column = find_column_by_name_in_relation(relation, column)
File "/code/db/columns/utils.py", line 72, in find_column_by_name_in_relation
col_name_split = col_name_string.split(sep='.', maxsplit=1)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 828, in __getattr__
util.raise_(
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
During handling of the above exception (Neither 'Column' object nor 'Comparator' object has an attribute 'split'), another exception occurred:
File "/code/db/records/operations/sort.py", line 57, in _get_sorted_column_obj_from_spec
column = col_utils.get_column_obj_from_relation(relation, sort_spec.field)
File "/code/db/columns/utils.py", line 56, in get_column_obj_from_relation
column = relation.columns[column.name]
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1158, in __getitem__
return self._index[key]
During handling of the above exception ('439__323__col__323'), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view
return view_func(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view
return self.dispatch(request, *args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch
response = self.handle_exception(exc)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception
response = exception_handler(exc, context)
File "/code/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler
raise exc
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch
response = handler(request, *args, **kwargs)
File "/code/mathesar/api/db/viewsets/records.py", line 155, in create
records = paginator.paginate_queryset(
File "/code/mathesar/api/pagination.py", line 89, in paginate_queryset
records = query.get_records(
File "/code/mathesar/models/query.py", line 172, in get_records
return self.db_query.get_records(**kwargs)
File "/code/db/queries/base.py", line 108, in get_records
return records_select.get_records_with_default_order(
File "/code/db/records/operations/select.py", line 28, in get_records_with_default_order
return get_records(
File "/code/db/records/operations/select.py", line 70, in get_records
relation = apply_transformations_deprecated(
File "/code/db/transforms/operations/apply.py", line 55, in apply_transformations_deprecated
relation = apply_transformations(relation, transforms)
File "/code/db/transforms/operations/apply.py", line 8, in apply_transformations
relation = _apply_transform(relation, transform)
File "/code/db/transforms/operations/apply.py", line 14, in _apply_transform
relation = transform.apply_to_relation(relation)
File "/code/db/transforms/base.py", line 124, in apply_to_relation
executable = rec_sort.apply_relation_sorting(relation, order_by)
File "/code/db/records/operations/sort.py", line 44, in apply_relation_sorting
order_by_list = [
File "/code/db/records/operations/sort.py", line 45, in <listcomp>
_get_sorted_column_obj_from_spec(relation, spec) for spec in sort_spec
File "/code/db/records/operations/sort.py", line 59, in _get_sorted_column_obj_from_spec
raise SortFieldNotFound(e)
Exception Type: SortFieldNotFound at /api/db/v0/tables/89/records/
Exception Value: '439__323__col__323'
```
</details>
https://user-images.githubusercontent.com/42411/207400140-bc4fdd90-6655-4213-95ce-a5cb111f6749.mp4
| main | sortfieldnotfound error when linking to a newly created table steps to reproduce begin with our library schema and the code in my pr from the schema page choose new table from scratch when the table page loads rename the table to reviews note that we get an error when loading the records that error is reported in i m not sure if it s related to the error i m reporting now create a new many to one link to the publications table note that since recording the video below the many to one option has been relocated to the first option attempt to add a new record to reviews observe the records post request to fail with the following error traceback environment request method post request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware mathesar middleware cursorclosedhandlermiddleware mathesar middleware passwordchangeneededmiddleware django userforeignkey middleware userforeignkeymiddleware django request cache middleware requestcachemiddleware traceback most recent call last file code db columns utils py line in find column by name in relation return relation columns file usr local lib site packages sqlalchemy sql base py line in getitem return self index during handling of the above exception column col integer table primary key true nullable false another exception occurred file usr local lib site packages sqlalchemy sql elements py line in getattr return getattr self comparator key the above exception comparator object has no attribute split was the direct cause of the following exception file code db columns utils py line in get column obj from relation column find column by name in relation relation column file code db columns utils py line in find column by name in relation col name split col name string split sep maxsplit file usr local lib site packages sqlalchemy sql elements py line in getattr util raise file usr local lib site packages sqlalchemy util compat py line in raise raise exception during handling of the above exception neither column object nor comparator object has an attribute split another exception occurred file code db records operations sort py line in get sorted column obj from spec column col utils get column obj from relation relation sort spec field file code db columns utils py line in get column obj from relation column relation columns file usr local lib site packages sqlalchemy sql base py line in getitem return self index during handling of the above exception col another exception occurred file usr local lib site packages django core handlers exception py line in inner response get response request file usr local lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr local lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file usr local lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file usr local lib site packages rest framework views py line in dispatch response self handle exception exc file usr local lib site packages rest framework views py line in handle exception response exception handler exc context file code mathesar exception handlers py line in mathesar exception handler raise exc file usr local lib site packages rest framework views py line in dispatch response handler request args kwargs file code mathesar api db viewsets records py line in create records paginator paginate queryset file code mathesar api pagination py line in paginate queryset records query get records file code mathesar models query py line in get records return self db query get records kwargs file code db queries base py line in get records return records select get records with default order file code db records operations select py line in get records with default order return get records file code db records operations select py line in get records relation apply transformations deprecated file code db transforms operations apply py line in apply transformations deprecated relation apply transformations relation transforms file code db transforms operations apply py line in apply transformations relation apply transform relation transform file code db transforms operations apply py line in apply transform relation transform apply to relation relation file code db transforms base py line in apply to relation executable rec sort apply relation sorting relation order by file code db records operations sort py line in apply relation sorting order by list file code db records operations sort py line in get sorted column obj from spec relation spec for spec in sort spec file code db records operations sort py line in get sorted column obj from spec raise sortfieldnotfound e exception type sortfieldnotfound at api db tables records exception value col | 1 |
82,034 | 23,649,117,146 | IssuesEvent | 2022-08-26 03:42:23 | curl/curl | https://api.github.com/repos/curl/curl | closed | Cannot build curl statically and use static dependencies - Mac OS X | build not-a-bug | Hi there,
I'm trying to build libcurl on Apple Mac OS X 11.
I'm doing a minimal build with HTTPS and HTTP/2.
I have built out openssl, zlib, and nghttp2 as static libraries -> libssl.a, libcrypto.a, libz.a. The files exist in that path as needed and the configure script picks them up correctly.
However, despite passing --disable-shared, the configure script errors out with:
`configure: error: one or more libs available at link-time are not available run-time. Libs used at link-time: -lnghttp2 -lssl -lcrypto -lz`
And of course the .dylibs for those don't exist - they're static. I'm trying to do a fully 100% independent build of curl that has no reliance on any other dependencies - except for mac os x frameworks. I want my static openssl, my static nghttp2, and my static zlib.
Here is my full configure command:
`./configure --disable-shared --with-secure-transport --without-librtmp --without-libidn2 --with-zlib=/Users/jammerxd/Desktop/Git/cplusplusdependencies/zlib-1.2.11/MacOSX/x86_64/ --with-openssl=/Users/jammerxd/Desktop/Git/cplusplusdependencies/openssl-1.1.1m/MacOSX/x86_64 --disable-ldaps --disable-ntlm --disable-telnet --disable-proxy --disable-debug --enable-optimize --disable-curldebug --enable-symbol-hiding --disable-ftp --disable-rtsp --disable-dict --disable-tftp --disable-pop3 --disable-imap --disable-smb --disable-smtp --disable-gopher --disable-mqtt --disable-ntlm-wb --disable-tls-srp --disable-http-auth --disable-doh --disable-netrc --enable-progress-meter --disable-alt-svc --disable-ldap --disable-file --disable-dependency-tracking --with-nghttp2=/Users/jammerxd/Desktop/Git/cplusplusdependencies/nghttp2-1.49.0/MacOSX/x86_64/`
If I try to prefix the above command with:
`LDFLAGS="-static" PKG_CONFIG="pkg-config --static"`
the configure tool errors out with:
```
checking whether the C compiler works... no
configure: error: in `/Users/jammerxd/Desktop/Git/cplusplusdependencies/curl-7.84.0':
configure: error: C compiler cannot create executables
See `config.log' for more details
```
Looking on how I'm supposed to be able to build curl statically then have it link to the dependencies statically as well.
Thanks,
jammerxd | 1.0 | Cannot build curl statically and use static dependencies - Mac OS X - Hi there,
I'm trying to build libcurl on Apple Mac OS X 11.
I'm doing a minimal build with HTTPS and HTTP/2.
I have built out openssl, zlib, and nghttp2 as static libraries -> libssl.a, libcrypto.a, libz.a. The files exist in that path as needed and the configure script picks them up correctly.
However, despite passing --disable-shared, the configure script errors out with:
`configure: error: one or more libs available at link-time are not available run-time. Libs used at link-time: -lnghttp2 -lssl -lcrypto -lz`
And of course the .dylibs for those don't exist - they're static. I'm trying to do a fully 100% independent build of curl that has no reliance on any other dependencies - except for mac os x frameworks. I want my static openssl, my static nghttp2, and my static zlib.
Here is my full configure command:
`./configure --disable-shared --with-secure-transport --without-librtmp --without-libidn2 --with-zlib=/Users/jammerxd/Desktop/Git/cplusplusdependencies/zlib-1.2.11/MacOSX/x86_64/ --with-openssl=/Users/jammerxd/Desktop/Git/cplusplusdependencies/openssl-1.1.1m/MacOSX/x86_64 --disable-ldaps --disable-ntlm --disable-telnet --disable-proxy --disable-debug --enable-optimize --disable-curldebug --enable-symbol-hiding --disable-ftp --disable-rtsp --disable-dict --disable-tftp --disable-pop3 --disable-imap --disable-smb --disable-smtp --disable-gopher --disable-mqtt --disable-ntlm-wb --disable-tls-srp --disable-http-auth --disable-doh --disable-netrc --enable-progress-meter --disable-alt-svc --disable-ldap --disable-file --disable-dependency-tracking --with-nghttp2=/Users/jammerxd/Desktop/Git/cplusplusdependencies/nghttp2-1.49.0/MacOSX/x86_64/`
If I try to prefix the above command with:
`LDFLAGS="-static" PKG_CONFIG="pkg-config --static"`
the configure tool errors out with:
```
checking whether the C compiler works... no
configure: error: in `/Users/jammerxd/Desktop/Git/cplusplusdependencies/curl-7.84.0':
configure: error: C compiler cannot create executables
See `config.log' for more details
```
Looking on how I'm supposed to be able to build curl statically then have it link to the dependencies statically as well.
Thanks,
jammerxd | non_main | cannot build curl statically and use static dependencies mac os x hi there i m trying to build libcurl on apple mac os x i m doing a minimal build with https and http i have built out openssl zlib and as static libraries libssl a libcrypto a libz a the files exist in that path as needed and the configure script picks them up correctly however despite passing disable shared the configure script errors out with configure error one or more libs available at link time are not available run time libs used at link time lssl lcrypto lz and of course the dylibs for those don t exist they re static i m trying to do a fully independent build of curl that has no reliance on any other dependencies except for mac os x frameworks i want my static openssl my static and my static zlib here is my full configure command configure disable shared with secure transport without librtmp without with zlib users jammerxd desktop git cplusplusdependencies zlib macosx with openssl users jammerxd desktop git cplusplusdependencies openssl macosx disable ldaps disable ntlm disable telnet disable proxy disable debug enable optimize disable curldebug enable symbol hiding disable ftp disable rtsp disable dict disable tftp disable disable imap disable smb disable smtp disable gopher disable mqtt disable ntlm wb disable tls srp disable http auth disable doh disable netrc enable progress meter disable alt svc disable ldap disable file disable dependency tracking with users jammerxd desktop git cplusplusdependencies macosx if i try to prefix the above command with ldflags static pkg config pkg config static the configure tool errors out with checking whether the c compiler works no configure error in users jammerxd desktop git cplusplusdependencies curl configure error c compiler cannot create executables see config log for more details looking on how i m supposed to be able to build curl statically then have it link to the dependencies statically as well thanks jammerxd | 0 |
74,221 | 14,220,967,715 | IssuesEvent | 2020-11-17 15:09:16 | paulevsGitch/BetterNether | https://api.github.com/repos/paulevsGitch/BetterNether | closed | Player on seats snaps to south-facing position, and body lags behind head when turning | bug in dev code | Whilst testing the three chair offerings added by Better Nether, I noticed that the player model visible via F5 or another player's perspective has issues with the player's body not facing the same way the head is. The player's body seems to always want to face south, and, when the player turns their camera, the head follows perfectly, but the body lags behind more and more the further the camera goes from dead south. When facing dead north, there is a point where the body flips 180* instantly when continuing to turn around. This issue also means that chairs that don't face south have the player seated with their legs clipping through the back of the chair by default.
I would hope that this issue can be fixed for a lovelier sitting-down experience.~
**Attachments**
Facing dead south.

Turning away from south, the body lags behind, about halfway between where the head is facing and south.

This picture is right on the edge of where the body would flip, facing north.
 | 1.0 | Player on seats snaps to south-facing position, and body lags behind head when turning - Whilst testing the three chair offerings added by Better Nether, I noticed that the player model visible via F5 or another player's perspective has issues with the player's body not facing the same way the head is. The player's body seems to always want to face south, and, when the player turns their camera, the head follows perfectly, but the body lags behind more and more the further the camera goes from dead south. When facing dead north, there is a point where the body flips 180* instantly when continuing to turn around. This issue also means that chairs that don't face south have the player seated with their legs clipping through the back of the chair by default.
I would hope that this issue can be fixed for a lovelier sitting-down experience.~
**Attachments**
Facing dead south.

Turning away from south, the body lags behind, about halfway between where the head is facing and south.

This picture is right on the edge of where the body would flip, facing north.
 | non_main | player on seats snaps to south facing position and body lags behind head when turning whilst testing the three chair offerings added by better nether i noticed that the player model visible via or another player s perspective has issues with the player s body not facing the same way the head is the player s body seems to always want to face south and when the player turns their camera the head follows perfectly but the body lags behind more and more the further the camera goes from dead south when facing dead north there is a point where the body flips instantly when continuing to turn around this issue also means that chairs that don t face south have the player seated with their legs clipping through the back of the chair by default i would hope that this issue can be fixed for a lovelier sitting down experience attachments facing dead south turning away from south the body lags behind about halfway between where the head is facing and south this picture is right on the edge of where the body would flip facing north | 0 |
4,383 | 22,304,098,274 | IssuesEvent | 2022-06-13 11:27:08 | precice/precice | https://api.github.com/repos/precice/precice | opened | Clarify mesh API | enhancement usability maintainability breaking change | **Please describe the problem you are trying to solve.**
The API for setting mesh primitives is confusing and tedious.
* Triangles have to be set using edges
* `setMeshTriangleWithEdges` sounds like it would take edges, but it actually takes vertices. #1057
* There are no bulk functions for setting edges and triangles. #465
* Adding `setMeshTetrahedron` requiring triangles would be a huge pain for users. #1314
* Exposing handles to connectivity (EdgeID) prevents us from optimizing meshes #1313
**Describe the solution you propose.**
1. Change the API to a vertex-only style. **:warning: breaking**
2. Remove `XWithEdges`
3. Add bulk functions
Function | Inputs + MeshID | Outputs
--- | --- | ---
setMeshVertex | Coords | VertexID
setMeshVertices | Count, Coords | VertexIDs
setMeshEdge | VertexIDs | |
setMeshEdges | Count, VertexIDs | |
setMeshTriangle | VertexIDs | |
setMeshTriangles | Count, VertexIDs | |
setMeshQuad | VertexIDs | |
setMeshQuads | Count, VertexIDs | |
setMeshTetrahedron | VertexIDs | |
setMeshTetrahedra | Count, VertexIDs | |
**Describe alternatives you've considered**
Leave it as it and end up with an increasingly confusing API.
| True | Clarify mesh API - **Please describe the problem you are trying to solve.**
The API for setting mesh primitives is confusing and tedious.
* Triangles have to be set using edges
* `setMeshTriangleWithEdges` sounds like it would take edges, but it actually takes vertices. #1057
* There are no bulk functions for setting edges and triangles. #465
* Adding `setMeshTetrahedron` requiring triangles would be a huge pain for users. #1314
* Exposing handles to connectivity (EdgeID) prevents us from optimizing meshes #1313
**Describe the solution you propose.**
1. Change the API to a vertex-only style. **:warning: breaking**
2. Remove `XWithEdges`
3. Add bulk functions
Function | Inputs + MeshID | Outputs
--- | --- | ---
setMeshVertex | Coords | VertexID
setMeshVertices | Count, Coords | VertexIDs
setMeshEdge | VertexIDs | |
setMeshEdges | Count, VertexIDs | |
setMeshTriangle | VertexIDs | |
setMeshTriangles | Count, VertexIDs | |
setMeshQuad | VertexIDs | |
setMeshQuads | Count, VertexIDs | |
setMeshTetrahedron | VertexIDs | |
setMeshTetrahedra | Count, VertexIDs | |
**Describe alternatives you've considered**
Leave it as it and end up with an increasingly confusing API.
| main | clarify mesh api please describe the problem you are trying to solve the api for setting mesh primitives is confusing and tedious triangles have to be set using edges setmeshtrianglewithedges sounds like it would take edges but it actually takes vertices there are no bulk functions for setting edges and triangles adding setmeshtetrahedron requiring triangles would be a huge pain for users exposing handles to connectivity edgeid prevents us from optimizing meshes describe the solution you propose change the api to a vertex only style warning breaking remove xwithedges add bulk functions function inputs meshid outputs setmeshvertex coords vertexid setmeshvertices count coords vertexids setmeshedge vertexids setmeshedges count vertexids setmeshtriangle vertexids setmeshtriangles count vertexids setmeshquad vertexids setmeshquads count vertexids setmeshtetrahedron vertexids setmeshtetrahedra count vertexids describe alternatives you ve considered leave it as it and end up with an increasingly confusing api | 1 |
294 | 3,072,891,779 | IssuesEvent | 2015-08-19 19:09:39 | cattolyst/datafinisher | https://api.github.com/repos/cattolyst/datafinisher | opened | Test DF on i2b2 server | maintainability qc | Since the anticipated production environment would be the i2b2 server at our site (because that's where the JSON files live) it's kind of important that it be able to run there. However, ticket #22 causes an error and prevents us from finding out if anything else will break further on down.
Once #22 is closed, need to test DF on i2b2. | True | Test DF on i2b2 server - Since the anticipated production environment would be the i2b2 server at our site (because that's where the JSON files live) it's kind of important that it be able to run there. However, ticket #22 causes an error and prevents us from finding out if anything else will break further on down.
Once #22 is closed, need to test DF on i2b2. | main | test df on server since the anticipated production environment would be the server at our site because that s where the json files live it s kind of important that it be able to run there however ticket causes an error and prevents us from finding out if anything else will break further on down once is closed need to test df on | 1 |
107,529 | 4,310,062,460 | IssuesEvent | 2016-07-21 17:59:41 | DDMAL/rodan-client | https://api.github.com/repos/DDMAL/rodan-client | closed | Provide form for user to set info | Priority: SHOW STOPPER Type: FEATURE | - password reset (INCREDIBLY IMPORTANT)
- name
- email (VERY IMPORTANT)
- get general Rodan notices (requires Rodan model update)
- get project notices (requires Rodan model update) | 1.0 | Provide form for user to set info - - password reset (INCREDIBLY IMPORTANT)
- name
- email (VERY IMPORTANT)
- get general Rodan notices (requires Rodan model update)
- get project notices (requires Rodan model update) | non_main | provide form for user to set info password reset incredibly important name email very important get general rodan notices requires rodan model update get project notices requires rodan model update | 0 |
216,550 | 24,281,564,666 | IssuesEvent | 2022-09-28 17:53:31 | liorzilberg/swagger-parser | https://api.github.com/repos/liorzilberg/swagger-parser | closed | CVE-2020-35728 (High) detected in jackson-databind-2.9.5.jar - autoclosed | security vulnerability | ## CVE-2020-35728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/target/lib/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/liorzilberg/swagger-parser/commits/299682f5b4a2ec420c0c3f91a170670051db10d0">299682f5b4a2ec420c0c3f91a170670051db10d0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.oracle.wls.shaded.org.apache.xalan.lib.sql.JNDIConnectionPool (aka embedded Xalan in org.glassfish.web/javax.servlet.jsp.jstl).
<p>Publish Date: 2020-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35728>CVE-2020-35728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728</a></p>
<p>Release Date: 2020-12-27</p>
<p>Fix Resolution: 2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-35728 (High) detected in jackson-databind-2.9.5.jar - autoclosed - ## CVE-2020-35728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/target/lib/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/liorzilberg/swagger-parser/commits/299682f5b4a2ec420c0c3f91a170670051db10d0">299682f5b4a2ec420c0c3f91a170670051db10d0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.oracle.wls.shaded.org.apache.xalan.lib.sql.JNDIConnectionPool (aka embedded Xalan in org.glassfish.web/javax.servlet.jsp.jstl).
<p>Publish Date: 2020-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35728>CVE-2020-35728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728</a></p>
<p>Release Date: 2020-12-27</p>
<p>Fix Resolution: 2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar target lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com oracle wls shaded org apache xalan lib sql jndiconnectionpool aka embedded xalan in org glassfish web javax servlet jsp jstl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
2,028 | 6,778,579,808 | IssuesEvent | 2017-10-28 12:49:56 | chocolatey/chocolatey-package-requests | https://api.github.com/repos/chocolatey/chocolatey-package-requests | closed | RFP sdkman | Status: Available For Maintainer(s) | http://sdkman.io is a package for technology built on JVM(java virtual machine) such as grails, groovy, gradel etc | True | RFP sdkman - http://sdkman.io is a package for technology built on JVM(java virtual machine) such as grails, groovy, gradel etc | main | rfp sdkman is a package for technology built on jvm java virtual machine such as grails groovy gradel etc | 1 |
4,857 | 24,996,288,577 | IssuesEvent | 2022-11-03 00:47:17 | usefulmove/comp | https://api.github.com/repos/usefulmove/comp | closed | Command response function code reuse | maintainability | There is a significant amount of code duplication in the command response functions (`c_...`). It may be possible to simplify the code in meaningful ways ( poss. using function composition ). | True | Command response function code reuse - There is a significant amount of code duplication in the command response functions (`c_...`). It may be possible to simplify the code in meaningful ways ( poss. using function composition ). | main | command response function code reuse there is a significant amount of code duplication in the command response functions c it may be possible to simplify the code in meaningful ways poss using function composition | 1 |
315,479 | 23,581,438,225 | IssuesEvent | 2022-08-23 08:01:24 | hyperledger/iroha-2-docs | https://api.github.com/repos/hyperledger/iroha-2-docs | closed | Document FFI bindgen name mangling semantics | documentation enhancement help-wanted iroha2 | ```
Aleksandr Petrosyan, [25.07.22 19:41]
That being said, we should document that convention too.
rust
module_name__method_name
StructName__inherent_fn
TraitName__assoc_fn_name
…
Marin, [25.07.22 19:56]
[In reply to Aleksandr Petrosyan]
1. module__name::__method_name -> we do this for functions that have the same API but are implemented on different types. Namely this is for trait impls, i.e. poor man's generics like Clone
2. yes
3. no, it's actually StructName__trait_method_name at the moment, but, in my opinion, should be reduced to 1. after 2488 (https://github.com/hyperledger/iroha/issues/2488)
Marin, [25.07.22 19:58]
1. for Clone in iroha_crypto gen_ffi_impl will create iroha_crypto::__clone in that library equivalent of Clone::clone
Aleksandr Petrosyan, [25.07.22 19:59]
This looks like we're mangling by hand.
Marin, [25.07.22 19:59]
which one?
Aleksandr Petrosyan, [25.07.22 20:00]
[In reply to Marin]
This one
Aleksandr Petrosyan, [25.07.22 20:00]
Under the circumstances I doubt we can do much better
Marin, [25.07.22 20:01]
I edited the comment
Aleksandr Petrosyan, [25.07.22 20:02]
[In reply to Marin]
Got it.
``` | 1.0 | Document FFI bindgen name mangling semantics - ```
Aleksandr Petrosyan, [25.07.22 19:41]
That being said, we should document that convention too.
rust
module_name__method_name
StructName__inherent_fn
TraitName__assoc_fn_name
…
Marin, [25.07.22 19:56]
[In reply to Aleksandr Petrosyan]
1. module__name::__method_name -> we do this for functions that have the same API but are implemented on different types. Namely this is for trait impls, i.e. poor man's generics like Clone
2. yes
3. no, it's actually StructName__trait_method_name at the moment, but, in my opinion, should be reduced to 1. after 2488 (https://github.com/hyperledger/iroha/issues/2488)
Marin, [25.07.22 19:58]
1. for Clone in iroha_crypto gen_ffi_impl will create iroha_crypto::__clone in that library equivalent of Clone::clone
Aleksandr Petrosyan, [25.07.22 19:59]
This looks like we're mangling by hand.
Marin, [25.07.22 19:59]
which one?
Aleksandr Petrosyan, [25.07.22 20:00]
[In reply to Marin]
This one
Aleksandr Petrosyan, [25.07.22 20:00]
Under the circumstances I doubt we can do much better
Marin, [25.07.22 20:01]
I edited the comment
Aleksandr Petrosyan, [25.07.22 20:02]
[In reply to Marin]
Got it.
``` | non_main | document ffi bindgen name mangling semantics aleksandr petrosyan that being said we should document that convention too rust module name method name structname inherent fn traitname assoc fn name … marin module name method name we do this for functions that have the same api but are implemented on different types namely this is for trait impls i e poor man s generics like clone yes no it s actually structname trait method name at the moment but in my opinion should be reduced to after marin for clone in iroha crypto gen ffi impl will create iroha crypto clone in that library equivalent of clone clone aleksandr petrosyan this looks like we re mangling by hand marin which one aleksandr petrosyan this one aleksandr petrosyan under the circumstances i doubt we can do much better marin i edited the comment aleksandr petrosyan got it | 0 |
4,932 | 25,350,832,636 | IssuesEvent | 2022-11-19 18:56:37 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Partial summarization does not consider partial grouping_expressions and aggregation_expressions | type: bug work: backend status: ready restricted: maintainers | ## Description
In the [partial summarizations PR](https://github.com/centerofci/mathesar/pull/1879), an important requirement that was discussed in earlier meetings have been possibly missed. It was decided that all `initial_columns` will be automatically identified as whether they are grouped/aggregated and included within either `grouping_expressions` or `aggregation_expressions`, and that the user will be able to partially specify both `grouping_expressions` and `aggregation_expressions`.
Currently, the API only performs "guessing" when 'grouping_expressions' and 'aggregation_expressions' are both not present in the request. When they are present (partially or even empty), it does not do anything.
cc: @dmos62 @mathemancer | True | Partial summarization does not consider partial grouping_expressions and aggregation_expressions - ## Description
In the [partial summarizations PR](https://github.com/centerofci/mathesar/pull/1879), an important requirement that was discussed in earlier meetings have been possibly missed. It was decided that all `initial_columns` will be automatically identified as whether they are grouped/aggregated and included within either `grouping_expressions` or `aggregation_expressions`, and that the user will be able to partially specify both `grouping_expressions` and `aggregation_expressions`.
Currently, the API only performs "guessing" when 'grouping_expressions' and 'aggregation_expressions' are both not present in the request. When they are present (partially or even empty), it does not do anything.
cc: @dmos62 @mathemancer | main | partial summarization does not consider partial grouping expressions and aggregation expressions description in the an important requirement that was discussed in earlier meetings have been possibly missed it was decided that all initial columns will be automatically identified as whether they are grouped aggregated and included within either grouping expressions or aggregation expressions and that the user will be able to partially specify both grouping expressions and aggregation expressions currently the api only performs guessing when grouping expressions and aggregation expressions are both not present in the request when they are present partially or even empty it does not do anything cc mathemancer | 1 |
210,050 | 16,332,333,168 | IssuesEvent | 2021-05-12 10:46:31 | oss-review-toolkit/ort | https://api.github.com/repos/oss-review-toolkit/ort | closed | package config: Placeholders in referenced version? | documentation question | Hello,
I'm just getting started with ORT and I'm wondering about one part of the docs. Referring to [this section of the package exclude documentation](https://github.com/oss-review-toolkit/ort/blob/master/docs/config-file-package-configuration-yml.md#defining-path-excludes-and-license-finding-curations): I want to exclude a dependency from the evaluation _regardless of its version_ because I own it and it is covered by another OSS analysis already.
The documentation tells:
> Each package configuration applies exactly to one package id and provenance which must be specified.
The ability to omit the version or use placeholders would be very helpful. If I understand the documentation correctly, one needs to update the package configuration whenever one bumps the version of any dependency that is mentioned therein. Isn't that causing needless maintenance toil?
Desired solution:
* possibility to declare an upwards-compatible version – as offered by the caret notation of npm: e.g. `^1.2.10` or completely omit the version
* If this is not currently offered due to implementation constraints, then the docs should say so.
* documentation refined in regards to this use-case
Pointer:
https://github.com/oss-review-toolkit/ort/commit/0aaff6b04c6121cf511d9438596f2b177d2fba2d by @fviernau | 1.0 | package config: Placeholders in referenced version? - Hello,
I'm just getting started with ORT and I'm wondering about one part of the docs. Referring to [this section of the package exclude documentation](https://github.com/oss-review-toolkit/ort/blob/master/docs/config-file-package-configuration-yml.md#defining-path-excludes-and-license-finding-curations): I want to exclude a dependency from the evaluation _regardless of its version_ because I own it and it is covered by another OSS analysis already.
The documentation tells:
> Each package configuration applies exactly to one package id and provenance which must be specified.
The ability to omit the version or use placeholders would be very helpful. If I understand the documentation correctly, one needs to update the package configuration whenever one bumps the version of any dependency that is mentioned therein. Isn't that causing needless maintenance toil?
Desired solution:
* possibility to declare an upwards-compatible version – as offered by the caret notation of npm: e.g. `^1.2.10` or completely omit the version
* If this is not currently offered due to implementation constraints, then the docs should say so.
* documentation refined in regards to this use-case
Pointer:
https://github.com/oss-review-toolkit/ort/commit/0aaff6b04c6121cf511d9438596f2b177d2fba2d by @fviernau | non_main | package config placeholders in referenced version hello i m just getting started with ort and i m wondering about one part of the docs referring to i want to exclude a dependency from the evaluation regardless of its version because i own it and it is covered by another oss analysis already the documentation tells each package configuration applies exactly to one package id and provenance which must be specified the ability to omit the version or use placeholders would be very helpful if i understand the documentation correctly one needs to update the package configuration whenever one bumps the version of any dependency that is mentioned therein isn t that causing needless maintenance toil desired solution possibility to declare an upwards compatible version – as offered by the caret notation of npm e g or completely omit the version if this is not currently offered due to implementation constraints then the docs should say so documentation refined in regards to this use case pointer by fviernau | 0 |
162,630 | 20,235,413,251 | IssuesEvent | 2022-02-14 01:06:39 | johnnymythology/material-blog-jp | https://api.github.com/repos/johnnymythology/material-blog-jp | closed | CVE-2010-3299 (Medium) detected in rails-5.0.7.2.gem - autoclosed | security vulnerability | ## CVE-2010-3299 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rails-5.0.7.2.gem</b></p></summary>
<p>Ruby on Rails is a full-stack web framework optimized for programmer happiness and sustainable productivity. It encourages beautiful code by favoring convention over configuration.</p>
<p>Library home page: <a href="https://rubygems.org/gems/rails-5.0.7.2.gem">https://rubygems.org/gems/rails-5.0.7.2.gem</a></p>
<p>Path to dependency file: /material-blog-jp/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/rails-5.0.7.2.gem</p>
<p>
Dependency Hierarchy:
- rogue-0.1.1.gem (Root Library)
- :x: **rails-5.0.7.2.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The encrypt/decrypt functions in Ruby on Rails 2.3 are vulnerable to padding oracle attacks.
<p>Publish Date: 2019-11-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2010-3299>CVE-2010-3299</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2010-3299">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2010-3299</a></p>
<p>Release Date: 2019-11-12</p>
<p>Fix Resolution: rails - 5.2.0.beta1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2010-3299 (Medium) detected in rails-5.0.7.2.gem - autoclosed - ## CVE-2010-3299 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rails-5.0.7.2.gem</b></p></summary>
<p>Ruby on Rails is a full-stack web framework optimized for programmer happiness and sustainable productivity. It encourages beautiful code by favoring convention over configuration.</p>
<p>Library home page: <a href="https://rubygems.org/gems/rails-5.0.7.2.gem">https://rubygems.org/gems/rails-5.0.7.2.gem</a></p>
<p>Path to dependency file: /material-blog-jp/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/rails-5.0.7.2.gem</p>
<p>
Dependency Hierarchy:
- rogue-0.1.1.gem (Root Library)
- :x: **rails-5.0.7.2.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The encrypt/decrypt functions in Ruby on Rails 2.3 are vulnerable to padding oracle attacks.
<p>Publish Date: 2019-11-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2010-3299>CVE-2010-3299</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2010-3299">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2010-3299</a></p>
<p>Release Date: 2019-11-12</p>
<p>Fix Resolution: rails - 5.2.0.beta1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in rails gem autoclosed cve medium severity vulnerability vulnerable library rails gem ruby on rails is a full stack web framework optimized for programmer happiness and sustainable productivity it encourages beautiful code by favoring convention over configuration library home page a href path to dependency file material blog jp gemfile lock path to vulnerable library var lib gems cache rails gem dependency hierarchy rogue gem root library x rails gem vulnerable library vulnerability details the encrypt decrypt functions in ruby on rails are vulnerable to padding oracle attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rails step up your open source security game with whitesource | 0 |
2,002 | 6,717,616,077 | IssuesEvent | 2017-10-14 23:49:09 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Shuttle subsystem recovery needs a do over | Bug Maintainability/Hinders improvements | It's very broken and you can tell by looking at the code. | True | Shuttle subsystem recovery needs a do over - It's very broken and you can tell by looking at the code. | main | shuttle subsystem recovery needs a do over it s very broken and you can tell by looking at the code | 1 |
143,718 | 13,081,329,393 | IssuesEvent | 2020-08-01 10:40:39 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | Documentation issue: Build Tutorial - Android: Example does not build | team-Android type: documentation untriaged | Documentation URL: https://docs.bazel.build/versions/3.3.0/tutorial/android-app.html
Installed bazel following the Ubuntu Linux using the apt repository on 18.04.4 LTS.
Stacktrace:
```
SEVERE: Unexpected
com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098
at com.google.common.base.Verify.verify(Verify.java:301)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668)
at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108)
at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258)
at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231)
at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81)
at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233)
at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177)
Jul 10, 2020 12:52:58 PM com.google.devtools.build.android.ResourceProcessorBusyBox processRequest
SEVERE: Error during processing
com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098
at com.google.common.base.Verify.verify(Verify.java:301)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668)
at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108)
at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258)
at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231)
at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81)
at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233)
at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177)
Exception in thread "main" com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098
at com.google.common.base.Verify.verify(Verify.java:301)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668)
at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108)
at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258)
at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231)
at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81)
at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233)
at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177)
Target //src/main:app failed to build
INFO: Elapsed time: 11.369s, Critical Path: 4.85s
INFO: 18 processes: 11 linux-sandbox, 7 worker.
FAILED: Build did NOT complete successfully
```
Patch with what I've done so far (on top of https://github.com/bazelbuild/examples/tree/e5d77821c2cd686f87310111499b11e70ca56f8e):
```
diff --git a/android/tutorial/WORKSPACE b/android/tutorial/WORKSPACE
new file mode 100644
index 0000000..6f41f30
--- /dev/null
+++ b/android/tutorial/WORKSPACE
@@ -0,0 +1,6 @@
+android_sdk_repository(
+ name = "androidsdk",
+ path = "/home/jorge/Android/Sdk",
+ api_level = 25,
+ build_tools_version = "26.0.1"
+)
diff --git a/android/tutorial/src/main/BUILD b/android/tutorial/src/main/BUILD
new file mode 100644
index 0000000..ce1490e
--- /dev/null
+++ b/android/tutorial/src/main/BUILD
@@ -0,0 +1,5 @@
+android_binary(
+ name = "app",
+ manifest = "AndroidManifest.xml",
+ deps = ["//src/main/java/com/example/bazel:greeter_activity"],
+)
diff --git a/android/tutorial/src/main/java/com/example/bazel/BUILD b/android/tutorial/src/main/java/com/example/bazel/BUILD
new file mode 100644
index 0000000..76f8ba7
--- /dev/null
+++ b/android/tutorial/src/main/java/com/example/bazel/BUILD
@@ -0,0 +1,13 @@
+package(
+ default_visibility = ["//src:__subpackages__"],
+)
+
+android_library(
+ name = "greeter_activity",
+ srcs = [
+ "Greeter.java",
+ "MainActivity.java",
+ ],
+ manifest = "AndroidManifest.xml",
+ resource_files = glob(["res/**"]),
+)
```
| 1.0 | Documentation issue: Build Tutorial - Android: Example does not build - Documentation URL: https://docs.bazel.build/versions/3.3.0/tutorial/android-app.html
Installed bazel following the Ubuntu Linux using the apt repository on 18.04.4 LTS.
Stacktrace:
```
SEVERE: Unexpected
com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098
at com.google.common.base.Verify.verify(Verify.java:301)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668)
at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108)
at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258)
at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231)
at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81)
at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233)
at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177)
Jul 10, 2020 12:52:58 PM com.google.devtools.build.android.ResourceProcessorBusyBox processRequest
SEVERE: Error during processing
com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098
at com.google.common.base.Verify.verify(Verify.java:301)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668)
at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108)
at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258)
at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231)
at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81)
at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233)
at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177)
Exception in thread "main" com.google.common.base.VerifyException: Unexpected magic number in generated/src/main/java/com/example/bazel/res/values_colors.arsc.flat: 470425098
at com.google.common.base.Verify.verify(Verify.java:301)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.readResourceContainer(AndroidCompiledDataDeserializer.java:719)
at com.google.devtools.build.android.AndroidCompiledDataDeserializer.read(AndroidCompiledDataDeserializer.java:668)
at com.google.devtools.build.android.SerializedAndroidData.deserialize(SerializedAndroidData.java:108)
at com.google.devtools.build.android.AndroidResourceMerger.mergeCompiledData(AndroidResourceMerger.java:258)
at com.google.devtools.build.android.AndroidCompiledResourceMergingAction.main(AndroidCompiledResourceMergingAction.java:231)
at com.google.devtools.build.android.ResourceProcessorBusyBox$Tool$3.call(ResourceProcessorBusyBox.java:81)
at com.google.devtools.build.android.ResourceProcessorBusyBox.processRequest(ResourceProcessorBusyBox.java:233)
at com.google.devtools.build.android.ResourceProcessorBusyBox.main(ResourceProcessorBusyBox.java:177)
Target //src/main:app failed to build
INFO: Elapsed time: 11.369s, Critical Path: 4.85s
INFO: 18 processes: 11 linux-sandbox, 7 worker.
FAILED: Build did NOT complete successfully
```
Patch with what I've done so far (on top of https://github.com/bazelbuild/examples/tree/e5d77821c2cd686f87310111499b11e70ca56f8e):
```
diff --git a/android/tutorial/WORKSPACE b/android/tutorial/WORKSPACE
new file mode 100644
index 0000000..6f41f30
--- /dev/null
+++ b/android/tutorial/WORKSPACE
@@ -0,0 +1,6 @@
+android_sdk_repository(
+ name = "androidsdk",
+ path = "/home/jorge/Android/Sdk",
+ api_level = 25,
+ build_tools_version = "26.0.1"
+)
diff --git a/android/tutorial/src/main/BUILD b/android/tutorial/src/main/BUILD
new file mode 100644
index 0000000..ce1490e
--- /dev/null
+++ b/android/tutorial/src/main/BUILD
@@ -0,0 +1,5 @@
+android_binary(
+ name = "app",
+ manifest = "AndroidManifest.xml",
+ deps = ["//src/main/java/com/example/bazel:greeter_activity"],
+)
diff --git a/android/tutorial/src/main/java/com/example/bazel/BUILD b/android/tutorial/src/main/java/com/example/bazel/BUILD
new file mode 100644
index 0000000..76f8ba7
--- /dev/null
+++ b/android/tutorial/src/main/java/com/example/bazel/BUILD
@@ -0,0 +1,13 @@
+package(
+ default_visibility = ["//src:__subpackages__"],
+)
+
+android_library(
+ name = "greeter_activity",
+ srcs = [
+ "Greeter.java",
+ "MainActivity.java",
+ ],
+ manifest = "AndroidManifest.xml",
+ resource_files = glob(["res/**"]),
+)
```
| non_main | documentation issue build tutorial android example does not build documentation url installed bazel following the ubuntu linux using the apt repository on lts stacktrace severe unexpected com google common base verifyexception unexpected magic number in generated src main java com example bazel res values colors arsc flat at com google common base verify verify verify java at com google devtools build android androidcompileddatadeserializer readresourcecontainer androidcompileddatadeserializer java at com google devtools build android androidcompileddatadeserializer read androidcompileddatadeserializer java at com google devtools build android serializedandroiddata deserialize serializedandroiddata java at com google devtools build android androidresourcemerger mergecompileddata androidresourcemerger java at com google devtools build android androidcompiledresourcemergingaction main androidcompiledresourcemergingaction java at com google devtools build android resourceprocessorbusybox tool call resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox processrequest resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox main resourceprocessorbusybox java jul pm com google devtools build android resourceprocessorbusybox processrequest severe error during processing com google common base verifyexception unexpected magic number in generated src main java com example bazel res values colors arsc flat at com google common base verify verify verify java at com google devtools build android androidcompileddatadeserializer readresourcecontainer androidcompileddatadeserializer java at com google devtools build android androidcompileddatadeserializer read androidcompileddatadeserializer java at com google devtools build android serializedandroiddata deserialize serializedandroiddata java at com google devtools build android androidresourcemerger mergecompileddata androidresourcemerger java at com google devtools build android androidcompiledresourcemergingaction main androidcompiledresourcemergingaction java at com google devtools build android resourceprocessorbusybox tool call resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox processrequest resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox main resourceprocessorbusybox java exception in thread main com google common base verifyexception unexpected magic number in generated src main java com example bazel res values colors arsc flat at com google common base verify verify verify java at com google devtools build android androidcompileddatadeserializer readresourcecontainer androidcompileddatadeserializer java at com google devtools build android androidcompileddatadeserializer read androidcompileddatadeserializer java at com google devtools build android serializedandroiddata deserialize serializedandroiddata java at com google devtools build android androidresourcemerger mergecompileddata androidresourcemerger java at com google devtools build android androidcompiledresourcemergingaction main androidcompiledresourcemergingaction java at com google devtools build android resourceprocessorbusybox tool call resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox processrequest resourceprocessorbusybox java at com google devtools build android resourceprocessorbusybox main resourceprocessorbusybox java target src main app failed to build info elapsed time critical path info processes linux sandbox worker failed build did not complete successfully patch with what i ve done so far on top of diff git a android tutorial workspace b android tutorial workspace new file mode index dev null b android tutorial workspace android sdk repository name androidsdk path home jorge android sdk api level build tools version diff git a android tutorial src main build b android tutorial src main build new file mode index dev null b android tutorial src main build android binary name app manifest androidmanifest xml deps diff git a android tutorial src main java com example bazel build b android tutorial src main java com example bazel build new file mode index dev null b android tutorial src main java com example bazel build package default visibility android library name greeter activity srcs greeter java mainactivity java manifest androidmanifest xml resource files glob | 0 |
352,754 | 10,545,856,233 | IssuesEvent | 2019-10-02 20:05:59 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | closed | Edit collection page closes when the language is changed | component: collections priority: p4 state: stale type: bug | STR:
1. Load AMO dev and open one of your collections
2. Click Edit collection details, make some changes (edit the text)
3. Change the language
4. Observe the screen
Actual results:
Collection's page is displayed not the edit page and edited text is lost.
Expected results:
The Edit collection page is still displayed.
Notes:
This issue is reproducible on AMO dev and AMO stage with FF62(Win10).

| 1.0 | Edit collection page closes when the language is changed - STR:
1. Load AMO dev and open one of your collections
2. Click Edit collection details, make some changes (edit the text)
3. Change the language
4. Observe the screen
Actual results:
Collection's page is displayed not the edit page and edited text is lost.
Expected results:
The Edit collection page is still displayed.
Notes:
This issue is reproducible on AMO dev and AMO stage with FF62(Win10).

| non_main | edit collection page closes when the language is changed str load amo dev and open one of your collections click edit collection details make some changes edit the text change the language observe the screen actual results collection s page is displayed not the edit page and edited text is lost expected results the edit collection page is still displayed notes this issue is reproducible on amo dev and amo stage with | 0 |
11,746 | 13,837,099,693 | IssuesEvent | 2020-10-14 02:50:45 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | reopened | Remove `NtIdent` hack for regressed crates | A-proc-macros C-future-compatibility T-compiler | ## What is this issue?
If you're a crate author who's been linked here, this issue tracks removing a backwards-compatibility hack in Rust.
Rust has a longstanding issue https://github.com/rust-lang/rust/issues/43081, which causes procedural macros to lose location and hygiene information (known as a "Span") under certain circumstances. Recently, pull request https://github.com/rust-lang/rust/pull/73084 was merged, which makes progress towards resolving https://github.com/rust-lang/rust/issues/43081.
Unfortunately, older versions of certain procedural macro crates (such as `proc-macro-hack v0.5.15` and `js-sys v0.3.39`) cannot handle the changes in input caused by the Rust bugfix. To allow these crates to continue to compile, a backward-compatibility hack was added to adjust the input passed to `proc-macro-hack` and `js-sys` specifically.
Eventually, we would like to remove this backwards-compatibility hack, since the compiler should not have hard-coded exceptions for certain crates. However, removing this hack will break any crates that depend on affected versions of `proc-macro-hack` or `js-sys`.
To ensure that your crate continues to work, you'll want to ensure that your `Cargo.lock` references `proc-macro-hack` v0.5.16 or above, and `js-sys` v0.3.40 or above. This can be done by running `cargo update -p proc-macro-hack` and `cargo update -p js-sys`. If you maintain a library crates (without a `Cargo.lock`, no action is needed on your part).
### Internal compiler details
In https://github.com/rust-lang/rust/pull/73084#issuecomment-652613950, I added a hack to change the behavior of `NtIdent`s passed to certain proc-macros. This was done by special-casing certain identifiers, and should be eventually be removed in favor of a proper solution.
If we decide to always wrap single identifiers in `None`-delimited groups, then we will need to wait until enough of the ecosystem has bumped `proc-macro-hack` and `wasm-bindgen`, to avoid breaking a large number of crates.
Crater run: https://crater-reports.s3.amazonaws.com/pr-73084-1/index.html
Triage: https://hackmd.io/O7icbSylRP6uVZyAQ9EDeA | True | Remove `NtIdent` hack for regressed crates - ## What is this issue?
If you're a crate author who's been linked here, this issue tracks removing a backwards-compatibility hack in Rust.
Rust has a longstanding issue https://github.com/rust-lang/rust/issues/43081, which causes procedural macros to lose location and hygiene information (known as a "Span") under certain circumstances. Recently, pull request https://github.com/rust-lang/rust/pull/73084 was merged, which makes progress towards resolving https://github.com/rust-lang/rust/issues/43081.
Unfortunately, older versions of certain procedural macro crates (such as `proc-macro-hack v0.5.15` and `js-sys v0.3.39`) cannot handle the changes in input caused by the Rust bugfix. To allow these crates to continue to compile, a backward-compatibility hack was added to adjust the input passed to `proc-macro-hack` and `js-sys` specifically.
Eventually, we would like to remove this backwards-compatibility hack, since the compiler should not have hard-coded exceptions for certain crates. However, removing this hack will break any crates that depend on affected versions of `proc-macro-hack` or `js-sys`.
To ensure that your crate continues to work, you'll want to ensure that your `Cargo.lock` references `proc-macro-hack` v0.5.16 or above, and `js-sys` v0.3.40 or above. This can be done by running `cargo update -p proc-macro-hack` and `cargo update -p js-sys`. If you maintain a library crates (without a `Cargo.lock`, no action is needed on your part).
### Internal compiler details
In https://github.com/rust-lang/rust/pull/73084#issuecomment-652613950, I added a hack to change the behavior of `NtIdent`s passed to certain proc-macros. This was done by special-casing certain identifiers, and should be eventually be removed in favor of a proper solution.
If we decide to always wrap single identifiers in `None`-delimited groups, then we will need to wait until enough of the ecosystem has bumped `proc-macro-hack` and `wasm-bindgen`, to avoid breaking a large number of crates.
Crater run: https://crater-reports.s3.amazonaws.com/pr-73084-1/index.html
Triage: https://hackmd.io/O7icbSylRP6uVZyAQ9EDeA | non_main | remove ntident hack for regressed crates what is this issue if you re a crate author who s been linked here this issue tracks removing a backwards compatibility hack in rust rust has a longstanding issue which causes procedural macros to lose location and hygiene information known as a span under certain circumstances recently pull request was merged which makes progress towards resolving unfortunately older versions of certain procedural macro crates such as proc macro hack and js sys cannot handle the changes in input caused by the rust bugfix to allow these crates to continue to compile a backward compatibility hack was added to adjust the input passed to proc macro hack and js sys specifically eventually we would like to remove this backwards compatibility hack since the compiler should not have hard coded exceptions for certain crates however removing this hack will break any crates that depend on affected versions of proc macro hack or js sys to ensure that your crate continues to work you ll want to ensure that your cargo lock references proc macro hack or above and js sys or above this can be done by running cargo update p proc macro hack and cargo update p js sys if you maintain a library crates without a cargo lock no action is needed on your part internal compiler details in i added a hack to change the behavior of ntident s passed to certain proc macros this was done by special casing certain identifiers and should be eventually be removed in favor of a proper solution if we decide to always wrap single identifiers in none delimited groups then we will need to wait until enough of the ecosystem has bumped proc macro hack and wasm bindgen to avoid breaking a large number of crates crater run triage | 0 |
5,719 | 30,232,146,500 | IssuesEvent | 2023-07-06 07:44:59 | UBCSailbot/raye-local-pathfinding | https://api.github.com/repos/UBCSailbot/raye-local-pathfinding | closed | Remove references to min_voltage | bug maintain | ### Purpose
<!-- What is the purpose of this issue? What does this issue hope to achieve? -->
The `min_voltage` ROS topic was removed in a recent PR to sailbot_msg. Local pathfinding still references `min_voltage`, which prevents the workspace from building.
### Changes
<!-- What changes will be made to an existing feature? How will it be better? -->
Remove/refactor all references to `min_voltage`.
| True | Remove references to min_voltage - ### Purpose
<!-- What is the purpose of this issue? What does this issue hope to achieve? -->
The `min_voltage` ROS topic was removed in a recent PR to sailbot_msg. Local pathfinding still references `min_voltage`, which prevents the workspace from building.
### Changes
<!-- What changes will be made to an existing feature? How will it be better? -->
Remove/refactor all references to `min_voltage`.
| main | remove references to min voltage purpose the min voltage ros topic was removed in a recent pr to sailbot msg local pathfinding still references min voltage which prevents the workspace from building changes remove refactor all references to min voltage | 1 |
4,955 | 25,455,594,941 | IssuesEvent | 2022-11-24 13:57:17 | pace/bricks | https://api.github.com/repos/pace/bricks | closed | Provide test helper for metrics | EST::Hours S::Ready T::Maintainance | ### Problem
It is not easy to test if the metrics are really collected. Currently, we rely on the availability `/metrics` and use that hardcoded in our tests.
### Solution
Provide a small `test/testmetrics` package that supports a user to see if metrics are collected.
The "/metrics" should be taken from the router instead of hard-coding it.
_Originally posted by @Teelevision in https://github.com/pace/bricks/pull/92_ | True | Provide test helper for metrics - ### Problem
It is not easy to test if the metrics are really collected. Currently, we rely on the availability `/metrics` and use that hardcoded in our tests.
### Solution
Provide a small `test/testmetrics` package that supports a user to see if metrics are collected.
The "/metrics" should be taken from the router instead of hard-coding it.
_Originally posted by @Teelevision in https://github.com/pace/bricks/pull/92_ | main | provide test helper for metrics problem it is not easy to test if the metrics are really collected currently we rely on the availability metrics and use that hardcoded in our tests solution provide a small test testmetrics package that supports a user to see if metrics are collected the metrics should be taken from the router instead of hard coding it originally posted by teelevision in | 1 |
2,837 | 10,209,489,549 | IssuesEvent | 2019-08-14 12:50:52 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | rax_facts needs to be renamed to _info and stop returning ansible_facts | affects_2.9 bug cloud module needs_maintainer needs_triage support:community | ##### SUMMARY
According to the [requirements](https://docs.ansible.com/ansible/latest/dev_guide/developing_modules_checklist.html#contributing-to-ansible-objective-requirements), `_facts` modules should return information that is specific to the host the module runs on as `ansible_facts`. (See also ansible/ansible#54280 for some discussion.)
I think that this module do not satisfy this condition: it returns information that is not specific to the host machine, but bound to a cloud account. Please verify whether this is actually the case.
To fix this, the core team thinks the best approach is to rename the `_facts` modules to `_info` modules, and change the return value so that if the module is called with the `_facts` name (should be deprecated, and will then be removed in Ansible 2.13), it should return `ansible_facts` as now (and print a deprecation message), and if it is called with the `_info` name, it no longer returns `ansible_facts`, so that the result needs to be registered. This informs users that they have to change something, won't break their playbooks, and allows them (by switching to the `_info` names) to verify that their roles and playbooks no longer use the facts.
If that approach doesn't work, or you want to change the module's interface (like rename the return value(s)), it is better to deprecate the `_facts` module and create a new `_info` module next to it. To see both approaches implemented, you can take a look at https://github.com/ansible/ansible/commit/2d98734ad51c3de557c3a7d1f7166bdd114ceb0f (ansible/ansible#60178).
If you have any input — preferred method to solve this, doubts that all modules should be renamed, ... — please write it here!
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
rax_facts
##### ANSIBLE VERSION
```paste below
2.9.0
```
| True | rax_facts needs to be renamed to _info and stop returning ansible_facts - ##### SUMMARY
According to the [requirements](https://docs.ansible.com/ansible/latest/dev_guide/developing_modules_checklist.html#contributing-to-ansible-objective-requirements), `_facts` modules should return information that is specific to the host the module runs on as `ansible_facts`. (See also ansible/ansible#54280 for some discussion.)
I think that this module do not satisfy this condition: it returns information that is not specific to the host machine, but bound to a cloud account. Please verify whether this is actually the case.
To fix this, the core team thinks the best approach is to rename the `_facts` modules to `_info` modules, and change the return value so that if the module is called with the `_facts` name (should be deprecated, and will then be removed in Ansible 2.13), it should return `ansible_facts` as now (and print a deprecation message), and if it is called with the `_info` name, it no longer returns `ansible_facts`, so that the result needs to be registered. This informs users that they have to change something, won't break their playbooks, and allows them (by switching to the `_info` names) to verify that their roles and playbooks no longer use the facts.
If that approach doesn't work, or you want to change the module's interface (like rename the return value(s)), it is better to deprecate the `_facts` module and create a new `_info` module next to it. To see both approaches implemented, you can take a look at https://github.com/ansible/ansible/commit/2d98734ad51c3de557c3a7d1f7166bdd114ceb0f (ansible/ansible#60178).
If you have any input — preferred method to solve this, doubts that all modules should be renamed, ... — please write it here!
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
rax_facts
##### ANSIBLE VERSION
```paste below
2.9.0
```
| main | rax facts needs to be renamed to info and stop returning ansible facts summary according to the facts modules should return information that is specific to the host the module runs on as ansible facts see also ansible ansible for some discussion i think that this module do not satisfy this condition it returns information that is not specific to the host machine but bound to a cloud account please verify whether this is actually the case to fix this the core team thinks the best approach is to rename the facts modules to info modules and change the return value so that if the module is called with the facts name should be deprecated and will then be removed in ansible it should return ansible facts as now and print a deprecation message and if it is called with the info name it no longer returns ansible facts so that the result needs to be registered this informs users that they have to change something won t break their playbooks and allows them by switching to the info names to verify that their roles and playbooks no longer use the facts if that approach doesn t work or you want to change the module s interface like rename the return value s it is better to deprecate the facts module and create a new info module next to it to see both approaches implemented you can take a look at ansible ansible if you have any input — preferred method to solve this doubts that all modules should be renamed — please write it here issue type bug report component name rax facts ansible version paste below | 1 |
965 | 4,707,894,644 | IssuesEvent | 2016-10-13 21:31:10 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Mutiple inputs to vsphere_guest module are silently ignored when launching from a template | affects_1.9 bug_report cloud vmware waiting_on_maintainer | [](https://www.bountysource.com/issues/16226901-mutiple-inputs-to-vsphere_guest-module-are-silently-ignored-when-launching-from-a-template?utm_source=16226901&utm_medium=shield&utm_campaign=ISSUE_BADGE)
#### Issue Type:
Bug Report
#### Ansible Version:
1.9.1
#### Ansible Configuration:
n/a
#### Environment:
OS X 10.10.3, vCenter 6
#### Summary:
Mutiple inputs to vsphere_guest module are ignored when launching from a template. Does not throw errors for invalid data either.
vm_nic, vm_disk & vm_hardware parameters are all ignored.
#### Steps To Reproduce:
This command creates VM with but ignores valid options. Invalid options are ignored as well.
```
- name: Create other-app VM for testing
connection: local
vsphere_guest:
vcenter_hostname: 110.120.113.333
username: foo
password: bar
guest: other-app-{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}
vm_disk:
disk1:
size_gb: 32
type: thin
datastore: foo_VMFS_1
vm_hardware:
memory_mb: 8192
num_cpus: 2
osid: ubuntu64Guest
scsi: lsi
vm_nic:
nic1:
type: vmxnet3
network: slartibartfast
network_type: dvs
vm_extra_config:
notes: this is a test VM
cluster: Ontrack-Cluster1
resource_pool: /Resources/foo-default
from_template: yes
template_src: ubuntu-14.04-template
esxi:
datacenter: foo-DC
```
#### Expected Results:
VM launched with specific NIC, disk and hardware settings. Invalid settings (such as mis-spelt NIC networks should produce an error).
#### Actual Results:
VM is launched but network, CPU and RAM that was originally specified in the template is retained.
```
<localhost> REMOTE_MODULE vsphere_guest template_src=ubuntu-14.04-template vcenter_hostname=110.120.113.333 cluster=Ontrack-Cluster1 guest=other-app-03770 password=VALUE_HIDDEN resource_pool=/Resources/foo-default username=foo
<localhost> EXEC ['/bin/sh', '-c', 'mkdir -p $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913 && chmod a+rx $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913 && echo $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913']
<localhost> PUT /var/folders/s0/jy7xj1915_bg41nmvc3cjcdw0000gn/T/tmpDWPVuq TO /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/vsphere_guest
<localhost> EXEC ['/bin/sh', '-c', u'LANG=en_US.UTF-8 LC_CTYPE=en_US.UTF-8 /usr/bin/python /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/vsphere_guest; rm -rf /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/ >/dev/null 2>&1']
changed: [localhost] => {"changed": true}
``` | True | Mutiple inputs to vsphere_guest module are silently ignored when launching from a template - [](https://www.bountysource.com/issues/16226901-mutiple-inputs-to-vsphere_guest-module-are-silently-ignored-when-launching-from-a-template?utm_source=16226901&utm_medium=shield&utm_campaign=ISSUE_BADGE)
#### Issue Type:
Bug Report
#### Ansible Version:
1.9.1
#### Ansible Configuration:
n/a
#### Environment:
OS X 10.10.3, vCenter 6
#### Summary:
Mutiple inputs to vsphere_guest module are ignored when launching from a template. Does not throw errors for invalid data either.
vm_nic, vm_disk & vm_hardware parameters are all ignored.
#### Steps To Reproduce:
This command creates VM with but ignores valid options. Invalid options are ignored as well.
```
- name: Create other-app VM for testing
connection: local
vsphere_guest:
vcenter_hostname: 110.120.113.333
username: foo
password: bar
guest: other-app-{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}{{ 9 |random}}
vm_disk:
disk1:
size_gb: 32
type: thin
datastore: foo_VMFS_1
vm_hardware:
memory_mb: 8192
num_cpus: 2
osid: ubuntu64Guest
scsi: lsi
vm_nic:
nic1:
type: vmxnet3
network: slartibartfast
network_type: dvs
vm_extra_config:
notes: this is a test VM
cluster: Ontrack-Cluster1
resource_pool: /Resources/foo-default
from_template: yes
template_src: ubuntu-14.04-template
esxi:
datacenter: foo-DC
```
#### Expected Results:
VM launched with specific NIC, disk and hardware settings. Invalid settings (such as mis-spelt NIC networks should produce an error).
#### Actual Results:
VM is launched but network, CPU and RAM that was originally specified in the template is retained.
```
<localhost> REMOTE_MODULE vsphere_guest template_src=ubuntu-14.04-template vcenter_hostname=110.120.113.333 cluster=Ontrack-Cluster1 guest=other-app-03770 password=VALUE_HIDDEN resource_pool=/Resources/foo-default username=foo
<localhost> EXEC ['/bin/sh', '-c', 'mkdir -p $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913 && chmod a+rx $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913 && echo $HOME/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913']
<localhost> PUT /var/folders/s0/jy7xj1915_bg41nmvc3cjcdw0000gn/T/tmpDWPVuq TO /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/vsphere_guest
<localhost> EXEC ['/bin/sh', '-c', u'LANG=en_US.UTF-8 LC_CTYPE=en_US.UTF-8 /usr/bin/python /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/vsphere_guest; rm -rf /Users/tpai/.ansible/tmp/ansible-tmp-1432138365.27-28886126071913/ >/dev/null 2>&1']
changed: [localhost] => {"changed": true}
``` | main | mutiple inputs to vsphere guest module are silently ignored when launching from a template issue type bug report ansible version ansible configuration n a environment os x vcenter summary mutiple inputs to vsphere guest module are ignored when launching from a template does not throw errors for invalid data either vm nic vm disk vm hardware parameters are all ignored steps to reproduce this command creates vm with but ignores valid options invalid options are ignored as well name create other app vm for testing connection local vsphere guest vcenter hostname username foo password bar guest other app random random random random random vm disk size gb type thin datastore foo vmfs vm hardware memory mb num cpus osid scsi lsi vm nic type network slartibartfast network type dvs vm extra config notes this is a test vm cluster ontrack resource pool resources foo default from template yes template src ubuntu template esxi datacenter foo dc expected results vm launched with specific nic disk and hardware settings invalid settings such as mis spelt nic networks should produce an error actual results vm is launched but network cpu and ram that was originally specified in the template is retained remote module vsphere guest template src ubuntu template vcenter hostname cluster ontrack guest other app password value hidden resource pool resources foo default username foo exec put var folders t tmpdwpvuq to users tpai ansible tmp ansible tmp vsphere guest exec changed changed true | 1 |
11,501 | 14,380,343,568 | IssuesEvent | 2020-12-02 02:34:05 | KevCor99/4a | https://api.github.com/repos/KevCor99/4a | closed | file_size_estimating_template | process-dashboard | -llenado de template de estimacion de lineas de coodigo en process dashboard
- correr el PROBE wizard | 1.0 | file_size_estimating_template - -llenado de template de estimacion de lineas de coodigo en process dashboard
- correr el PROBE wizard | non_main | file size estimating template llenado de template de estimacion de lineas de coodigo en process dashboard correr el probe wizard | 0 |
250,733 | 21,332,376,563 | IssuesEvent | 2022-04-18 10:08:36 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | opened | [CI] CCRIT.testAutoFollowing | >test-failure :Distributed/CCR | ### CI Link
https://gradle-enterprise.elastic.co/s/xqkefsonks3ni/tests/:client:rest-high-level:asyncIntegTest/org.elasticsearch.client.CCRIT/testAutoFollowing?top-execution=1
### Repro line
./gradlew ':client:rest-high-level:asyncIntegTest' --tests "org.elasticsearch.client.CCRIT.testAutoFollowing" -Dtests.seed=BD8CF40508D1FA6A -Dtests.locale=en-US -Dtests.timezone=Pacific/Saipan -Druntime.java=17
### Does it reproduce?
Yes
### Applicable branches
master
### Failure history
_No response_
### Failure excerpt
```
org.elasticsearch.client.CCRIT > testAutoFollowing FAILED
java.lang.AssertionError:
Expected: not null
but: was null
at __randomizedtesting.SeedInfo.seed([BD8CF40508D1FA6A:1262D2050C785147]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.client.CCRIT.lambda$testAutoFollowing$3(CCRIT.java:293)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1069)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1042)
at org.elasticsearch.client.CCRIT.testAutoFollowing(CCRIT.java:289)
```
The test fails locally too but with a different error:
```
org.elasticsearch.client.CCRIT > testAutoFollowing FAILED
org.elasticsearch.ElasticsearchStatusException: Elasticsearch exception [type=status_exception, reason=can not fetch remote cluster state as the license state of the remote cluster [local_cluster] could not be determined]
at __randomizedtesting.SeedInfo.seed([BD8CF40508D1FA6A:1262D2050C785147]:0)
at app//org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:178)
at app//org.elasticsearch.client.RestHighLevelClient.parseEntity(RestHighLevelClient.java:2484)
at app//org.elasticsearch.client.RestHighLevelClient.parseResponseException(RestHighLevelClient.java:2461)
at app//org.elasticsearch.client.RestHighLevelClient$1.onFailure(RestHighLevelClient.java:2373)
at app//org.elasticsearch.client.RestClient$FailureTrackingResponseListener.onDefinitiveFailure(RestClient.java:672)
at app//org.elasticsearch.client.RestClient$1.completed(RestClient.java:408)
at app//org.elasticsearch.client.RestClient$1.completed(RestClient.java:392)
at app//org.apache.http.concurrent.BasicFuture.completed(BasicFuture.java:122)
at app//org.apache.http.impl.nio.client.DefaultClientExchangeHandlerImpl.responseCompleted(DefaultClientExchangeHandlerImpl.java:181)
at app//org.apache.http.nio.protocol.HttpAsyncRequestExecutor.processResponse(HttpAsyncRequestExecutor.java:448)
at app//org.apache.http.nio.protocol.HttpAsyncRequestExecutor.inputReady(HttpAsyncRequestExecutor.java:338)
at app//org.apache.http.impl.nio.DefaultNHttpClientConnection.consumeInput(DefaultNHttpClientConnection.java:265)
at app//org.apache.http.impl.nio.client.InternalIODispatch.onInputReady(InternalIODispatch.java:81)
at app//org.apache.http.impl.nio.client.InternalIODispatch.onInputReady(InternalIODispatch.java:39)
at app//org.apache.http.impl.nio.reactor.AbstractIODispatch.inputReady(AbstractIODispatch.java:114)
at app//org.apache.http.impl.nio.reactor.BaseIOReactor.readable(BaseIOReactor.java:162)
at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.processEvent(AbstractIOReactor.java:337)
at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.processEvents(AbstractIOReactor.java:315)
at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.execute(AbstractIOReactor.java:276)
at app//org.apache.http.impl.nio.reactor.BaseIOReactor.execute(BaseIOReactor.java:104)
at app//org.apache.http.impl.nio.reactor.AbstractMultiworkerIOReactor$Worker.run(AbstractMultiworkerIOReactor.java:591)
at java.base@18/java.lang.Thread.run(Thread.java:833)
Caused by:
org.elasticsearch.ElasticsearchException: Elasticsearch exception [type=resource_not_found_exception, reason=license info is missing for cluster [local_cluster]]
at app//org.elasticsearch.ElasticsearchException.innerFromXContent(ElasticsearchException.java:490)
at app//org.elasticsearch.ElasticsearchException.fromXContent(ElasticsearchException.java:401)
at app//org.elasticsearch.ElasticsearchException.innerFromXContent(ElasticsearchException.java:431)
at app//org.elasticsearch.ElasticsearchException.failureFromXContent(ElasticsearchException.java:597)
at app//org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:170)
... 21 more
```
| 1.0 | [CI] CCRIT.testAutoFollowing - ### CI Link
https://gradle-enterprise.elastic.co/s/xqkefsonks3ni/tests/:client:rest-high-level:asyncIntegTest/org.elasticsearch.client.CCRIT/testAutoFollowing?top-execution=1
### Repro line
./gradlew ':client:rest-high-level:asyncIntegTest' --tests "org.elasticsearch.client.CCRIT.testAutoFollowing" -Dtests.seed=BD8CF40508D1FA6A -Dtests.locale=en-US -Dtests.timezone=Pacific/Saipan -Druntime.java=17
### Does it reproduce?
Yes
### Applicable branches
master
### Failure history
_No response_
### Failure excerpt
```
org.elasticsearch.client.CCRIT > testAutoFollowing FAILED
java.lang.AssertionError:
Expected: not null
but: was null
at __randomizedtesting.SeedInfo.seed([BD8CF40508D1FA6A:1262D2050C785147]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.client.CCRIT.lambda$testAutoFollowing$3(CCRIT.java:293)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1069)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1042)
at org.elasticsearch.client.CCRIT.testAutoFollowing(CCRIT.java:289)
```
The test fails locally too but with a different error:
```
org.elasticsearch.client.CCRIT > testAutoFollowing FAILED
org.elasticsearch.ElasticsearchStatusException: Elasticsearch exception [type=status_exception, reason=can not fetch remote cluster state as the license state of the remote cluster [local_cluster] could not be determined]
at __randomizedtesting.SeedInfo.seed([BD8CF40508D1FA6A:1262D2050C785147]:0)
at app//org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:178)
at app//org.elasticsearch.client.RestHighLevelClient.parseEntity(RestHighLevelClient.java:2484)
at app//org.elasticsearch.client.RestHighLevelClient.parseResponseException(RestHighLevelClient.java:2461)
at app//org.elasticsearch.client.RestHighLevelClient$1.onFailure(RestHighLevelClient.java:2373)
at app//org.elasticsearch.client.RestClient$FailureTrackingResponseListener.onDefinitiveFailure(RestClient.java:672)
at app//org.elasticsearch.client.RestClient$1.completed(RestClient.java:408)
at app//org.elasticsearch.client.RestClient$1.completed(RestClient.java:392)
at app//org.apache.http.concurrent.BasicFuture.completed(BasicFuture.java:122)
at app//org.apache.http.impl.nio.client.DefaultClientExchangeHandlerImpl.responseCompleted(DefaultClientExchangeHandlerImpl.java:181)
at app//org.apache.http.nio.protocol.HttpAsyncRequestExecutor.processResponse(HttpAsyncRequestExecutor.java:448)
at app//org.apache.http.nio.protocol.HttpAsyncRequestExecutor.inputReady(HttpAsyncRequestExecutor.java:338)
at app//org.apache.http.impl.nio.DefaultNHttpClientConnection.consumeInput(DefaultNHttpClientConnection.java:265)
at app//org.apache.http.impl.nio.client.InternalIODispatch.onInputReady(InternalIODispatch.java:81)
at app//org.apache.http.impl.nio.client.InternalIODispatch.onInputReady(InternalIODispatch.java:39)
at app//org.apache.http.impl.nio.reactor.AbstractIODispatch.inputReady(AbstractIODispatch.java:114)
at app//org.apache.http.impl.nio.reactor.BaseIOReactor.readable(BaseIOReactor.java:162)
at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.processEvent(AbstractIOReactor.java:337)
at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.processEvents(AbstractIOReactor.java:315)
at app//org.apache.http.impl.nio.reactor.AbstractIOReactor.execute(AbstractIOReactor.java:276)
at app//org.apache.http.impl.nio.reactor.BaseIOReactor.execute(BaseIOReactor.java:104)
at app//org.apache.http.impl.nio.reactor.AbstractMultiworkerIOReactor$Worker.run(AbstractMultiworkerIOReactor.java:591)
at java.base@18/java.lang.Thread.run(Thread.java:833)
Caused by:
org.elasticsearch.ElasticsearchException: Elasticsearch exception [type=resource_not_found_exception, reason=license info is missing for cluster [local_cluster]]
at app//org.elasticsearch.ElasticsearchException.innerFromXContent(ElasticsearchException.java:490)
at app//org.elasticsearch.ElasticsearchException.fromXContent(ElasticsearchException.java:401)
at app//org.elasticsearch.ElasticsearchException.innerFromXContent(ElasticsearchException.java:431)
at app//org.elasticsearch.ElasticsearchException.failureFromXContent(ElasticsearchException.java:597)
at app//org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:170)
... 21 more
```
| non_main | ccrit testautofollowing ci link repro line gradlew client rest high level asyncintegtest tests org elasticsearch client ccrit testautofollowing dtests seed dtests locale en us dtests timezone pacific saipan druntime java does it reproduce yes applicable branches master failure history no response failure excerpt org elasticsearch client ccrit testautofollowing failed java lang assertionerror expected not null but was null at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch client ccrit lambda testautofollowing ccrit java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch client ccrit testautofollowing ccrit java the test fails locally too but with a different error org elasticsearch client ccrit testautofollowing failed org elasticsearch elasticsearchstatusexception elasticsearch exception could not be determined at randomizedtesting seedinfo seed at app org elasticsearch rest bytesrestresponse errorfromxcontent bytesrestresponse java at app org elasticsearch client resthighlevelclient parseentity resthighlevelclient java at app org elasticsearch client resthighlevelclient parseresponseexception resthighlevelclient java at app org elasticsearch client resthighlevelclient onfailure resthighlevelclient java at app org elasticsearch client restclient failuretrackingresponselistener ondefinitivefailure restclient java at app org elasticsearch client restclient completed restclient java at app org elasticsearch client restclient completed restclient java at app org apache http concurrent basicfuture completed basicfuture java at app org apache http impl nio client defaultclientexchangehandlerimpl responsecompleted defaultclientexchangehandlerimpl java at app org apache http nio protocol httpasyncrequestexecutor processresponse httpasyncrequestexecutor java at app org apache http nio protocol httpasyncrequestexecutor inputready httpasyncrequestexecutor java at app org apache http impl nio defaultnhttpclientconnection consumeinput defaultnhttpclientconnection java at app org apache http impl nio client internaliodispatch oninputready internaliodispatch java at app org apache http impl nio client internaliodispatch oninputready internaliodispatch java at app org apache http impl nio reactor abstractiodispatch inputready abstractiodispatch java at app org apache http impl nio reactor baseioreactor readable baseioreactor java at app org apache http impl nio reactor abstractioreactor processevent abstractioreactor java at app org apache http impl nio reactor abstractioreactor processevents abstractioreactor java at app org apache http impl nio reactor abstractioreactor execute abstractioreactor java at app org apache http impl nio reactor baseioreactor execute baseioreactor java at app org apache http impl nio reactor abstractmultiworkerioreactor worker run abstractmultiworkerioreactor java at java base java lang thread run thread java caused by org elasticsearch elasticsearchexception elasticsearch exception at app org elasticsearch elasticsearchexception innerfromxcontent elasticsearchexception java at app org elasticsearch elasticsearchexception fromxcontent elasticsearchexception java at app org elasticsearch elasticsearchexception innerfromxcontent elasticsearchexception java at app org elasticsearch elasticsearchexception failurefromxcontent elasticsearchexception java at app org elasticsearch rest bytesrestresponse errorfromxcontent bytesrestresponse java more | 0 |
111,623 | 17,030,479,286 | IssuesEvent | 2021-07-04 13:07:13 | turkdevops/next-auth | https://api.github.com/repos/turkdevops/next-auth | closed | CVE-2021-28092 (High) detected in is-svg-3.0.0.tgz - autoclosed | security vulnerability | ## CVE-2021-28092 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary>
<p>Check if a string or buffer is SVG</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p>
<p>Path to dependency file: next-auth/package.json</p>
<p>Path to vulnerable library: next-auth/node_modules/is-svg/package.json,next-auth/www/node_modules/is-svg/package.json</p>
<p>
Dependency Hierarchy:
- cssnano-4.1.10.tgz (Root Library)
- cssnano-preset-default-4.0.7.tgz
- postcss-svgo-4.0.2.tgz
- :x: **is-svg-3.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/next-auth/commit/d96c3a73550059033f41882dbe71f9064e527688">d96c3a73550059033f41882dbe71f9064e527688</a></p>
<p>Found in base branch: <b>canary</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The is-svg package 2.1.0 through 4.2.1 for Node.js uses a regular expression that is vulnerable to Regular Expression Denial of Service (ReDoS). If an attacker provides a malicious string, is-svg will get stuck processing the input for a very long time.
<p>Publish Date: 2021-03-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092>CVE-2021-28092</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092</a></p>
<p>Release Date: 2021-03-12</p>
<p>Fix Resolution: v4.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-28092 (High) detected in is-svg-3.0.0.tgz - autoclosed - ## CVE-2021-28092 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary>
<p>Check if a string or buffer is SVG</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p>
<p>Path to dependency file: next-auth/package.json</p>
<p>Path to vulnerable library: next-auth/node_modules/is-svg/package.json,next-auth/www/node_modules/is-svg/package.json</p>
<p>
Dependency Hierarchy:
- cssnano-4.1.10.tgz (Root Library)
- cssnano-preset-default-4.0.7.tgz
- postcss-svgo-4.0.2.tgz
- :x: **is-svg-3.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/next-auth/commit/d96c3a73550059033f41882dbe71f9064e527688">d96c3a73550059033f41882dbe71f9064e527688</a></p>
<p>Found in base branch: <b>canary</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The is-svg package 2.1.0 through 4.2.1 for Node.js uses a regular expression that is vulnerable to Regular Expression Denial of Service (ReDoS). If an attacker provides a malicious string, is-svg will get stuck processing the input for a very long time.
<p>Publish Date: 2021-03-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092>CVE-2021-28092</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092</a></p>
<p>Release Date: 2021-03-12</p>
<p>Fix Resolution: v4.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in is svg tgz autoclosed cve high severity vulnerability vulnerable library is svg tgz check if a string or buffer is svg library home page a href path to dependency file next auth package json path to vulnerable library next auth node modules is svg package json next auth www node modules is svg package json dependency hierarchy cssnano tgz root library cssnano preset default tgz postcss svgo tgz x is svg tgz vulnerable library found in head commit a href found in base branch canary vulnerability details the is svg package through for node js uses a regular expression that is vulnerable to regular expression denial of service redos if an attacker provides a malicious string is svg will get stuck processing the input for a very long time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
4,180 | 20,112,696,646 | IssuesEvent | 2022-02-07 16:25:36 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [a11y]: <AVT1> The 'id' "downshift-8-label" specified for the ARIA property 'aria-labelledby' value is not valid | type: a11y ♿ component: dropdown status: waiting for maintainer response 💬 | ### Package
carbon-components-react
### Browser
Chrome
### Operating System
MacOS
### Package version
v7.50.0
### React version
v16.2.0
### Automated testing tool and ruleset
IBM Equal Access Accessibility Checker
### Assistive technology
_No response_
### Description

Issue can be reproduced in https://react.carbondesignsystem.com/?path=/docs/components-dropdown--default
Issue reproduced in code sandbox

Originally issue was opened in dns-ui
https://github.ibm.com/NetworkTribe/dns-ui/issues/1521
### WCAG 2.1 Violation
_No response_
### CodeSandbox example
https://codesandbox.io/s/5xrmq?file=/index.js
### Steps to reproduce
1. Render a dropdown with few items
2. Inspect on dropdown menu
3. You should see the "'id' specified for the ARIA property 'aria-labelledby' value is not valid"
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | True | [a11y]: <AVT1> The 'id' "downshift-8-label" specified for the ARIA property 'aria-labelledby' value is not valid - ### Package
carbon-components-react
### Browser
Chrome
### Operating System
MacOS
### Package version
v7.50.0
### React version
v16.2.0
### Automated testing tool and ruleset
IBM Equal Access Accessibility Checker
### Assistive technology
_No response_
### Description

Issue can be reproduced in https://react.carbondesignsystem.com/?path=/docs/components-dropdown--default
Issue reproduced in code sandbox

Originally issue was opened in dns-ui
https://github.ibm.com/NetworkTribe/dns-ui/issues/1521
### WCAG 2.1 Violation
_No response_
### CodeSandbox example
https://codesandbox.io/s/5xrmq?file=/index.js
### Steps to reproduce
1. Render a dropdown with few items
2. Inspect on dropdown menu
3. You should see the "'id' specified for the ARIA property 'aria-labelledby' value is not valid"
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | main | the id downshift label specified for the aria property aria labelledby value is not valid package carbon components react browser chrome operating system macos package version react version automated testing tool and ruleset ibm equal access accessibility checker assistive technology no response description issue can be reproduced in issue reproduced in code sandbox originally issue was opened in dns ui wcag violation no response codesandbox example steps to reproduce render a dropdown with few items inspect on dropdown menu you should see the id specified for the aria property aria labelledby value is not valid code of conduct i agree to follow this project s i checked the for duplicate problems | 1 |
203,906 | 15,890,724,780 | IssuesEvent | 2021-04-10 16:28:27 | veg-share/frontend-ui | https://api.github.com/repos/veg-share/frontend-ui | closed | Dev Issue : Initial SetUp | Developer Phase 1 documentation set-up | **What is the issue?**
> Set up initial build of React app
install updated dependencies
establish local/remote repository connections for all members
-[ ] https://www.npmjs.com/package/react-modal
| 1.0 | Dev Issue : Initial SetUp - **What is the issue?**
> Set up initial build of React app
install updated dependencies
establish local/remote repository connections for all members
-[ ] https://www.npmjs.com/package/react-modal
| non_main | dev issue initial setup what is the issue set up initial build of react app install updated dependencies establish local remote repository connections for all members | 0 |
17,763 | 10,765,919,647 | IssuesEvent | 2019-11-01 12:26:48 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Can we change the service principal id/secret on agent nodes at runtime? | Pri1 container-service/svc cxp product-question triaged | Hi, let's say I specified a service principal "A" when creating my AKS cluster, and as described above the service principal info will be stored in file "/etc/kubernetes/azure.json" on agent node VMs. Due to some security concern, can I change the "azure.json" file to set another service principal "B" on agent node VMs only without changing the whole service principal used by AKS (which means not use "az aks update-credentials" command).
We did some experiment that it works, but we'd also like to get confirmation from AKS production team to see whether this is an official supported approach and whether there is any side-effect of doing so. Thanks.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 25a0d502-ce71-76d4-bf8a-060e63913b52
* Version Independent ID: 69d7fa26-1a37-e477-0a2c-db58ce19a84f
* Content: [Service principals for Azure Kubernetes Services (AKS)](https://docs.microsoft.com/en-us/azure/aks/kubernetes-service-principal)
* Content Source: [articles/aks/kubernetes-service-principal.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/kubernetes-service-principal.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | 1.0 | Can we change the service principal id/secret on agent nodes at runtime? - Hi, let's say I specified a service principal "A" when creating my AKS cluster, and as described above the service principal info will be stored in file "/etc/kubernetes/azure.json" on agent node VMs. Due to some security concern, can I change the "azure.json" file to set another service principal "B" on agent node VMs only without changing the whole service principal used by AKS (which means not use "az aks update-credentials" command).
We did some experiment that it works, but we'd also like to get confirmation from AKS production team to see whether this is an official supported approach and whether there is any side-effect of doing so. Thanks.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 25a0d502-ce71-76d4-bf8a-060e63913b52
* Version Independent ID: 69d7fa26-1a37-e477-0a2c-db58ce19a84f
* Content: [Service principals for Azure Kubernetes Services (AKS)](https://docs.microsoft.com/en-us/azure/aks/kubernetes-service-principal)
* Content Source: [articles/aks/kubernetes-service-principal.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/kubernetes-service-principal.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | non_main | can we change the service principal id secret on agent nodes at runtime hi let s say i specified a service principal a when creating my aks cluster and as described above the service principal info will be stored in file etc kubernetes azure json on agent node vms due to some security concern can i change the azure json file to set another service principal b on agent node vms only without changing the whole service principal used by aks which means not use az aks update credentials command we did some experiment that it works but we d also like to get confirmation from aks production team to see whether this is an official supported approach and whether there is any side effect of doing so thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned | 0 |
144,545 | 19,287,858,868 | IssuesEvent | 2021-12-11 08:53:39 | turkdevops/electron | https://api.github.com/repos/turkdevops/electron | closed | CVE-2012-6708 (Medium) detected in jquery-1.7.2.min.js - autoclosed | security vulnerability | ## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: electron/node_modules/jmespath/index.html</p>
<p>Path to vulnerable library: /node_modules/jmespath/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/electron/commit/5da3bcb84e69be57f16313e78dd435a9af067d0d">5da3bcb84e69be57f16313e78dd435a9af067d0d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2012-6708 (Medium) detected in jquery-1.7.2.min.js - autoclosed - ## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: electron/node_modules/jmespath/index.html</p>
<p>Path to vulnerable library: /node_modules/jmespath/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/electron/commit/5da3bcb84e69be57f16313e78dd435a9af067d0d">5da3bcb84e69be57f16313e78dd435a9af067d0d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in jquery min js autoclosed cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file electron node modules jmespath index html path to vulnerable library node modules jmespath index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
78,822 | 10,090,549,621 | IssuesEvent | 2019-07-26 12:00:12 | XanaduAI/pennylane | https://api.github.com/repos/XanaduAI/pennylane | closed | QubitUnitary does not act on all wires | documentation | #### Issue description
Description of the issue - In the documentation, QubitUnitary operator:
1. is stated to act on `all wires` in the system. I tested it and it can act on subsytem of wires.
2. is said to support finite difference as its gradient recipe. This can not be right as we can only input a fixed, non-differentiable matrix to this operator. So the gradient recipe should be `None`.
I am making a PR for this issue. | 1.0 | QubitUnitary does not act on all wires - #### Issue description
Description of the issue - In the documentation, QubitUnitary operator:
1. is stated to act on `all wires` in the system. I tested it and it can act on subsytem of wires.
2. is said to support finite difference as its gradient recipe. This can not be right as we can only input a fixed, non-differentiable matrix to this operator. So the gradient recipe should be `None`.
I am making a PR for this issue. | non_main | qubitunitary does not act on all wires issue description description of the issue in the documentation qubitunitary operator is stated to act on all wires in the system i tested it and it can act on subsytem of wires is said to support finite difference as its gradient recipe this can not be right as we can only input a fixed non differentiable matrix to this operator so the gradient recipe should be none i am making a pr for this issue | 0 |
220,760 | 17,259,335,804 | IssuesEvent | 2021-07-22 04:08:26 | DeFiCh/jellyfish | https://api.github.com/repos/DeFiCh/jellyfish | closed | `MasterNodeRegTestContainer` block gen feature should be controlled by testcontainers | area/testcontainers kind/refactor priority/important-soon triage/accepted | <!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
`MasterNodeRegTestContainer` block generation should be automated by testcontainers. `@defichain/testcontainers` should be created such that block generation can only be done by code within testcontainers. `gen=1` was set to automated the generation process on ain side.
#### Why is this needed:
Having automated block gen on the ain side is unpredictable and creates too many conditions for "side-effect" when testing on jellyfish. This effect is cascaded down to whale, playground and potentially wallet implementation.
#### Additional comments:
https://github.com/DeFiCh/playground uses its own block generation interval, we can follow that implementation reference.
https://github.com/DeFiCh/playground/blob/09b29fe1d47f48ac75283ecdda85c234c4bf94c2/src/module.playground/playground.block.ts#L13-L17 | 1.0 | `MasterNodeRegTestContainer` block gen feature should be controlled by testcontainers - <!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
`MasterNodeRegTestContainer` block generation should be automated by testcontainers. `@defichain/testcontainers` should be created such that block generation can only be done by code within testcontainers. `gen=1` was set to automated the generation process on ain side.
#### Why is this needed:
Having automated block gen on the ain side is unpredictable and creates too many conditions for "side-effect" when testing on jellyfish. This effect is cascaded down to whale, playground and potentially wallet implementation.
#### Additional comments:
https://github.com/DeFiCh/playground uses its own block generation interval, we can follow that implementation reference.
https://github.com/DeFiCh/playground/blob/09b29fe1d47f48ac75283ecdda85c234c4bf94c2/src/module.playground/playground.block.ts#L13-L17 | non_main | masternoderegtestcontainer block gen feature should be controlled by testcontainers what would you like to be added masternoderegtestcontainer block generation should be automated by testcontainers defichain testcontainers should be created such that block generation can only be done by code within testcontainers gen was set to automated the generation process on ain side why is this needed having automated block gen on the ain side is unpredictable and creates too many conditions for side effect when testing on jellyfish this effect is cascaded down to whale playground and potentially wallet implementation additional comments uses its own block generation interval we can follow that implementation reference | 0 |
769,335 | 27,002,020,038 | IssuesEvent | 2023-02-10 08:40:16 | shaka-project/shaka-player | https://api.github.com/repos/shaka-project/shaka-player | closed | DASH unprotected content stuck on the first frame on Samsung Tizen 5.5 | type: bug flag: seeking PR status: waiting on response priority: P2 component: DASH platform: Tizen | @joeyparrish I have a potentially similar issue, both to #3046 and to #2620. Basically in a few of several attempts, the DASH stream fails to start properly, it's just stuck on the first frame, but can be initiated if the user seeks forward.
From the logs, I can see that the Shaka player has detected a stall, and tries to unstall by pause/play, but with no success.
Only reproducible on Tizen 5.5 (2020). I've been testing on model: UE43TU7072UXXH, but it's reproducible on other 5.5 models as well.
Shaka player version: 3.1.2
The stream where this can be reproduced (not consistently): https://trailers.redbox.com/assets/211568/dash/211568.mpd
And it's not specific to this DASH stream, can be reproduced with others as well, I think that it's happening a bit more often with it.
Logs:
<details>
<summary>Details</summary>


</details>
Shaka player config:
```js
{
streaming: {
ignoreTextStreamFailures: true,
// https://shaka-player-demo.appspot.com/docs/api/tutorial-network-and-buffering-config.html
bufferingGoal: 60,
// how much to buffer while the video is playing, to have enough content in case of network hiccups
rebufferingGoal: 3,
// how much to buffer before resuming playing, aka on video start and after seeking video
bufferBehind: 30,
jumpLargeGaps: true,
stallSkip: 0.1,
retryParameters: {
maxAttempts: 10,
baseDelay: 1000,
backoffFactor: 1.75,
fuzzFactor: 0.5,
timeout: 10000,
},
},
manifest: {
dash: {
ignoreMinBufferTime: true,
},
retryParameters: {
maxAttempts: 5,
baseDelay: 1000,
backoffFactor: 2,
fuzzFactor: 0.5,
timeout: 25000,
},
},
drm: {
servers: getDrmLicenseServers(),
advanced: {
'com.widevine.alpha': {
videoRobustness: 'SW_SECURE_CRYPTO',
audioRobustness: 'SW_SECURE_CRYPTO',
},
'com.microsoft.playready': withAltTVODDrmConfig
? {
distinctiveIdentifierRequired: true,
sessionType: 'persistent-license',
}
: {},
},
retryParameters: {
maxAttempts: 5,
baseDelay: 1000,
backoffFactor: 2,
fuzzFactor: 0.5,
timeout: 10000,
},
},
restrictions: {
minHeight: 360,
},
preferredAudioChannelCount: 6,
}
```
I've also tried playing with the default configuration instead of the one above, and the issue persists. The content sometimes starts playing after some time, but I could not reproduce this consistently. Also, looks like there is a lot of calls to switch after the shaka player loads the content:
```log
13:01:41.638 mathRound.install
13:01:41.642 MediaSource.install
13:01:41.645 Using native MSE as-is.
13:01:41.649 VideoPlayPromise.install
13:01:41.652 Using native VTTCue.
13:01:41.656 MediaCapabilities: install
13:01:41.659 MediaCapabilities: Native mediaCapabilities support found.
13:01:42.060 mathRound.install
13:01:42.073 MediaSource.install
13:01:42.076 Using native MSE as-is.
13:01:42.079 VideoPlayPromise.install
13:01:42.081 Using native VTTCue.
13:01:42.084 MediaCapabilities: install
13:01:42.087 MediaCapabilities: Native mediaCapabilities support found.
13:01:43.912 Starting attach...
13:01:45.306 Starting load of https://trailers.redbox.com/assets/212464/dash/212464.mpd...
13:01:45.433 Found variant with audio and video content, so filtering out audio-only content.
13:01:45.482 codecs avc1-mp4a avg bandwidth 8596000
13:01:45.543 init: completed initial Stream setup
13:01:45.567 After load: https://trailers.redbox.com/assets/212464/dash/212464.mpd
13:01:52.201 Calling switch_(), bandwidth=1048 kbps
13:01:52.210 switch_
13:01:52.218 switch: switching to Stream (video:6)
13:01:52.233 switch: Stream (audio:8) already active
13:02:00.656 Calling switch_(), bandwidth=826 kbps
13:02:00.667 switch_
13:02:09.045 Calling switch_(), bandwidth=987 kbps
13:02:09.057 switch_
13:02:17.317 Calling switch_(), bandwidth=951 kbps
13:02:17.329 switch_
13:02:25.496 Calling switch_(), bandwidth=934 kbps
13:02:25.507 switch_
13:02:34.148 Calling switch_(), bandwidth=891 kbps
13:02:34.158 switch_
13:02:42.549 Calling switch_(), bandwidth=898 kbps
13:02:42.560 switch_
13:02:51.073 Calling switch_(), bandwidth=891 kbps
13:02:51.086 switch_
13:02:59.432 Calling switch_(), bandwidth=856 kbps
13:02:59.443 switch_
13:03:07.906 Calling switch_(), bandwidth=844 kbps
13:03:07.917 switch_
13:03:16.035 Calling switch_(), bandwidth=819 kbps
13:03:16.047 switch_
13:03:24.335 Calling switch_(), bandwidth=850 kbps
13:03:24.348 switch_
13:03:33.144 Calling switch_(), bandwidth=822 kbps
13:03:33.157 switch_
13:03:41.333 Calling switch_(), bandwidth=807 kbps
13:03:41.348 switch_
13:03:49.476 Calling switch_(), bandwidth=808 kbps
13:03:49.489 switch_
13:03:57.587 Calling switch_(), bandwidth=808 kbps
13:03:57.599 switch_
13:04:05.866 Calling switch_(), bandwidth=796 kbps
13:04:05.876 switch_
13:04:14.165 Calling switch_(), bandwidth=747 kbps
13:04:14.176 switch_
13:04:22.345 Calling switch_(), bandwidth=731 kbps
13:04:22.357 switch_
13:04:30.521 Calling switch_(), bandwidth=732 kbps
13:04:30.531 switch_
13:04:39.179 Calling switch_(), bandwidth=706 kbps
13:04:39.196 switch_
13:04:47.848 Calling switch_(), bandwidth=659 kbps
13:04:47.863 switch_
13:04:56.334 Calling switch_(), bandwidth=699 kbps
13:04:56.344 switch_
13:05:04.836 Calling switch_(), bandwidth=700 kbps
13:05:04.846 switch_
13:05:13.258 Calling switch_(), bandwidth=715 kbps
13:05:13.267 switch_
13:05:21.740 Calling switch_(), bandwidth=662 kbps
13:05:21.750 switch_
13:05:30.741 Calling switch_(), bandwidth=704 kbps
13:05:30.749 switch_
13:05:39.415 Calling switch_(), bandwidth=694 kbps
13:05:39.424 switch_
13:05:47.876 Calling switch_(), bandwidth=655 kbps
13:05:47.885 switch_
13:05:56.757 Calling switch_(), bandwidth=631 kbps
13:05:56.768 switch_
13:06:05.725 Calling switch_(), bandwidth=656 kbps
13:06:05.734 switch_
13:06:14.687 Calling switch_(), bandwidth=664 kbps
13:06:14.697 switch_
13:06:23.384 Calling switch_(), bandwidth=645 kbps
13:06:23.394 switch_
13:06:31.451 Calling switch_(), bandwidth=620 kbps
13:06:31.459 switch_
13:06:40.143 Calling switch_(), bandwidth=625 kbps
13:06:40.152 switch_
13:06:48.708 Calling switch_(), bandwidth=638 kbps
13:06:48.719 switch_
13:06:57.602 Calling switch_(), bandwidth=625 kbps
13:06:57.613 switch_
13:07:05.908 Calling switch_(), bandwidth=600 kbps
13:07:05.918 switch_
13:07:14.040 Calling switch_(), bandwidth=623 kbps
13:07:14.053 switch_
13:07:22.273 Calling switch_(), bandwidth=622 kbps
13:07:22.282 switch_
13:07:30.844 Calling switch_(), bandwidth=547 kbps
13:07:30.855 switch_
13:07:39.932 Calling switch_(), bandwidth=550 kbps
13:07:39.944 switch_
13:07:48.012 Calling switch_(), bandwidth=543 kbps
13:07:48.024 switch_
13:07:56.416 Calling switch_(), bandwidth=584 kbps
13:07:56.428 switch_
13:08:05.005 Calling switch_(), bandwidth=595 kbps
13:08:05.017 switch_
13:08:13.812 Calling switch_(), bandwidth=534 kbps
13:08:13.829 switch_
13:08:20.219 Jumping forward 0.033 seconds because of gap before start time of 0.033
13:08:21.137 (all): seeked: buffered seek: presentationTime=0.033
13:08:25.570 Jumping forward 0.033 seconds because of gap before start time of 0.033
13:08:25.596 (all): seeked: buffered seek: presentationTime=0.033
13:08:27.965 Calling switch_(), bandwidth=534 kbps
13:08:27.972 switch_
```
In the log above, the playback started after 6 minutes... Because of this, I would say that this is also related to #3076.
Any suggestion on how to debug/fix this would be much appreciated.
_Originally posted by @Puritanic in https://github.com/google/shaka-player/issues/3046#issuecomment-901113523_ | 1.0 | DASH unprotected content stuck on the first frame on Samsung Tizen 5.5 - @joeyparrish I have a potentially similar issue, both to #3046 and to #2620. Basically in a few of several attempts, the DASH stream fails to start properly, it's just stuck on the first frame, but can be initiated if the user seeks forward.
From the logs, I can see that the Shaka player has detected a stall, and tries to unstall by pause/play, but with no success.
Only reproducible on Tizen 5.5 (2020). I've been testing on model: UE43TU7072UXXH, but it's reproducible on other 5.5 models as well.
Shaka player version: 3.1.2
The stream where this can be reproduced (not consistently): https://trailers.redbox.com/assets/211568/dash/211568.mpd
And it's not specific to this DASH stream, can be reproduced with others as well, I think that it's happening a bit more often with it.
Logs:
<details>
<summary>Details</summary>


</details>
Shaka player config:
```js
{
streaming: {
ignoreTextStreamFailures: true,
// https://shaka-player-demo.appspot.com/docs/api/tutorial-network-and-buffering-config.html
bufferingGoal: 60,
// how much to buffer while the video is playing, to have enough content in case of network hiccups
rebufferingGoal: 3,
// how much to buffer before resuming playing, aka on video start and after seeking video
bufferBehind: 30,
jumpLargeGaps: true,
stallSkip: 0.1,
retryParameters: {
maxAttempts: 10,
baseDelay: 1000,
backoffFactor: 1.75,
fuzzFactor: 0.5,
timeout: 10000,
},
},
manifest: {
dash: {
ignoreMinBufferTime: true,
},
retryParameters: {
maxAttempts: 5,
baseDelay: 1000,
backoffFactor: 2,
fuzzFactor: 0.5,
timeout: 25000,
},
},
drm: {
servers: getDrmLicenseServers(),
advanced: {
'com.widevine.alpha': {
videoRobustness: 'SW_SECURE_CRYPTO',
audioRobustness: 'SW_SECURE_CRYPTO',
},
'com.microsoft.playready': withAltTVODDrmConfig
? {
distinctiveIdentifierRequired: true,
sessionType: 'persistent-license',
}
: {},
},
retryParameters: {
maxAttempts: 5,
baseDelay: 1000,
backoffFactor: 2,
fuzzFactor: 0.5,
timeout: 10000,
},
},
restrictions: {
minHeight: 360,
},
preferredAudioChannelCount: 6,
}
```
I've also tried playing with the default configuration instead of the one above, and the issue persists. The content sometimes starts playing after some time, but I could not reproduce this consistently. Also, looks like there is a lot of calls to switch after the shaka player loads the content:
```log
13:01:41.638 mathRound.install
13:01:41.642 MediaSource.install
13:01:41.645 Using native MSE as-is.
13:01:41.649 VideoPlayPromise.install
13:01:41.652 Using native VTTCue.
13:01:41.656 MediaCapabilities: install
13:01:41.659 MediaCapabilities: Native mediaCapabilities support found.
13:01:42.060 mathRound.install
13:01:42.073 MediaSource.install
13:01:42.076 Using native MSE as-is.
13:01:42.079 VideoPlayPromise.install
13:01:42.081 Using native VTTCue.
13:01:42.084 MediaCapabilities: install
13:01:42.087 MediaCapabilities: Native mediaCapabilities support found.
13:01:43.912 Starting attach...
13:01:45.306 Starting load of https://trailers.redbox.com/assets/212464/dash/212464.mpd...
13:01:45.433 Found variant with audio and video content, so filtering out audio-only content.
13:01:45.482 codecs avc1-mp4a avg bandwidth 8596000
13:01:45.543 init: completed initial Stream setup
13:01:45.567 After load: https://trailers.redbox.com/assets/212464/dash/212464.mpd
13:01:52.201 Calling switch_(), bandwidth=1048 kbps
13:01:52.210 switch_
13:01:52.218 switch: switching to Stream (video:6)
13:01:52.233 switch: Stream (audio:8) already active
13:02:00.656 Calling switch_(), bandwidth=826 kbps
13:02:00.667 switch_
13:02:09.045 Calling switch_(), bandwidth=987 kbps
13:02:09.057 switch_
13:02:17.317 Calling switch_(), bandwidth=951 kbps
13:02:17.329 switch_
13:02:25.496 Calling switch_(), bandwidth=934 kbps
13:02:25.507 switch_
13:02:34.148 Calling switch_(), bandwidth=891 kbps
13:02:34.158 switch_
13:02:42.549 Calling switch_(), bandwidth=898 kbps
13:02:42.560 switch_
13:02:51.073 Calling switch_(), bandwidth=891 kbps
13:02:51.086 switch_
13:02:59.432 Calling switch_(), bandwidth=856 kbps
13:02:59.443 switch_
13:03:07.906 Calling switch_(), bandwidth=844 kbps
13:03:07.917 switch_
13:03:16.035 Calling switch_(), bandwidth=819 kbps
13:03:16.047 switch_
13:03:24.335 Calling switch_(), bandwidth=850 kbps
13:03:24.348 switch_
13:03:33.144 Calling switch_(), bandwidth=822 kbps
13:03:33.157 switch_
13:03:41.333 Calling switch_(), bandwidth=807 kbps
13:03:41.348 switch_
13:03:49.476 Calling switch_(), bandwidth=808 kbps
13:03:49.489 switch_
13:03:57.587 Calling switch_(), bandwidth=808 kbps
13:03:57.599 switch_
13:04:05.866 Calling switch_(), bandwidth=796 kbps
13:04:05.876 switch_
13:04:14.165 Calling switch_(), bandwidth=747 kbps
13:04:14.176 switch_
13:04:22.345 Calling switch_(), bandwidth=731 kbps
13:04:22.357 switch_
13:04:30.521 Calling switch_(), bandwidth=732 kbps
13:04:30.531 switch_
13:04:39.179 Calling switch_(), bandwidth=706 kbps
13:04:39.196 switch_
13:04:47.848 Calling switch_(), bandwidth=659 kbps
13:04:47.863 switch_
13:04:56.334 Calling switch_(), bandwidth=699 kbps
13:04:56.344 switch_
13:05:04.836 Calling switch_(), bandwidth=700 kbps
13:05:04.846 switch_
13:05:13.258 Calling switch_(), bandwidth=715 kbps
13:05:13.267 switch_
13:05:21.740 Calling switch_(), bandwidth=662 kbps
13:05:21.750 switch_
13:05:30.741 Calling switch_(), bandwidth=704 kbps
13:05:30.749 switch_
13:05:39.415 Calling switch_(), bandwidth=694 kbps
13:05:39.424 switch_
13:05:47.876 Calling switch_(), bandwidth=655 kbps
13:05:47.885 switch_
13:05:56.757 Calling switch_(), bandwidth=631 kbps
13:05:56.768 switch_
13:06:05.725 Calling switch_(), bandwidth=656 kbps
13:06:05.734 switch_
13:06:14.687 Calling switch_(), bandwidth=664 kbps
13:06:14.697 switch_
13:06:23.384 Calling switch_(), bandwidth=645 kbps
13:06:23.394 switch_
13:06:31.451 Calling switch_(), bandwidth=620 kbps
13:06:31.459 switch_
13:06:40.143 Calling switch_(), bandwidth=625 kbps
13:06:40.152 switch_
13:06:48.708 Calling switch_(), bandwidth=638 kbps
13:06:48.719 switch_
13:06:57.602 Calling switch_(), bandwidth=625 kbps
13:06:57.613 switch_
13:07:05.908 Calling switch_(), bandwidth=600 kbps
13:07:05.918 switch_
13:07:14.040 Calling switch_(), bandwidth=623 kbps
13:07:14.053 switch_
13:07:22.273 Calling switch_(), bandwidth=622 kbps
13:07:22.282 switch_
13:07:30.844 Calling switch_(), bandwidth=547 kbps
13:07:30.855 switch_
13:07:39.932 Calling switch_(), bandwidth=550 kbps
13:07:39.944 switch_
13:07:48.012 Calling switch_(), bandwidth=543 kbps
13:07:48.024 switch_
13:07:56.416 Calling switch_(), bandwidth=584 kbps
13:07:56.428 switch_
13:08:05.005 Calling switch_(), bandwidth=595 kbps
13:08:05.017 switch_
13:08:13.812 Calling switch_(), bandwidth=534 kbps
13:08:13.829 switch_
13:08:20.219 Jumping forward 0.033 seconds because of gap before start time of 0.033
13:08:21.137 (all): seeked: buffered seek: presentationTime=0.033
13:08:25.570 Jumping forward 0.033 seconds because of gap before start time of 0.033
13:08:25.596 (all): seeked: buffered seek: presentationTime=0.033
13:08:27.965 Calling switch_(), bandwidth=534 kbps
13:08:27.972 switch_
```
In the log above, the playback started after 6 minutes... Because of this, I would say that this is also related to #3076.
Any suggestion on how to debug/fix this would be much appreciated.
_Originally posted by @Puritanic in https://github.com/google/shaka-player/issues/3046#issuecomment-901113523_ | non_main | dash unprotected content stuck on the first frame on samsung tizen joeyparrish i have a potentially similar issue both to and to basically in a few of several attempts the dash stream fails to start properly it s just stuck on the first frame but can be initiated if the user seeks forward from the logs i can see that the shaka player has detected a stall and tries to unstall by pause play but with no success only reproducible on tizen i ve been testing on model but it s reproducible on other models as well shaka player version the stream where this can be reproduced not consistently and it s not specific to this dash stream can be reproduced with others as well i think that it s happening a bit more often with it logs details shaka player config js streaming ignoretextstreamfailures true bufferinggoal how much to buffer while the video is playing to have enough content in case of network hiccups rebufferinggoal how much to buffer before resuming playing aka on video start and after seeking video bufferbehind jumplargegaps true stallskip retryparameters maxattempts basedelay backofffactor fuzzfactor timeout manifest dash ignoreminbuffertime true retryparameters maxattempts basedelay backofffactor fuzzfactor timeout drm servers getdrmlicenseservers advanced com widevine alpha videorobustness sw secure crypto audiorobustness sw secure crypto com microsoft playready withalttvoddrmconfig distinctiveidentifierrequired true sessiontype persistent license retryparameters maxattempts basedelay backofffactor fuzzfactor timeout restrictions minheight preferredaudiochannelcount i ve also tried playing with the default configuration instead of the one above and the issue persists the content sometimes starts playing after some time but i could not reproduce this consistently also looks like there is a lot of calls to switch after the shaka player loads the content log mathround install mediasource install using native mse as is videoplaypromise install using native vttcue mediacapabilities install mediacapabilities native mediacapabilities support found mathround install mediasource install using native mse as is videoplaypromise install using native vttcue mediacapabilities install mediacapabilities native mediacapabilities support found starting attach starting load of found variant with audio and video content so filtering out audio only content codecs avg bandwidth init completed initial stream setup after load calling switch bandwidth kbps switch switch switching to stream video switch stream audio already active calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch calling switch bandwidth kbps switch jumping forward seconds because of gap before start time of all seeked buffered seek presentationtime jumping forward seconds because of gap before start time of all seeked buffered seek presentationtime calling switch bandwidth kbps switch in the log above the playback started after minutes because of this i would say that this is also related to any suggestion on how to debug fix this would be much appreciated originally posted by puritanic in | 0 |
652,736 | 21,560,106,163 | IssuesEvent | 2022-05-01 03:03:55 | devs-immortal/Paradise-Lost | https://api.github.com/repos/devs-immortal/Paradise-Lost | closed | Portal not working!!! | Bug duplicate Priority More Info Required | ### What happened?
he animation goes along with the sound, but the portal doesn't work.
I tried to play without other mods, it did not help
(Maybe the portal does not work well with worlds that were created before installing).
https://www.youtube.com/watch?v=U3udsIqGWI4
### Mod Version
1.6.5 - 1.18
### Fabric API Version
last
### Relevant log output
_No response_
### Other mods
_No response_
### Additional Information
https://www.youtube.com/watch?v=U3udsIqGWI4 | 1.0 | Portal not working!!! - ### What happened?
he animation goes along with the sound, but the portal doesn't work.
I tried to play without other mods, it did not help
(Maybe the portal does not work well with worlds that were created before installing).
https://www.youtube.com/watch?v=U3udsIqGWI4
### Mod Version
1.6.5 - 1.18
### Fabric API Version
last
### Relevant log output
_No response_
### Other mods
_No response_
### Additional Information
https://www.youtube.com/watch?v=U3udsIqGWI4 | non_main | portal not working what happened he animation goes along with the sound but the portal doesn t work i tried to play without other mods it did not help maybe the portal does not work well with worlds that were created before installing mod version fabric api version last relevant log output no response other mods no response additional information | 0 |
737 | 4,330,673,379 | IssuesEvent | 2016-07-26 20:47:13 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | opened | Untappd: Update triggering? Example query not working | Maintainer Input Requested | One of the supplied example queries doesn't show the Untappd IA: https://duckduckgo.com/?q=bommen%20%26%20granaten%20barleywine&ia=food&drink
------
IA Page: http://duck.co/ia/view/untappd
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @urbanserj | True | Untappd: Update triggering? Example query not working - One of the supplied example queries doesn't show the Untappd IA: https://duckduckgo.com/?q=bommen%20%26%20granaten%20barleywine&ia=food&drink
------
IA Page: http://duck.co/ia/view/untappd
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @urbanserj | main | untappd update triggering example query not working one of the supplied example queries doesn t show the untappd ia ia page urbanserj | 1 |
2,592 | 8,817,635,485 | IssuesEvent | 2018-12-31 03:07:10 | mlabbe/nativefiledialog | https://api.github.com/repos/mlabbe/nativefiledialog | closed | ShellItem not released on error in Windows implementation | bug maintainer-confirmed windows | Hi,
The following `ShellItem` is not released if an error is returned while using it (the function jumps to the end):
https://github.com/mlabbe/nativefiledialog/blob/5cfe5002eb0fac1e49777a17dec70134147931e2/src/nfd_win.cpp#L403
Similarly, the file dialog itself is never released.
Compare it to [the official example](https://msdn.microsoft.com/en-us/library/windows/desktop/bb776913(v=vs.85).aspx#usage). | True | ShellItem not released on error in Windows implementation - Hi,
The following `ShellItem` is not released if an error is returned while using it (the function jumps to the end):
https://github.com/mlabbe/nativefiledialog/blob/5cfe5002eb0fac1e49777a17dec70134147931e2/src/nfd_win.cpp#L403
Similarly, the file dialog itself is never released.
Compare it to [the official example](https://msdn.microsoft.com/en-us/library/windows/desktop/bb776913(v=vs.85).aspx#usage). | main | shellitem not released on error in windows implementation hi the following shellitem is not released if an error is returned while using it the function jumps to the end similarly the file dialog itself is never released compare it to | 1 |
121,747 | 16,019,876,350 | IssuesEvent | 2021-04-20 21:08:24 | equinor/design-system | https://api.github.com/repos/equinor/design-system | opened | Create color scheme for oil and gas products | design feature request | Unwritten standards exist in Equinor for visualizing products.
Oil = green
gas = red
NGL = yellow
Condensate and/or water = blue
Some info about this [here from NPD:](https://statoilsrm.sharepoint.com/sites/Designsystem/_layouts/OneNote.aspx?id=%2Fsites%2FDesignsystem%2FShared%20Documents%2FGeneral%2FPower%20BI%2FExamples%2FPower%20BI%20examples&wd=target%28Color%20schemes.one%7CC461E3B1-30B9-4CD3-AE59-C6112EFAADA9%2FNPD%7CE603561B-EE51-49CE-BF9F-40F19141BBDF%2F%29
onenote:https://statoilsrm.sharepoint.com/sites/Designsystem/Shared%20Documents/General/Power%20BI/Examples/Power%20BI%20examples/Color%20schemes.one#NPD§ion-id={C461E3B1-30B9-4CD3-AE59-C6112EFAADA9}&page-id={E603561B-EE51-49CE-BF9F-40F19141BBDF}&end) | 1.0 | Create color scheme for oil and gas products - Unwritten standards exist in Equinor for visualizing products.
Oil = green
gas = red
NGL = yellow
Condensate and/or water = blue
Some info about this [here from NPD:](https://statoilsrm.sharepoint.com/sites/Designsystem/_layouts/OneNote.aspx?id=%2Fsites%2FDesignsystem%2FShared%20Documents%2FGeneral%2FPower%20BI%2FExamples%2FPower%20BI%20examples&wd=target%28Color%20schemes.one%7CC461E3B1-30B9-4CD3-AE59-C6112EFAADA9%2FNPD%7CE603561B-EE51-49CE-BF9F-40F19141BBDF%2F%29
onenote:https://statoilsrm.sharepoint.com/sites/Designsystem/Shared%20Documents/General/Power%20BI/Examples/Power%20BI%20examples/Color%20schemes.one#NPD§ion-id={C461E3B1-30B9-4CD3-AE59-C6112EFAADA9}&page-id={E603561B-EE51-49CE-BF9F-40F19141BBDF}&end) | non_main | create color scheme for oil and gas products unwritten standards exist in equinor for visualizing products oil green gas red ngl yellow condensate and or water blue some info about this onenote | 0 |
39,048 | 10,282,202,527 | IssuesEvent | 2019-08-26 10:27:34 | letscontrolit/ESPEasy | https://api.github.com/repos/letscontrolit/ESPEasy | closed | Puya image - where to find or what to use | Category: Build Type: Question | ESP-01S to be flashed with puya image. Read a lot, but what image to be used? In the releases there is no image name with "puya" in it?
https://github.com/letscontrolit/ESPEasy/releases
| 1.0 | Puya image - where to find or what to use - ESP-01S to be flashed with puya image. Read a lot, but what image to be used? In the releases there is no image name with "puya" in it?
https://github.com/letscontrolit/ESPEasy/releases
| non_main | puya image where to find or what to use esp to be flashed with puya image read a lot but what image to be used in the releases there is no image name with puya in it | 0 |
4,957 | 25,457,718,309 | IssuesEvent | 2022-11-24 15:31:22 | permon/permon | https://api.github.com/repos/permon/permon | opened | Review petscfix functions | maintainability | These functions should be contributed to PETSc or removed if already fixed | True | Review petscfix functions - These functions should be contributed to PETSc or removed if already fixed | main | review petscfix functions these functions should be contributed to petsc or removed if already fixed | 1 |
4,938 | 25,376,858,620 | IssuesEvent | 2022-11-21 14:43:15 | ClubInfoInsaT/application-amicale | https://api.github.com/repos/ClubInfoInsaT/application-amicale | closed | Mettre à jour le logo de proximo | enhancement maintainers | Le nouveau logo est sur le discord.
Se fait directement sur le site etud. | True | Mettre à jour le logo de proximo - Le nouveau logo est sur le discord.
Se fait directement sur le site etud. | main | mettre à jour le logo de proximo le nouveau logo est sur le discord se fait directement sur le site etud | 1 |
244,881 | 18,768,825,426 | IssuesEvent | 2021-11-06 12:48:43 | Team-Hydra-Discord/Feedback | https://api.github.com/repos/Team-Hydra-Discord/Feedback | closed | [Documentation] Create AppBot Docs | Documentation | ### Describe The Issue With This Content
Create AppBot docs.
### Where Does This Issue Reside?
```bash
Team Hydra Docs
```
### Expected Content
Full AppBot Docs
### Additional Context
_No response_ | 1.0 | [Documentation] Create AppBot Docs - ### Describe The Issue With This Content
Create AppBot docs.
### Where Does This Issue Reside?
```bash
Team Hydra Docs
```
### Expected Content
Full AppBot Docs
### Additional Context
_No response_ | non_main | create appbot docs describe the issue with this content create appbot docs where does this issue reside bash team hydra docs expected content full appbot docs additional context no response | 0 |
3,219 | 12,337,823,476 | IssuesEvent | 2020-05-14 15:32:13 | chaoss/website | https://api.github.com/repos/chaoss/website | closed | Request to get my article featured on CHAOSS Website Blog | Maintainer Task | Good day all,
I got accepted as an Outreachy intern for CHAOSS, we are expected to blog every 2 weeks. The articles are meant to be featured in the community website blog or any blog currently maintained by the community.
I wrote my first article on "How i got accepted into the Outreachy Internship". Please read here
https://docs.google.com/document/d/1vF69kwXMYtuzl9S0jXyOyS0AeSDga9SJet054ooIIWU/edit?usp=sharing
| True | Request to get my article featured on CHAOSS Website Blog - Good day all,
I got accepted as an Outreachy intern for CHAOSS, we are expected to blog every 2 weeks. The articles are meant to be featured in the community website blog or any blog currently maintained by the community.
I wrote my first article on "How i got accepted into the Outreachy Internship". Please read here
https://docs.google.com/document/d/1vF69kwXMYtuzl9S0jXyOyS0AeSDga9SJet054ooIIWU/edit?usp=sharing
| main | request to get my article featured on chaoss website blog good day all i got accepted as an outreachy intern for chaoss we are expected to blog every weeks the articles are meant to be featured in the community website blog or any blog currently maintained by the community i wrote my first article on how i got accepted into the outreachy internship please read here | 1 |
5,701 | 30,022,705,657 | IssuesEvent | 2023-06-27 01:50:40 | darekkay/todoist-export | https://api.github.com/repos/darekkay/todoist-export | closed | Improve state passing | Type: Maintainance | Form options (CSV/JSON) are currently passed through the OAuth `state` string parameter. This complicates passing state and is nonstandard. Instead, we should be storing this state locally with the `state` token as a key (and ideally validating the request):
> If state is used for carrying application state, and integrity of its contents is a concern, clients MUST protect state against tampering and swapping. This can be achieved by binding the contents of state to the browser session and/or signed/encrypted state values
See:
* https://tools.ietf.org/html/draft-bradley-oauth-jwt-encoded-state-09
* http://www.thread-safe.com/2014/05/the-correct-use-of-state-parameter-in.html
* https://stackoverflow.com/questions/52459476/does-oauth-state-mitigate-any-genuinely-dangerous-attacks
* https://pipedrive.readme.io/docs/marketplace-oauth-authorization-state-parameter | True | Improve state passing - Form options (CSV/JSON) are currently passed through the OAuth `state` string parameter. This complicates passing state and is nonstandard. Instead, we should be storing this state locally with the `state` token as a key (and ideally validating the request):
> If state is used for carrying application state, and integrity of its contents is a concern, clients MUST protect state against tampering and swapping. This can be achieved by binding the contents of state to the browser session and/or signed/encrypted state values
See:
* https://tools.ietf.org/html/draft-bradley-oauth-jwt-encoded-state-09
* http://www.thread-safe.com/2014/05/the-correct-use-of-state-parameter-in.html
* https://stackoverflow.com/questions/52459476/does-oauth-state-mitigate-any-genuinely-dangerous-attacks
* https://pipedrive.readme.io/docs/marketplace-oauth-authorization-state-parameter | main | improve state passing form options csv json are currently passed through the oauth state string parameter this complicates passing state and is nonstandard instead we should be storing this state locally with the state token as a key and ideally validating the request if state is used for carrying application state and integrity of its contents is a concern clients must protect state against tampering and swapping this can be achieved by binding the contents of state to the browser session and or signed encrypted state values see | 1 |
226,849 | 25,008,692,955 | IssuesEvent | 2022-11-03 13:49:11 | temporalio/sdk-go | https://api.github.com/repos/temporalio/sdk-go | closed | golang.org/x/tools-v0.1.10: 3 vulnerabilities (highest severity is: 6.1) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p></summary>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (golang.org/x/tools-v0.1.10 version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.1.10 | Direct | jQuery - 3.5.0 | ❌ |
| [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | golang.org/x/tools-v0.1.10 | Direct | jQuery - 3.0.0 | ❌ |
| [CVE-2012-6708](https://www.mend.io/vulnerability-database/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | golang.org/x/tools-v0.1.10 | Direct | jQuery - v1.9.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2015-9251</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - 3.0.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2012-6708</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
<p></p>
</details> | True | golang.org/x/tools-v0.1.10: 3 vulnerabilities (highest severity is: 6.1) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p></summary>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (golang.org/x/tools-v0.1.10 version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.1.10 | Direct | jQuery - 3.5.0 | ❌ |
| [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | golang.org/x/tools-v0.1.10 | Direct | jQuery - 3.0.0 | ❌ |
| [CVE-2012-6708](https://www.mend.io/vulnerability-database/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | golang.org/x/tools-v0.1.10 | Direct | jQuery - v1.9.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2015-9251</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - 3.0.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2012-6708</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.1.10</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.1.10.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **golang.org/x/tools-v0.1.10** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/b5942aefecb0379859bab42b44fa267ca8f3f8d8">b5942aefecb0379859bab42b44fa267ca8f3f8d8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
<p></p>
</details> | non_main | golang org x tools vulnerabilities highest severity is autoclosed vulnerable library golang org x tools library home page a href found in head commit a href vulnerabilities cve severity cvss dependency type fixed in golang org x tools version remediation available medium golang org x tools direct jquery low golang org x tools direct jquery low golang org x tools direct jquery details cve vulnerable library golang org x tools library home page a href dependency hierarchy x golang org x tools vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library golang org x tools library home page a href dependency hierarchy x golang org x tools vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library golang org x tools library home page a href dependency hierarchy x golang org x tools vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery | 0 |
234,916 | 18,022,591,209 | IssuesEvent | 2021-09-16 21:36:12 | AlmaLinux/build-system | https://api.github.com/repos/AlmaLinux/build-system | opened | Add a link to the Build System project board to README | documentation | We are using a GitHub's kanban board https://github.com/orgs/AlmaLinux/projects/1 for the Build System project. It would be nice to mention that fact in the README file so that people know where progress can be tracked. | 1.0 | Add a link to the Build System project board to README - We are using a GitHub's kanban board https://github.com/orgs/AlmaLinux/projects/1 for the Build System project. It would be nice to mention that fact in the README file so that people know where progress can be tracked. | non_main | add a link to the build system project board to readme we are using a github s kanban board for the build system project it would be nice to mention that fact in the readme file so that people know where progress can be tracked | 0 |
283,858 | 21,335,390,265 | IssuesEvent | 2022-04-18 14:01:08 | MetLife/TlsTestGate | https://api.github.com/repos/MetLife/TlsTestGate | closed | Wrong screenshot? | documentation | in https://marketplace.visualstudio.com/items?itemName=gattjoe.TlsTestGate i see:

in [manual](https://github.com/MetLife/TlsTestGate#yaml-sample) i see:
`steps:
- task: JoeGatt.TlsTestGate.custom-build-release-task.TlsTestGate@1
displayName: 'github.com SSL/TLS Test Gate'
inputs:
baseURL: github.com
port: 443
dnsserver: 8.8.8.8
**failBuild**: true` | 1.0 | Wrong screenshot? - in https://marketplace.visualstudio.com/items?itemName=gattjoe.TlsTestGate i see:

in [manual](https://github.com/MetLife/TlsTestGate#yaml-sample) i see:
`steps:
- task: JoeGatt.TlsTestGate.custom-build-release-task.TlsTestGate@1
displayName: 'github.com SSL/TLS Test Gate'
inputs:
baseURL: github.com
port: 443
dnsserver: 8.8.8.8
**failBuild**: true` | non_main | wrong screenshot in i see in i see steps task joegatt tlstestgate custom build release task tlstestgate displayname github com ssl tls test gate inputs baseurl github com port dnsserver failbuild true | 0 |
351,659 | 10,521,765,214 | IssuesEvent | 2019-09-30 07:04:52 | AY1920S1-CS2103T-T09-2/main | https://api.github.com/repos/AY1920S1-CS2103T-T09-2/main | opened | As a student who wants to get stronger i want to be advised on how much increment I should make for each exercise | priority.Low type.Story | so that i can train progressively | 1.0 | As a student who wants to get stronger i want to be advised on how much increment I should make for each exercise - so that i can train progressively | non_main | as a student who wants to get stronger i want to be advised on how much increment i should make for each exercise so that i can train progressively | 0 |
5,039 | 25,841,221,827 | IssuesEvent | 2022-12-13 00:37:36 | ElasticPerch/websocket | https://api.github.com/repos/ElasticPerch/websocket | opened | Add Upgrader.NegotiateSubprotocols | waiting on new maintainer feature request | From websocket created by [garyburd](https://github.com/garyburd): gorilla/websocket#480
See discussion [in this comment](https://github.com/gorilla/websocket/issues/399#issuecomment-413661967). | True | Add Upgrader.NegotiateSubprotocols - From websocket created by [garyburd](https://github.com/garyburd): gorilla/websocket#480
See discussion [in this comment](https://github.com/gorilla/websocket/issues/399#issuecomment-413661967). | main | add upgrader negotiatesubprotocols from websocket created by gorilla websocket see discussion | 1 |
300,993 | 22,708,219,232 | IssuesEvent | 2022-07-05 16:29:14 | Issafalcon/lsp-overloads.nvim | https://api.github.com/repos/Issafalcon/lsp-overloads.nvim | closed | Plugin documentation | documentation | Generate vim docs and README:
- Write README to include:
- Summary of the plugin
- Gif of signature helper
- Setup and configuration
- Acknowledgements
- Use github action to generate and commit Pandoc conversions of README to vim docs | 1.0 | Plugin documentation - Generate vim docs and README:
- Write README to include:
- Summary of the plugin
- Gif of signature helper
- Setup and configuration
- Acknowledgements
- Use github action to generate and commit Pandoc conversions of README to vim docs | non_main | plugin documentation generate vim docs and readme write readme to include summary of the plugin gif of signature helper setup and configuration acknowledgements use github action to generate and commit pandoc conversions of readme to vim docs | 0 |
54,458 | 6,388,330,830 | IssuesEvent | 2017-08-03 15:22:13 | WormBase/website | https://api.github.com/repos/WormBase/website | closed | Broken expression rest endpoint live/staging | Bug Report HelpDesk Under testing Webteam |
*Submitted from the feedback form on the WormBase website.*
<p>Server Error: 500 at <a href="http://www.wormbase.org/rest/widget/gene/WBGene00000412/expression">rest/widget/gene/WBGene00000412/expression</a></p><p>
</p><p>Error Message:<br> Caught exception in WormBase::Web::Controller::REST->widget "http://rest.wormbase.org/rest/widget/gene/WBGene00000412/expression failed with 500 at /var/lib/jenkins/jobs/staging_build/workspace/lib/WormBase/Web/Controller/REST.pm line 915." </p>
**Reported by:** abio*******************
**Submitted from:** <a target="_blank" href="http://www.wormbase.org//rest/widget/gene/WBGene00000412/expression">/rest/widget/gene/WBGene00000412/expression</a>
**Browser:** Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.115 Safari/537.36
| 1.0 | Broken expression rest endpoint live/staging -
*Submitted from the feedback form on the WormBase website.*
<p>Server Error: 500 at <a href="http://www.wormbase.org/rest/widget/gene/WBGene00000412/expression">rest/widget/gene/WBGene00000412/expression</a></p><p>
</p><p>Error Message:<br> Caught exception in WormBase::Web::Controller::REST->widget "http://rest.wormbase.org/rest/widget/gene/WBGene00000412/expression failed with 500 at /var/lib/jenkins/jobs/staging_build/workspace/lib/WormBase/Web/Controller/REST.pm line 915." </p>
**Reported by:** abio*******************
**Submitted from:** <a target="_blank" href="http://www.wormbase.org//rest/widget/gene/WBGene00000412/expression">/rest/widget/gene/WBGene00000412/expression</a>
**Browser:** Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.115 Safari/537.36
| non_main | broken expression rest endpoint live staging submitted from the feedback form on the wormbase website server error at a href error message caught exception in wormbase web controller rest gt widget failed with at var lib jenkins jobs staging build workspace lib wormbase web controller rest pm line reported by abio submitted from a target blank href browser mozilla windows nt applewebkit khtml like gecko chrome safari | 0 |
95,551 | 12,004,320,581 | IssuesEvent | 2020-04-09 11:18:55 | Western-Health-Covid19-Collaboration/wh_covid19_app | https://api.github.com/repos/Western-Health-Covid19-Collaboration/wh_covid19_app | closed | Change intubation checklist infographic | #1 Priority Design | The intubation checklist infographic (accessible through tertiary PDF button) requires updating to reflect changes in guidelines form 9th April.
Attached is new infographic.
@lukesleeman @Western-Health-Covid19-Collaboration/designers This might have to go through design first of all prior to devs?
[9thAprSAS COVID 19 Intubaiton Checklist WH version 1 (1).pdf](https://github.com/Western-Health-Covid19-Collaboration/wh_covid19_app/files/4454181/9thAprSAS.COVID.19.Intubaiton.Checklist.WH.version.1.1.pdf)
| 1.0 | Change intubation checklist infographic - The intubation checklist infographic (accessible through tertiary PDF button) requires updating to reflect changes in guidelines form 9th April.
Attached is new infographic.
@lukesleeman @Western-Health-Covid19-Collaboration/designers This might have to go through design first of all prior to devs?
[9thAprSAS COVID 19 Intubaiton Checklist WH version 1 (1).pdf](https://github.com/Western-Health-Covid19-Collaboration/wh_covid19_app/files/4454181/9thAprSAS.COVID.19.Intubaiton.Checklist.WH.version.1.1.pdf)
| non_main | change intubation checklist infographic the intubation checklist infographic accessible through tertiary pdf button requires updating to reflect changes in guidelines form april attached is new infographic lukesleeman western health collaboration designers this might have to go through design first of all prior to devs | 0 |
768,426 | 26,962,837,021 | IssuesEvent | 2023-02-08 19:37:51 | nanovms/nanos | https://api.github.com/repos/nanovms/nanos | closed | tracking tkt for elasticsearch - low priority | jvm low-priority | get a few variations of this
```
warning: overwriting existing file /etc/resolv.conf hostpath old: /home/eyberg/.ops/.staging/java_1.8.0_191/sysroot/etc/resolv.conf new: /home/eyberg/.ops/java_temp/resolv.conf
warning: overwriting existing file /proc/sys/kernel/hostname hostpath old: /home/eyberg/.ops/.staging/java_1.8.0_191/sysroot/proc/sys/kernel/hostname new: /home/eyberg/.ops/java_temp/hostname
[java -Xms1g -Xmx1g -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=75 -XX:+UseCMSInitiatingOccupancyOnly -Des.networkaddress.cache.ttl=60 -Des.networkaddress.cache.negative.ttl=10 -XX:+AlwaysPreTouch -Xss1m -Djava.awt.headless=true -Dfile.encoding=UTF-8 -Djna.nosys=true -XX:-OmitStackTraceInFastThrow -Dio.netty.noUnsafe=true -Dio.netty.noKeySetOptimization=true -Dio.netty.recycler.maxCapacityPerThread=0 -Dlog4j.shutdownHookEnabled=false -Dlog4j2.disable.jmx=true -Dos.name=Linux -Djava.io.tmpdir=/elasticsearch-9822391376708845111 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=data -XX:ErrorFile=logs/hs_err_pid%p.log -Djava.locale.providers=COMPAT -Dio.netty.allocator.type=unpooled -Des.path.home=/ -Des.path.conf=/home/eyberg/el/elasticsearch-7.0.0/config -Des.distribution.flavor=default -Des.distribution.type=tar -Des.bundled_jdk=true -cp /lib/* org.elasticsearch.bootstrap.Elasticsearch]
booting /home/eyberg/.ops/images/java.img ...
qemu-system-x86_64: warning: TCG doesn't support requested feature: CPUID.01H:ECX.vmx [bit 5]
assigned: 10.0.2.15
OpenJDK 64-Bit Server VM warning: Can't detect initial thread stack location - find_vma failed
unix_fault_page error: no vmap found for vaddr 0x1008b18
Unhandled: 000000000000000e
Page fault
interrupt: 000000000000000e
frame: 0000000100e01500
error code: 0000000000000000
address: 0000000001008b18
rax: 0000000000000028
rbx: 0000000001008b10
rcx: 0000000000000000
rdx: 0000000100207400
rsi: 00000070280b3d28
rdi: 00000070280b3d28
rbp: 00000070bf8ff9d0
rsp: 00000070bf8ff9b0
r8: 0000000000000000
r9: 0000000000000000
r10: 0000000000000000
r11: 0000000000000006
r12: 0000000100a3a9c0
r13: 00000070280b3d28
r14: 00000070280b3d28
r15: 0000000000000000
rip: 000000007f00d60a (table_find + 0000000000000074/000000000000009d)
flags: 0000000000000002
frame trace:
000000007f01f767 (soft_create_futex + 0000000000000028/0000000000000073)
000000007f01fe81 (futex + 0000000000000077/0000000000000566)
000000007f01b38f (syscall_debug + 000000000000028f/000000000000029c)
000000007f0002b0 (syscall_enter + 0000000000000062/000000000000006c)
00000070a98a8a55
00000070a98a9c78
00000070a9996228
00000070a9ab79de
00000070a9ab7dc7
00000070a98ea992
00000070938076db
stack trace:
0000000000000081
000000007f043608 (bootstrap_region + 0000000000000b28/0000000000200000)
00000070280b3d28
0000000100e01700
00000070bf8ffa00
000000007f01f767 (soft_create_futex + 0000000000000028/0000000000000073)
0000000000000081
00000070280b3d28
0000000000000000
0000000000000000
00000070bf8ffa90
000000007f01fe81 (futex + 0000000000000077/0000000000000566)
00000000000000ff
0000007039132169
0000000000000000
00000001fb01beec
0000000000000000
0000000100203400
ffffffffffffffda
0000000100203400
00000070280b3d28
1bd4100efb01beec
00000070bf8ffa90
0000000100203400
halt
exit status 255
```
```
eyberg@s1:~/el/elasticsearch-7.0.0$ cat config.json
{
"Args": ["-Xms1g", "-Xmx1g", "-XX:+UseConcMarkSweepGC", "-XX:CMSInitiatingOccupancyFraction=75",
"-XX:+UseCMSInitiatingOccupancyOnly", "-Des.networkaddress.cache.ttl=60", "-Des.networkaddress.cache.negative.ttl=10",
"-XX:+AlwaysPreTouch", "-Xss1m", "-Djava.awt.headless=true", "-Dfile.encoding=UTF-8", "-Djna.nosys=true",
"-XX:-OmitStackTraceInFastThrow", "-Dio.netty.noUnsafe=true", "-Dio.netty.noKeySetOptimization=true",
"-Dio.netty.recycler.maxCapacityPerThread=0", "-Dlog4j.shutdownHookEnabled=false", "-Dlog4j2.disable.jmx=true",
"-Dos.name=Linux",
"-Djava.io.tmpdir=/elasticsearch-9822391376708845111", "-XX:+HeapDumpOnOutOfMemoryError", "-XX:HeapDumpPath=data",
"-XX:ErrorFile=logs/hs_err_pid%p.log", "-Djava.locale.providers=COMPAT", "-Dio.netty.allocator.type=unpooled", "-Des.path.home=/",
"-Des.path.conf=/home/eyberg/el/elasticsearch-7.0.0/config", "-Des.distribution.flavor=default", "-Des.distribution.type=tar",
"-Des.bundled_jdk=true", "-cp", "/lib/*", "org.elasticsearch.bootstrap.Elasticsearch"],
"Dirs":["lib", "elasticsearch-9822391376708845111"],
"RunConfig": {
"Memory": "3G"
}
}
``` | 1.0 | tracking tkt for elasticsearch - low priority - get a few variations of this
```
warning: overwriting existing file /etc/resolv.conf hostpath old: /home/eyberg/.ops/.staging/java_1.8.0_191/sysroot/etc/resolv.conf new: /home/eyberg/.ops/java_temp/resolv.conf
warning: overwriting existing file /proc/sys/kernel/hostname hostpath old: /home/eyberg/.ops/.staging/java_1.8.0_191/sysroot/proc/sys/kernel/hostname new: /home/eyberg/.ops/java_temp/hostname
[java -Xms1g -Xmx1g -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=75 -XX:+UseCMSInitiatingOccupancyOnly -Des.networkaddress.cache.ttl=60 -Des.networkaddress.cache.negative.ttl=10 -XX:+AlwaysPreTouch -Xss1m -Djava.awt.headless=true -Dfile.encoding=UTF-8 -Djna.nosys=true -XX:-OmitStackTraceInFastThrow -Dio.netty.noUnsafe=true -Dio.netty.noKeySetOptimization=true -Dio.netty.recycler.maxCapacityPerThread=0 -Dlog4j.shutdownHookEnabled=false -Dlog4j2.disable.jmx=true -Dos.name=Linux -Djava.io.tmpdir=/elasticsearch-9822391376708845111 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=data -XX:ErrorFile=logs/hs_err_pid%p.log -Djava.locale.providers=COMPAT -Dio.netty.allocator.type=unpooled -Des.path.home=/ -Des.path.conf=/home/eyberg/el/elasticsearch-7.0.0/config -Des.distribution.flavor=default -Des.distribution.type=tar -Des.bundled_jdk=true -cp /lib/* org.elasticsearch.bootstrap.Elasticsearch]
booting /home/eyberg/.ops/images/java.img ...
qemu-system-x86_64: warning: TCG doesn't support requested feature: CPUID.01H:ECX.vmx [bit 5]
assigned: 10.0.2.15
OpenJDK 64-Bit Server VM warning: Can't detect initial thread stack location - find_vma failed
unix_fault_page error: no vmap found for vaddr 0x1008b18
Unhandled: 000000000000000e
Page fault
interrupt: 000000000000000e
frame: 0000000100e01500
error code: 0000000000000000
address: 0000000001008b18
rax: 0000000000000028
rbx: 0000000001008b10
rcx: 0000000000000000
rdx: 0000000100207400
rsi: 00000070280b3d28
rdi: 00000070280b3d28
rbp: 00000070bf8ff9d0
rsp: 00000070bf8ff9b0
r8: 0000000000000000
r9: 0000000000000000
r10: 0000000000000000
r11: 0000000000000006
r12: 0000000100a3a9c0
r13: 00000070280b3d28
r14: 00000070280b3d28
r15: 0000000000000000
rip: 000000007f00d60a (table_find + 0000000000000074/000000000000009d)
flags: 0000000000000002
frame trace:
000000007f01f767 (soft_create_futex + 0000000000000028/0000000000000073)
000000007f01fe81 (futex + 0000000000000077/0000000000000566)
000000007f01b38f (syscall_debug + 000000000000028f/000000000000029c)
000000007f0002b0 (syscall_enter + 0000000000000062/000000000000006c)
00000070a98a8a55
00000070a98a9c78
00000070a9996228
00000070a9ab79de
00000070a9ab7dc7
00000070a98ea992
00000070938076db
stack trace:
0000000000000081
000000007f043608 (bootstrap_region + 0000000000000b28/0000000000200000)
00000070280b3d28
0000000100e01700
00000070bf8ffa00
000000007f01f767 (soft_create_futex + 0000000000000028/0000000000000073)
0000000000000081
00000070280b3d28
0000000000000000
0000000000000000
00000070bf8ffa90
000000007f01fe81 (futex + 0000000000000077/0000000000000566)
00000000000000ff
0000007039132169
0000000000000000
00000001fb01beec
0000000000000000
0000000100203400
ffffffffffffffda
0000000100203400
00000070280b3d28
1bd4100efb01beec
00000070bf8ffa90
0000000100203400
halt
exit status 255
```
```
eyberg@s1:~/el/elasticsearch-7.0.0$ cat config.json
{
"Args": ["-Xms1g", "-Xmx1g", "-XX:+UseConcMarkSweepGC", "-XX:CMSInitiatingOccupancyFraction=75",
"-XX:+UseCMSInitiatingOccupancyOnly", "-Des.networkaddress.cache.ttl=60", "-Des.networkaddress.cache.negative.ttl=10",
"-XX:+AlwaysPreTouch", "-Xss1m", "-Djava.awt.headless=true", "-Dfile.encoding=UTF-8", "-Djna.nosys=true",
"-XX:-OmitStackTraceInFastThrow", "-Dio.netty.noUnsafe=true", "-Dio.netty.noKeySetOptimization=true",
"-Dio.netty.recycler.maxCapacityPerThread=0", "-Dlog4j.shutdownHookEnabled=false", "-Dlog4j2.disable.jmx=true",
"-Dos.name=Linux",
"-Djava.io.tmpdir=/elasticsearch-9822391376708845111", "-XX:+HeapDumpOnOutOfMemoryError", "-XX:HeapDumpPath=data",
"-XX:ErrorFile=logs/hs_err_pid%p.log", "-Djava.locale.providers=COMPAT", "-Dio.netty.allocator.type=unpooled", "-Des.path.home=/",
"-Des.path.conf=/home/eyberg/el/elasticsearch-7.0.0/config", "-Des.distribution.flavor=default", "-Des.distribution.type=tar",
"-Des.bundled_jdk=true", "-cp", "/lib/*", "org.elasticsearch.bootstrap.Elasticsearch"],
"Dirs":["lib", "elasticsearch-9822391376708845111"],
"RunConfig": {
"Memory": "3G"
}
}
``` | non_main | tracking tkt for elasticsearch low priority get a few variations of this warning overwriting existing file etc resolv conf hostpath old home eyberg ops staging java sysroot etc resolv conf new home eyberg ops java temp resolv conf warning overwriting existing file proc sys kernel hostname hostpath old home eyberg ops staging java sysroot proc sys kernel hostname new home eyberg ops java temp hostname booting home eyberg ops images java img qemu system warning tcg doesn t support requested feature cpuid ecx vmx assigned openjdk bit server vm warning can t detect initial thread stack location find vma failed unix fault page error no vmap found for vaddr unhandled page fault interrupt frame error code address rax rbx rcx rdx rsi rdi rbp rsp rip table find flags frame trace soft create futex futex syscall debug syscall enter stack trace bootstrap region soft create futex futex ffffffffffffffda halt exit status eyberg el elasticsearch cat config json args xx useconcmarksweepgc xx cmsinitiatingoccupancyfraction xx usecmsinitiatingoccupancyonly des networkaddress cache ttl des networkaddress cache negative ttl xx alwayspretouch djava awt headless true dfile encoding utf djna nosys true xx omitstacktraceinfastthrow dio netty nounsafe true dio netty nokeysetoptimization true dio netty recycler maxcapacityperthread shutdownhookenabled false disable jmx true dos name linux djava io tmpdir elasticsearch xx heapdumponoutofmemoryerror xx heapdumppath data xx errorfile logs hs err pid p log djava locale providers compat dio netty allocator type unpooled des path home des path conf home eyberg el elasticsearch config des distribution flavor default des distribution type tar des bundled jdk true cp lib org elasticsearch bootstrap elasticsearch dirs runconfig memory | 0 |
1,638 | 6,572,661,924 | IssuesEvent | 2017-09-11 04:11:13 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | firewalld: Allow Service definition and Source IP/Block specification | affects_2.2 feature_idea waiting_on_maintainer | ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
firewalld
##### ANSIBLE VERSION
```
ansible 2.2.0.0
```
##### OS / ENVIRONMENT
Redhat 7+
##### SUMMARY
The firewalld module should be improved to allow more basic functions. Right now, you can allow everything from a host, everything to a port, everything to a previously defined service, or a rich rule, which includes everything else. There should be a way to add a new service and then allow connections to that service from an IP or IP block. Match rules, tps rules, and the like, should still be within rich rules, but defining services in firewalld and slightly more granular control of source IPs should really be within the ability of this module.
| True | firewalld: Allow Service definition and Source IP/Block specification - ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
firewalld
##### ANSIBLE VERSION
```
ansible 2.2.0.0
```
##### OS / ENVIRONMENT
Redhat 7+
##### SUMMARY
The firewalld module should be improved to allow more basic functions. Right now, you can allow everything from a host, everything to a port, everything to a previously defined service, or a rich rule, which includes everything else. There should be a way to add a new service and then allow connections to that service from an IP or IP block. Match rules, tps rules, and the like, should still be within rich rules, but defining services in firewalld and slightly more granular control of source IPs should really be within the ability of this module.
| main | firewalld allow service definition and source ip block specification issue type feature idea component name firewalld ansible version ansible os environment redhat summary the firewalld module should be improved to allow more basic functions right now you can allow everything from a host everything to a port everything to a previously defined service or a rich rule which includes everything else there should be a way to add a new service and then allow connections to that service from an ip or ip block match rules tps rules and the like should still be within rich rules but defining services in firewalld and slightly more granular control of source ips should really be within the ability of this module | 1 |
12,601 | 3,631,332,747 | IssuesEvent | 2016-02-11 00:49:17 | systemd/systemd | https://api.github.com/repos/systemd/systemd | closed | Manpage for systemctl is incorrect, full service name isn't needed for enable/disable | documentation systemctl | The manpage (http://www.freedesktop.org/software/systemd/man/systemctl.html) says:
>For unit file commands, the specified NAME should be the full name of the unit file, or the absolute path to the unit file:
>
> systemctl enable foo.service
>or
> systemctl link /path/to/foo.service
but you can actually use systemctl enable foo without any problems.
| 1.0 | Manpage for systemctl is incorrect, full service name isn't needed for enable/disable - The manpage (http://www.freedesktop.org/software/systemd/man/systemctl.html) says:
>For unit file commands, the specified NAME should be the full name of the unit file, or the absolute path to the unit file:
>
> systemctl enable foo.service
>or
> systemctl link /path/to/foo.service
but you can actually use systemctl enable foo without any problems.
| non_main | manpage for systemctl is incorrect full service name isn t needed for enable disable the manpage says for unit file commands the specified name should be the full name of the unit file or the absolute path to the unit file systemctl enable foo service or systemctl link path to foo service but you can actually use systemctl enable foo without any problems | 0 |
1,889 | 6,577,532,919 | IssuesEvent | 2017-09-12 01:34:29 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | iam_policy module errors out when using policy_json field | affects_2.0 aws bug_report cloud waiting_on_maintainer | ##### Issue Type:
- Bug Report
##### Plugin Name:
iam_policy
##### Ansible Version:
2.0.1.0
##### Ansible Configuration:
n/a
##### Environment:
n/a
##### Summary:
iam_policy module errors out if you use the policy_json field. PR #2730 introduced this bug.
##### Steps To Reproduce:
Example:
```
- iam_policy:
iam_type: role
iam_name: my-role-name
policy_name: my-policy-name
policy_json: "{{ lookup( 'template', 'policy.json.j2', convert_data=False) }}"
state: present
```
##### Expected Results:
Expected to succeed.
##### Actual Results:
```
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/root/.ansible/tmp/ansible-tmp-1457652747.27-251976645440330/iam_policy", line 2528, in <module>
main()
File "/root/.ansible/tmp/ansible-tmp-1457652747.27-251976645440330/iam_policy", line 306, in main
if not isinstance(pdoc, basestring):
UnboundLocalError: local variable 'pdoc' referenced before assignment
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "iam_policy"}, "parsed": false}
```
| True | iam_policy module errors out when using policy_json field - ##### Issue Type:
- Bug Report
##### Plugin Name:
iam_policy
##### Ansible Version:
2.0.1.0
##### Ansible Configuration:
n/a
##### Environment:
n/a
##### Summary:
iam_policy module errors out if you use the policy_json field. PR #2730 introduced this bug.
##### Steps To Reproduce:
Example:
```
- iam_policy:
iam_type: role
iam_name: my-role-name
policy_name: my-policy-name
policy_json: "{{ lookup( 'template', 'policy.json.j2', convert_data=False) }}"
state: present
```
##### Expected Results:
Expected to succeed.
##### Actual Results:
```
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/root/.ansible/tmp/ansible-tmp-1457652747.27-251976645440330/iam_policy", line 2528, in <module>
main()
File "/root/.ansible/tmp/ansible-tmp-1457652747.27-251976645440330/iam_policy", line 306, in main
if not isinstance(pdoc, basestring):
UnboundLocalError: local variable 'pdoc' referenced before assignment
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "iam_policy"}, "parsed": false}
```
| main | iam policy module errors out when using policy json field issue type bug report plugin name iam policy ansible version ansible configuration n a environment n a summary iam policy module errors out if you use the policy json field pr introduced this bug steps to reproduce example iam policy iam type role iam name my role name policy name my policy name policy json lookup template policy json convert data false state present expected results expected to succeed actual results an exception occurred during task execution the full traceback is traceback most recent call last file root ansible tmp ansible tmp iam policy line in main file root ansible tmp ansible tmp iam policy line in main if not isinstance pdoc basestring unboundlocalerror local variable pdoc referenced before assignment fatal failed changed false failed true invocation module name iam policy parsed false | 1 |
1,422 | 3,738,859,406 | IssuesEvent | 2016-03-09 00:53:53 | Royedc4/caYca | https://api.github.com/repos/Royedc4/caYca | closed | Presentation 4 korea | requirement | Required by Alba Isabel. New specific presentation of caYca to Korea people. | 1.0 | Presentation 4 korea - Required by Alba Isabel. New specific presentation of caYca to Korea people. | non_main | presentation korea required by alba isabel new specific presentation of cayca to korea people | 0 |
21,915 | 11,424,548,328 | IssuesEvent | 2020-02-03 17:59:09 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Keras RNN training speed significantly slower with eager execution/control flow v2 | TF 2.0 comp:keras type:performance | **System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): yes
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A
- TensorFlow installed from (source or binary): source
- TensorFlow version (use command below): 2.0.0
- Python version: 3.6.8
- Bazel version (if compiling from source): N/A
- GCC/Compiler version (if compiling from source): N/A
- CUDA/cuDNN version: 10.0/7.6
- GPU model and memory: GTX 980 Ti
**Describe the current behavior**
Enabling eager execution or control flow v2 causes RNN training speed to decrease significantly.
**Describe the expected behavior**
Enabling eager mode or control flow v2 should not affect the training time (or improve it, ideally).
**Code to reproduce the issue**
``` python
import tensorflow as tf
import numpy as np
import timeit
use_eager = False
use_v2 = False
if not use_eager:
tf.compat.v1.disable_eager_execution()
if not use_v2:
tf.compat.v1.disable_control_flow_v2()
n_steps = 1000
n_input = 100
n_hidden = 1000
batch_size = 64
inputs = tf.keras.Input((n_steps, n_input))
outputs = tf.keras.layers.SimpleRNN(units=n_hidden, return_sequences=True)(inputs)
outputs = tf.keras.layers.Dense(units=n_input)(outputs)
model = tf.keras.Model(inputs=inputs, outputs=outputs)
model.compile(optimizer=tf.optimizers.SGD(0.1), loss="mse")
x = np.ones((batch_size, n_steps, n_input))
y = np.ones((batch_size, n_steps, n_input))
# warmup
model.fit(x, y, epochs=1)
start = timeit.default_timer()
model.fit(x, y, epochs=10)
print("Execution time:", timeit.default_timer() - start)
```
**Other info / logs**
On my machine the results look like:
- use_eager=False, use_v2=False: 5.90s
- use_eager=False, use_v2=True: 8.08s
- use_eager=True, use_v2=False: 9.81s
- use_eager=True, use_v2=True: 10.10s
So, overall a >60% increase in training time comparing no eager and no v2 to the current defaults.
| True | Keras RNN training speed significantly slower with eager execution/control flow v2 - **System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): yes
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A
- TensorFlow installed from (source or binary): source
- TensorFlow version (use command below): 2.0.0
- Python version: 3.6.8
- Bazel version (if compiling from source): N/A
- GCC/Compiler version (if compiling from source): N/A
- CUDA/cuDNN version: 10.0/7.6
- GPU model and memory: GTX 980 Ti
**Describe the current behavior**
Enabling eager execution or control flow v2 causes RNN training speed to decrease significantly.
**Describe the expected behavior**
Enabling eager mode or control flow v2 should not affect the training time (or improve it, ideally).
**Code to reproduce the issue**
``` python
import tensorflow as tf
import numpy as np
import timeit
use_eager = False
use_v2 = False
if not use_eager:
tf.compat.v1.disable_eager_execution()
if not use_v2:
tf.compat.v1.disable_control_flow_v2()
n_steps = 1000
n_input = 100
n_hidden = 1000
batch_size = 64
inputs = tf.keras.Input((n_steps, n_input))
outputs = tf.keras.layers.SimpleRNN(units=n_hidden, return_sequences=True)(inputs)
outputs = tf.keras.layers.Dense(units=n_input)(outputs)
model = tf.keras.Model(inputs=inputs, outputs=outputs)
model.compile(optimizer=tf.optimizers.SGD(0.1), loss="mse")
x = np.ones((batch_size, n_steps, n_input))
y = np.ones((batch_size, n_steps, n_input))
# warmup
model.fit(x, y, epochs=1)
start = timeit.default_timer()
model.fit(x, y, epochs=10)
print("Execution time:", timeit.default_timer() - start)
```
**Other info / logs**
On my machine the results look like:
- use_eager=False, use_v2=False: 5.90s
- use_eager=False, use_v2=True: 8.08s
- use_eager=True, use_v2=False: 9.81s
- use_eager=True, use_v2=True: 10.10s
So, overall a >60% increase in training time comparing no eager and no v2 to the current defaults.
| non_main | keras rnn training speed significantly slower with eager execution control flow system information have i written custom code as opposed to using a stock example script provided in tensorflow yes os platform and distribution e g linux ubuntu windows mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device n a tensorflow installed from source or binary source tensorflow version use command below python version bazel version if compiling from source n a gcc compiler version if compiling from source n a cuda cudnn version gpu model and memory gtx ti describe the current behavior enabling eager execution or control flow causes rnn training speed to decrease significantly describe the expected behavior enabling eager mode or control flow should not affect the training time or improve it ideally code to reproduce the issue python import tensorflow as tf import numpy as np import timeit use eager false use false if not use eager tf compat disable eager execution if not use tf compat disable control flow n steps n input n hidden batch size inputs tf keras input n steps n input outputs tf keras layers simplernn units n hidden return sequences true inputs outputs tf keras layers dense units n input outputs model tf keras model inputs inputs outputs outputs model compile optimizer tf optimizers sgd loss mse x np ones batch size n steps n input y np ones batch size n steps n input warmup model fit x y epochs start timeit default timer model fit x y epochs print execution time timeit default timer start other info logs on my machine the results look like use eager false use false use eager false use true use eager true use false use eager true use true so overall a increase in training time comparing no eager and no to the current defaults | 0 |
14,985 | 9,635,603,485 | IssuesEvent | 2019-05-16 01:58:25 | PuzzleServer/mainpuzzleserver | https://api.github.com/repos/PuzzleServer/mainpuzzleserver | closed | [P0] Add team blurbs | enhancement usability | Allow teams to fill in a blurb that will be shown during matchmaking to help them find players that are good fits for their teams | True | [P0] Add team blurbs - Allow teams to fill in a blurb that will be shown during matchmaking to help them find players that are good fits for their teams | non_main | add team blurbs allow teams to fill in a blurb that will be shown during matchmaking to help them find players that are good fits for their teams | 0 |
557,441 | 16,509,691,989 | IssuesEvent | 2021-05-26 01:22:28 | eclipse-ee4j/glassfish | https://api.github.com/repos/eclipse-ee4j/glassfish | closed | appclient -user xxx option is ignored unless -passwordfile is given | Component: security ERR: Assignee Priority: Major Stale Type: Bug | Steps to reproduce:
* appclient -name myname -client MyClient.jar
Expected result:
* Login dialog should default user name to "myname".
Actual result:
* Login dialog defaults user name to Windows Account.
#### Environment
GlassFish 3.1.1, Win 7 Pro SP 1 (64 Bit), JDK 1.7.0_21 | 1.0 | appclient -user xxx option is ignored unless -passwordfile is given - Steps to reproduce:
* appclient -name myname -client MyClient.jar
Expected result:
* Login dialog should default user name to "myname".
Actual result:
* Login dialog defaults user name to Windows Account.
#### Environment
GlassFish 3.1.1, Win 7 Pro SP 1 (64 Bit), JDK 1.7.0_21 | non_main | appclient user xxx option is ignored unless passwordfile is given steps to reproduce appclient name myname client myclient jar expected result login dialog should default user name to myname actual result login dialog defaults user name to windows account environment glassfish win pro sp bit jdk | 0 |
1,305 | 5,545,189,143 | IssuesEvent | 2017-03-22 20:58:11 | WhitestormJS/whitestorm.js | https://api.github.com/repos/WhitestormJS/whitestorm.js | opened | Spec/tests closer to src code | MAINTAINANCE | Ideally the tests should be closer to the src code.
Why?
Incentive to test the code. Not having to search for where existing specs are etc.
If not possible/difficult with our current setup, let's at least bring structure to the specs we have in the `test` folder.
Or discuss. what we have right now is far from ideal.
###### Version:
- [x] v2.x.x
- [ ] v1.x.x
###### Issue type:
- [ ] Bug
- [ ] Proposal/Enhancement
- [ ] Question
- [ ] Discussion
------
<details>
<summary> <b>Tested on: </b> </summary>
###### Desktop
- [ ] Chrome
- [ ] Chrome Canary
- [ ] Chrome dev-channel
- [ ] Firefox
- [ ] Opera
- [ ] Microsoft IE
- [ ] Microsoft Edge
###### Android
- [ ] Chrome
- [ ] Firefox
- [ ] Opera
###### IOS
- [ ] Chrome
- [ ] Firefox
- [ ] Opera
</details>
| True | Spec/tests closer to src code - Ideally the tests should be closer to the src code.
Why?
Incentive to test the code. Not having to search for where existing specs are etc.
If not possible/difficult with our current setup, let's at least bring structure to the specs we have in the `test` folder.
Or discuss. what we have right now is far from ideal.
###### Version:
- [x] v2.x.x
- [ ] v1.x.x
###### Issue type:
- [ ] Bug
- [ ] Proposal/Enhancement
- [ ] Question
- [ ] Discussion
------
<details>
<summary> <b>Tested on: </b> </summary>
###### Desktop
- [ ] Chrome
- [ ] Chrome Canary
- [ ] Chrome dev-channel
- [ ] Firefox
- [ ] Opera
- [ ] Microsoft IE
- [ ] Microsoft Edge
###### Android
- [ ] Chrome
- [ ] Firefox
- [ ] Opera
###### IOS
- [ ] Chrome
- [ ] Firefox
- [ ] Opera
</details>
| main | spec tests closer to src code ideally the tests should be closer to the src code why incentive to test the code not having to search for where existing specs are etc if not possible difficult with our current setup let s at least bring structure to the specs we have in the test folder or discuss what we have right now is far from ideal version x x x x issue type bug proposal enhancement question discussion tested on desktop chrome chrome canary chrome dev channel firefox opera microsoft ie microsoft edge android chrome firefox opera ios chrome firefox opera | 1 |
3,116 | 11,904,991,493 | IssuesEvent | 2020-03-30 17:47:30 | diofant/diofant | https://api.github.com/repos/diofant/diofant | opened | Port remaining polys/dense*.py methods to sparse polys | maintainability polys | Ex., implement PolyElement.decompose(), per [this TODO](https://github.com/diofant/diofant/blob/3d08f9ab8cd77359f97411382ad754b5dc09b96e/diofant/polys/rings.py#L2058-L2059). | True | Port remaining polys/dense*.py methods to sparse polys - Ex., implement PolyElement.decompose(), per [this TODO](https://github.com/diofant/diofant/blob/3d08f9ab8cd77359f97411382ad754b5dc09b96e/diofant/polys/rings.py#L2058-L2059). | main | port remaining polys dense py methods to sparse polys ex implement polyelement decompose per | 1 |
3,581 | 14,367,640,018 | IssuesEvent | 2020-12-01 07:03:55 | adda-team/adda | https://api.github.com/repos/adda-team/adda | closed | Upgrade to clFFT and clBLAS | OpenCL comp-Library enhancement maintainability pri-High | ```
AMD has open-sourced its projects clAmdFft and clAmdBlas some time ago. Now
they are called clFFT and clBLAS respectively. And these projects have new
binary releases:
https://github.com/clMathLibraries/clFFT/releases
https://github.com/clMathLibraries/clBLAS/releases
The changelog is probably not that impressive yet, but it make sense to upgrade
to it before the next release. Then the user will be able to benefit from
further development of clFFT and clBLAS simply by downloading new DLLs from
this projects.
The upgrade should be a matter of removing "Amd" from all names of headers,
libraries, and functions. But also wiki-s with installation instructions for
these packages should be updated.
```
Original issue reported on code.google.com by `yurkin` on 3 Aug 2014 at 4:49
- Blocking: #157
| True | Upgrade to clFFT and clBLAS - ```
AMD has open-sourced its projects clAmdFft and clAmdBlas some time ago. Now
they are called clFFT and clBLAS respectively. And these projects have new
binary releases:
https://github.com/clMathLibraries/clFFT/releases
https://github.com/clMathLibraries/clBLAS/releases
The changelog is probably not that impressive yet, but it make sense to upgrade
to it before the next release. Then the user will be able to benefit from
further development of clFFT and clBLAS simply by downloading new DLLs from
this projects.
The upgrade should be a matter of removing "Amd" from all names of headers,
libraries, and functions. But also wiki-s with installation instructions for
these packages should be updated.
```
Original issue reported on code.google.com by `yurkin` on 3 Aug 2014 at 4:49
- Blocking: #157
| main | upgrade to clfft and clblas amd has open sourced its projects clamdfft and clamdblas some time ago now they are called clfft and clblas respectively and these projects have new binary releases the changelog is probably not that impressive yet but it make sense to upgrade to it before the next release then the user will be able to benefit from further development of clfft and clblas simply by downloading new dlls from this projects the upgrade should be a matter of removing amd from all names of headers libraries and functions but also wiki s with installation instructions for these packages should be updated original issue reported on code google com by yurkin on aug at blocking | 1 |
60,289 | 8,414,787,861 | IssuesEvent | 2018-10-13 07:08:41 | deeplearning4j/deeplearning4j | https://api.github.com/repos/deeplearning4j/deeplearning4j | closed | ND4J: Add instructions for AVX artifacts | Documentation ND4J | The instructions to use AVX binaries on the following page isn't available in the new docs:
https://github.com/deeplearning4j/nd4j/blob/gh-pages/getstarted.md | 1.0 | ND4J: Add instructions for AVX artifacts - The instructions to use AVX binaries on the following page isn't available in the new docs:
https://github.com/deeplearning4j/nd4j/blob/gh-pages/getstarted.md | non_main | add instructions for avx artifacts the instructions to use avx binaries on the following page isn t available in the new docs | 0 |
104,009 | 11,387,474,641 | IssuesEvent | 2020-01-29 15:05:06 | vtex-apps/io-documentation | https://api.github.com/repos/vtex-apps/io-documentation | closed | vtex-apps/helloworld has no documentation yet | no-documentation | [vtex-apps/helloworld](https://github.com/vtex-apps/helloworld) hasn't created any README file yet or is not using Docs Builder | 1.0 | vtex-apps/helloworld has no documentation yet - [vtex-apps/helloworld](https://github.com/vtex-apps/helloworld) hasn't created any README file yet or is not using Docs Builder | non_main | vtex apps helloworld has no documentation yet hasn t created any readme file yet or is not using docs builder | 0 |
815,533 | 30,560,275,691 | IssuesEvent | 2023-07-20 14:13:09 | intersystems-community/vscode-objectscript | https://api.github.com/repos/intersystems-community/vscode-objectscript | closed | revisit support export/import xml | enhancement upstream priority/important-soon | Hi All,
with the deprecation of InterSystems Studio, can we look into getting export/import of xml packaged source code implemented?
As a developer i often have the issue that i need to transport code from e.g. a non CCR controlled scratch environment to the BASE environment which then is under CCR source control.
This usually is done via ISC Studio export/import functionality.
ISC Studio import functionality also triggers the question on source controlled environment to add all imported code to source control (here CCR).
Using UDL as the export format can be prohibitive if the code consists of multiple classes/routines in multiple packages. (e.g. InterSystems Integration projects)
Also the current export functionality is rather clunky in large namespaces (i.e. thousands of classes) and also i found a bit hit and miss in regards to it working (got it working 2 out of 10 tries, usually not working when i try to present the functionality to others) | 1.0 | revisit support export/import xml - Hi All,
with the deprecation of InterSystems Studio, can we look into getting export/import of xml packaged source code implemented?
As a developer i often have the issue that i need to transport code from e.g. a non CCR controlled scratch environment to the BASE environment which then is under CCR source control.
This usually is done via ISC Studio export/import functionality.
ISC Studio import functionality also triggers the question on source controlled environment to add all imported code to source control (here CCR).
Using UDL as the export format can be prohibitive if the code consists of multiple classes/routines in multiple packages. (e.g. InterSystems Integration projects)
Also the current export functionality is rather clunky in large namespaces (i.e. thousands of classes) and also i found a bit hit and miss in regards to it working (got it working 2 out of 10 tries, usually not working when i try to present the functionality to others) | non_main | revisit support export import xml hi all with the deprecation of intersystems studio can we look into getting export import of xml packaged source code implemented as a developer i often have the issue that i need to transport code from e g a non ccr controlled scratch environment to the base environment which then is under ccr source control this usually is done via isc studio export import functionality isc studio import functionality also triggers the question on source controlled environment to add all imported code to source control here ccr using udl as the export format can be prohibitive if the code consists of multiple classes routines in multiple packages e g intersystems integration projects also the current export functionality is rather clunky in large namespaces i e thousands of classes and also i found a bit hit and miss in regards to it working got it working out of tries usually not working when i try to present the functionality to others | 0 |
1,262 | 5,353,855,504 | IssuesEvent | 2017-02-20 07:54:14 | espeak-ng/espeak-ng | https://api.github.com/repos/espeak-ng/espeak-ng | closed | Merge the android branch into master. | maintainability portability resolved/fixed | Now that espeak-ng has diverged from espeak, it makes sense to have the android branch merged into the main development line. This will make it easier to maintain the Android support in the future and keep it up-to-date.
- [x] Merge the android code into the master branch.
- [x] Fix building the JNI and libespeak-ng code for Android.
- [x] Incorporate the Android README text into the README file.
- [x] Add ucd-tools as a git subtree.
- [x] Use ucd-tools to create a compatibility shim for the wide-character APIs. | True | Merge the android branch into master. - Now that espeak-ng has diverged from espeak, it makes sense to have the android branch merged into the main development line. This will make it easier to maintain the Android support in the future and keep it up-to-date.
- [x] Merge the android code into the master branch.
- [x] Fix building the JNI and libespeak-ng code for Android.
- [x] Incorporate the Android README text into the README file.
- [x] Add ucd-tools as a git subtree.
- [x] Use ucd-tools to create a compatibility shim for the wide-character APIs. | main | merge the android branch into master now that espeak ng has diverged from espeak it makes sense to have the android branch merged into the main development line this will make it easier to maintain the android support in the future and keep it up to date merge the android code into the master branch fix building the jni and libespeak ng code for android incorporate the android readme text into the readme file add ucd tools as a git subtree use ucd tools to create a compatibility shim for the wide character apis | 1 |
7,797 | 8,056,695,841 | IssuesEvent | 2018-08-02 13:29:07 | spring-io/initializr | https://api.github.com/repos/spring-io/initializr | closed | Remove JTA entries | type: service-upgrade | We currently have three entries for JTA and the traffic on them is quite low. While supporting JTA is important for Spring Boot, I am not entirely sure we should promote its usage for new projects.
In our effort to focus our proposition, I am wondering if we shouldn't remove them. | 1.0 | Remove JTA entries - We currently have three entries for JTA and the traffic on them is quite low. While supporting JTA is important for Spring Boot, I am not entirely sure we should promote its usage for new projects.
In our effort to focus our proposition, I am wondering if we shouldn't remove them. | non_main | remove jta entries we currently have three entries for jta and the traffic on them is quite low while supporting jta is important for spring boot i am not entirely sure we should promote its usage for new projects in our effort to focus our proposition i am wondering if we shouldn t remove them | 0 |
975 | 4,726,024,203 | IssuesEvent | 2016-10-18 08:54:17 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | win_owner fails with recurse option specified | affects_2.1 bug_report waiting_on_maintainer windows | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
win_owner
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
```
##### CONFIGURATION
<!---
-->
N/A
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Windows
##### SUMMARY
<!--- Explain the problem briefly -->
Ansible converts yes/true/no/false to booleans before passing them to win_owner but win_owner's ``recurse`` parameter seems is expecting yes/no string instead of a boolean.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- win_owner:
path: "{{ app_dir }}"
user: user@domain
recurse: yes
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Set ownership of specified directory recursively without errors.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
fatal: [hostname]: FAILED! => {"changed": false, "failed": true, "msg": "Argument recurse needs to be one of no,yes but was True."}
```
| True | win_owner fails with recurse option specified - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
win_owner
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
```
##### CONFIGURATION
<!---
-->
N/A
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Windows
##### SUMMARY
<!--- Explain the problem briefly -->
Ansible converts yes/true/no/false to booleans before passing them to win_owner but win_owner's ``recurse`` parameter seems is expecting yes/no string instead of a boolean.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- win_owner:
path: "{{ app_dir }}"
user: user@domain
recurse: yes
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Set ownership of specified directory recursively without errors.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
fatal: [hostname]: FAILED! => {"changed": false, "failed": true, "msg": "Argument recurse needs to be one of no,yes but was True."}
```
| main | win owner fails with recurse option specified issue type bug report component name win owner ansible version ansible configuration n a os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific windows summary ansible converts yes true no false to booleans before passing them to win owner but win owner s recurse parameter seems is expecting yes no string instead of a boolean steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used win owner path app dir user user domain recurse yes expected results set ownership of specified directory recursively without errors actual results fatal failed changed false failed true msg argument recurse needs to be one of no yes but was true | 1 |
515 | 3,883,049,194 | IssuesEvent | 2016-04-13 12:31:37 | duckduckgo/zeroclickinfo-goodies | https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies | closed | Quack | Improvement Maintainer Input Requested | The "Need a Slack Invitation?" message needs updating after the Slackbot is deployed.
------
IA Page: http://duck.co/ia/view/quackhack_cheat_sheet
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @zekiel | True | Quack - The "Need a Slack Invitation?" message needs updating after the Slackbot is deployed.
------
IA Page: http://duck.co/ia/view/quackhack_cheat_sheet
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @zekiel | main | quack the need a slack invitation message needs updating after the slackbot is deployed ia page zekiel | 1 |
2,170 | 7,603,104,890 | IssuesEvent | 2018-04-29 10:50:00 | lansuite/lansuite | https://api.github.com/repos/lansuite/lansuite | opened | Scheduling a release / TODOs for the next release | pending-maintainer-response question | ## Context
The last release of LanSuite was v4.2 from Mar 15, 2015.
Since then we did a couple of changes here and there.
What changes do we want/need to do before we ship a new release?
- [ ] Add a `CHANGELOG.md` with the listed changes
- [ ] Fix #72
- [ ] Fix #9
- [ ] Fix #89
- [ ] Fix #145
- [ ] Fix #185
- [ ] Fix #222
What else comes into your mind @M4LuZ ?
What about #183 ?
I know that we talked about to test the new code at a party before we release it.
The thing is: Not so many people run this code at a party on master version (maybe you and me, but that's it).
And if we wait and keep changing, we pile up changes and changes and many people are "afraid" to upgrade.
So the coin has two sides.
What do you think? | True | Scheduling a release / TODOs for the next release - ## Context
The last release of LanSuite was v4.2 from Mar 15, 2015.
Since then we did a couple of changes here and there.
What changes do we want/need to do before we ship a new release?
- [ ] Add a `CHANGELOG.md` with the listed changes
- [ ] Fix #72
- [ ] Fix #9
- [ ] Fix #89
- [ ] Fix #145
- [ ] Fix #185
- [ ] Fix #222
What else comes into your mind @M4LuZ ?
What about #183 ?
I know that we talked about to test the new code at a party before we release it.
The thing is: Not so many people run this code at a party on master version (maybe you and me, but that's it).
And if we wait and keep changing, we pile up changes and changes and many people are "afraid" to upgrade.
So the coin has two sides.
What do you think? | main | scheduling a release todos for the next release context the last release of lansuite was from mar since then we did a couple of changes here and there what changes do we want need to do before we ship a new release add a changelog md with the listed changes fix fix fix fix fix fix what else comes into your mind what about i know that we talked about to test the new code at a party before we release it the thing is not so many people run this code at a party on master version maybe you and me but that s it and if we wait and keep changing we pile up changes and changes and many people are afraid to upgrade so the coin has two sides what do you think | 1 |
2,714 | 9,544,095,952 | IssuesEvent | 2019-05-01 13:06:38 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | closed | EventArgs should not implement interfaces | Area: analyzer Area: maintainability backlog feature | `EventArgs` are by themselves contracts. Therefore, they should not implement any additional interface. | True | EventArgs should not implement interfaces - `EventArgs` are by themselves contracts. Therefore, they should not implement any additional interface. | main | eventargs should not implement interfaces eventargs are by themselves contracts therefore they should not implement any additional interface | 1 |
4,438 | 23,057,650,776 | IssuesEvent | 2022-07-25 06:56:32 | samkatakouzinosracing/nspace.hu-soundshift | https://api.github.com/repos/samkatakouzinosracing/nspace.hu-soundshift | closed | Edit Global Mazda MX-5 Cup shift point to 7200 rpm | vehicle maintain | It's set at 6400 rpm now, which is hit almost immediately after changing up a gear. I have set it at 7200 rpm now, which is also the exact moment the red light comes up in game and gives you just enough time to change gears before hitting the limiter. In the MX-5, afaik, it's important to stay in a gear as long as possible to get the best acceleration.
So, can you please update it to 7200 rpm? | True | Edit Global Mazda MX-5 Cup shift point to 7200 rpm - It's set at 6400 rpm now, which is hit almost immediately after changing up a gear. I have set it at 7200 rpm now, which is also the exact moment the red light comes up in game and gives you just enough time to change gears before hitting the limiter. In the MX-5, afaik, it's important to stay in a gear as long as possible to get the best acceleration.
So, can you please update it to 7200 rpm? | main | edit global mazda mx cup shift point to rpm it s set at rpm now which is hit almost immediately after changing up a gear i have set it at rpm now which is also the exact moment the red light comes up in game and gives you just enough time to change gears before hitting the limiter in the mx afaik it s important to stay in a gear as long as possible to get the best acceleration so can you please update it to rpm | 1 |
35,237 | 16,992,924,901 | IssuesEvent | 2021-07-01 00:02:53 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | kvserver: limit MVCC scan in Raft Entries retrieval | A-kv-replication C-bug C-performance N-followup O-postmortem T-kv | In a recent customer escalation, we saw a newly elected Raft leader spend so long to send 100 log entries to catch up a follower that it missed its heartbeat. The log entries were found to be about 132 kb each, and we limit these messages to be a maximum of 32 kb (but at least 1 entry), so responses were being sent with a single entry each. It seemed to be taking 60 ms to retrieve a single entry, totalling 6s for 100 log entries. The log was also huge, at ~900,000 entries, with the current commit index at 1500.
Digging into the Raft code, we find that `raft.maybeSendAppend()` calls int `raftLog.entries()` to fetch entries to send to the follower. Notice how it limits this solely by `maxsize` (32 kb), otherwise using the entire Raft log (`lastIndex()+1`) as an upper bound:
https://github.com/cockroachdb/vendored/blob/310ea3cd09b55055b290f0a4cf9991a91725797e/go.etcd.io/etcd/raft/v3/log.go#L287-L292
This in turn gets passed into `Storage.Entries()`:
https://github.com/cockroachdb/vendored/blob/310ea3cd09b55055b290f0a4cf9991a91725797e/go.etcd.io/etcd/raft/v3/log.go#L350
Our implementation of `Entries()`, however, does not pass the max size bound to the MVCC iterator, only the key bounds, instead relying on the passed scan function to error when the size limit is exceeded:
https://github.com/cockroachdb/cockroach/blob/6b9168de33628257f3f1fdb17af1f8205ee21d32/pkg/kv/kvserver/replica_raftstorage.go#L236-L243
The problem is that `MVCCIterate` will actually fetch 1000 entries before calling scanFunc:
https://github.com/cockroachdb/cockroach/blob/8f5231d3443f372c88632c8b77c7d954890cc334/pkg/storage/mvcc.go#L2662-L2666
https://github.com/cockroachdb/cockroach/blob/8f5231d3443f372c88632c8b77c7d954890cc334/pkg/storage/mvcc.go#L2679-L2686
Since we're using the entire Raft log as an upper bound, not passing in a `MVCCScanOptions.MaxSize` for the iterator, and only doing size limiting in the passed `scanFunc`, this means that every time we fetch a single log entry we're actually fetching and decoding 1000 entries, each 132 kb large.
The simplest fix here is probably to just pass in the max size from Raft in `MVCCScanOptions`.
/cc @cockroachdb/kv | True | kvserver: limit MVCC scan in Raft Entries retrieval - In a recent customer escalation, we saw a newly elected Raft leader spend so long to send 100 log entries to catch up a follower that it missed its heartbeat. The log entries were found to be about 132 kb each, and we limit these messages to be a maximum of 32 kb (but at least 1 entry), so responses were being sent with a single entry each. It seemed to be taking 60 ms to retrieve a single entry, totalling 6s for 100 log entries. The log was also huge, at ~900,000 entries, with the current commit index at 1500.
Digging into the Raft code, we find that `raft.maybeSendAppend()` calls int `raftLog.entries()` to fetch entries to send to the follower. Notice how it limits this solely by `maxsize` (32 kb), otherwise using the entire Raft log (`lastIndex()+1`) as an upper bound:
https://github.com/cockroachdb/vendored/blob/310ea3cd09b55055b290f0a4cf9991a91725797e/go.etcd.io/etcd/raft/v3/log.go#L287-L292
This in turn gets passed into `Storage.Entries()`:
https://github.com/cockroachdb/vendored/blob/310ea3cd09b55055b290f0a4cf9991a91725797e/go.etcd.io/etcd/raft/v3/log.go#L350
Our implementation of `Entries()`, however, does not pass the max size bound to the MVCC iterator, only the key bounds, instead relying on the passed scan function to error when the size limit is exceeded:
https://github.com/cockroachdb/cockroach/blob/6b9168de33628257f3f1fdb17af1f8205ee21d32/pkg/kv/kvserver/replica_raftstorage.go#L236-L243
The problem is that `MVCCIterate` will actually fetch 1000 entries before calling scanFunc:
https://github.com/cockroachdb/cockroach/blob/8f5231d3443f372c88632c8b77c7d954890cc334/pkg/storage/mvcc.go#L2662-L2666
https://github.com/cockroachdb/cockroach/blob/8f5231d3443f372c88632c8b77c7d954890cc334/pkg/storage/mvcc.go#L2679-L2686
Since we're using the entire Raft log as an upper bound, not passing in a `MVCCScanOptions.MaxSize` for the iterator, and only doing size limiting in the passed `scanFunc`, this means that every time we fetch a single log entry we're actually fetching and decoding 1000 entries, each 132 kb large.
The simplest fix here is probably to just pass in the max size from Raft in `MVCCScanOptions`.
/cc @cockroachdb/kv | non_main | kvserver limit mvcc scan in raft entries retrieval in a recent customer escalation we saw a newly elected raft leader spend so long to send log entries to catch up a follower that it missed its heartbeat the log entries were found to be about kb each and we limit these messages to be a maximum of kb but at least entry so responses were being sent with a single entry each it seemed to be taking ms to retrieve a single entry totalling for log entries the log was also huge at entries with the current commit index at digging into the raft code we find that raft maybesendappend calls int raftlog entries to fetch entries to send to the follower notice how it limits this solely by maxsize kb otherwise using the entire raft log lastindex as an upper bound this in turn gets passed into storage entries our implementation of entries however does not pass the max size bound to the mvcc iterator only the key bounds instead relying on the passed scan function to error when the size limit is exceeded the problem is that mvcciterate will actually fetch entries before calling scanfunc since we re using the entire raft log as an upper bound not passing in a mvccscanoptions maxsize for the iterator and only doing size limiting in the passed scanfunc this means that every time we fetch a single log entry we re actually fetching and decoding entries each kb large the simplest fix here is probably to just pass in the max size from raft in mvccscanoptions cc cockroachdb kv | 0 |
1,338 | 5,721,482,647 | IssuesEvent | 2017-04-20 06:46:16 | tomchentw/react-google-maps | https://api.github.com/repos/tomchentw/react-google-maps | closed | triggerEvent('resize') does...what? | CALL_FOR_MAINTAINERS | http://react-google-maps.tomchentw.com/#/?_k=vkf06f
If you load this in Chrome nightly and open the console, you can see the logs dump on the window resize event, but the map is not being updated to reflect the changing screen size.
The example code just logs the component instance on resize. But how does one actually trigger the resize _on the map_, with access only to the _component_? Looking at the source code, it looks like `triggerEvent` just spreads the args, but my understanding of the gMaps API is that I need to run something like `map.setCenter` or `setBounds` or something on the event?
What's the way to do this on the given implementation? I wouldn't ask, but it just doesn't seem to be working in the demo example.
Thanks!
| True | triggerEvent('resize') does...what? - http://react-google-maps.tomchentw.com/#/?_k=vkf06f
If you load this in Chrome nightly and open the console, you can see the logs dump on the window resize event, but the map is not being updated to reflect the changing screen size.
The example code just logs the component instance on resize. But how does one actually trigger the resize _on the map_, with access only to the _component_? Looking at the source code, it looks like `triggerEvent` just spreads the args, but my understanding of the gMaps API is that I need to run something like `map.setCenter` or `setBounds` or something on the event?
What's the way to do this on the given implementation? I wouldn't ask, but it just doesn't seem to be working in the demo example.
Thanks!
| main | triggerevent resize does what if you load this in chrome nightly and open the console you can see the logs dump on the window resize event but the map is not being updated to reflect the changing screen size the example code just logs the component instance on resize but how does one actually trigger the resize on the map with access only to the component looking at the source code it looks like triggerevent just spreads the args but my understanding of the gmaps api is that i need to run something like map setcenter or setbounds or something on the event what s the way to do this on the given implementation i wouldn t ask but it just doesn t seem to be working in the demo example thanks | 1 |
1,030 | 4,827,515,754 | IssuesEvent | 2016-11-07 13:52:44 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | temporary AWS Access Keys results in auth failures | affects_1.9 aws bug_report cloud waiting_on_maintainer | #### Issue Type:
Bug Report
#### Component Name:
ec2 module
#### Ansible Version:
ansible 1.9.4
#### Ansible Configuration:
none
#### Environment:
Mac OSX 10.11 / Not applicable
#### Summary:
Using temporary AWS Access Keys results in auth failures
#### Steps to reproduce:
generate temporary access keys (eg. via STS or SAML provider)
attempt to create ec2 resource
#### Expected results:
ec2 resource is created
#### Actual results:
AWS AuthFailure exception
I originally lodged this against ansible/ansible as https://github.com/ansible/ansible/issues/12959 but I think maybe this should be resolved in the core modules code.
There's a comment in the code that suggests, that perhaps the modules just need to be modified to use `connect_to_aws()`
```
def get_ec2_creds(module):
''' for compatibility mode with old modules that don't/can't yet
use ec2_connect method '''
```
```
failed: [localhost] => {"failed": true, "parsed": false}
Traceback (most recent call last):
File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 3070, in <module>
main()
File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 1249, in main
(instance_dict_array, new_instance_ids, changed) = create_instances(module, ec2, vpc)
File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 792, in create_instances
vpc_id = vpc.get_all_subnets(subnet_ids=[vpc_subnet_id])[0].vpc_id
File "/Library/Python/2.7/site-packages/boto/vpc/__init__.py", line 1153, in get_all_subnets
return self.get_list('DescribeSubnets', params, [('item', Subnet)])
File "/Library/Python/2.7/site-packages/boto/connection.py", line 1186, in get_list
raise self.ResponseError(response.status, response.reason, body)
boto.exception.EC2ResponseError: EC2ResponseError: 401 Unauthorized
<?xml version="1.0" encoding="UTF-8"?>
<Response><Errors><Error><Code>AuthFailure</Code><Message>AWS was not able to validate the provided access credentials</Message></Error></Errors><RequestID>bcce0f14-b8d4-46e0-a582-17993365b18b</RequestID></Response>
from my investigation the issue appears in the module_utils/ec2.py get_ec2_creds( ) which returns
ec2_url, boto_params['aws_access_key_id'], boto_params['aws_secret_access_key'], region
```
since the aws_access_key_id in this example will only work with a security_token, the method is effectively broken.
I think the function should at least warn when it detects a security token or an access key that starts with ASIA instead of AKIA.
| True | temporary AWS Access Keys results in auth failures - #### Issue Type:
Bug Report
#### Component Name:
ec2 module
#### Ansible Version:
ansible 1.9.4
#### Ansible Configuration:
none
#### Environment:
Mac OSX 10.11 / Not applicable
#### Summary:
Using temporary AWS Access Keys results in auth failures
#### Steps to reproduce:
generate temporary access keys (eg. via STS or SAML provider)
attempt to create ec2 resource
#### Expected results:
ec2 resource is created
#### Actual results:
AWS AuthFailure exception
I originally lodged this against ansible/ansible as https://github.com/ansible/ansible/issues/12959 but I think maybe this should be resolved in the core modules code.
There's a comment in the code that suggests, that perhaps the modules just need to be modified to use `connect_to_aws()`
```
def get_ec2_creds(module):
''' for compatibility mode with old modules that don't/can't yet
use ec2_connect method '''
```
```
failed: [localhost] => {"failed": true, "parsed": false}
Traceback (most recent call last):
File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 3070, in <module>
main()
File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 1249, in main
(instance_dict_array, new_instance_ids, changed) = create_instances(module, ec2, vpc)
File "/Users/secole/.ansible/tmp/ansible-tmp-1446095011.34-234704611276689/ec2", line 792, in create_instances
vpc_id = vpc.get_all_subnets(subnet_ids=[vpc_subnet_id])[0].vpc_id
File "/Library/Python/2.7/site-packages/boto/vpc/__init__.py", line 1153, in get_all_subnets
return self.get_list('DescribeSubnets', params, [('item', Subnet)])
File "/Library/Python/2.7/site-packages/boto/connection.py", line 1186, in get_list
raise self.ResponseError(response.status, response.reason, body)
boto.exception.EC2ResponseError: EC2ResponseError: 401 Unauthorized
<?xml version="1.0" encoding="UTF-8"?>
<Response><Errors><Error><Code>AuthFailure</Code><Message>AWS was not able to validate the provided access credentials</Message></Error></Errors><RequestID>bcce0f14-b8d4-46e0-a582-17993365b18b</RequestID></Response>
from my investigation the issue appears in the module_utils/ec2.py get_ec2_creds( ) which returns
ec2_url, boto_params['aws_access_key_id'], boto_params['aws_secret_access_key'], region
```
since the aws_access_key_id in this example will only work with a security_token, the method is effectively broken.
I think the function should at least warn when it detects a security token or an access key that starts with ASIA instead of AKIA.
| main | temporary aws access keys results in auth failures issue type bug report component name module ansible version ansible ansible configuration none environment mac osx not applicable summary using temporary aws access keys results in auth failures steps to reproduce generate temporary access keys eg via sts or saml provider attempt to create resource expected results resource is created actual results aws authfailure exception i originally lodged this against ansible ansible as but i think maybe this should be resolved in the core modules code there s a comment in the code that suggests that perhaps the modules just need to be modified to use connect to aws def get creds module for compatibility mode with old modules that don t can t yet use connect method failed failed true parsed false traceback most recent call last file users secole ansible tmp ansible tmp line in main file users secole ansible tmp ansible tmp line in main instance dict array new instance ids changed create instances module vpc file users secole ansible tmp ansible tmp line in create instances vpc id vpc get all subnets subnet ids vpc id file library python site packages boto vpc init py line in get all subnets return self get list describesubnets params file library python site packages boto connection py line in get list raise self responseerror response status response reason body boto exception unauthorized authfailure aws was not able to validate the provided access credentials from my investigation the issue appears in the module utils py get creds which returns url boto params boto params region since the aws access key id in this example will only work with a security token the method is effectively broken i think the function should at least warn when it detects a security token or an access key that starts with asia instead of akia | 1 |
285,789 | 31,155,588,778 | IssuesEvent | 2023-08-16 12:56:53 | nidhi7598/linux-4.1.15_CVE-2018-5873 | https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2018-5873 | opened | CVE-2022-22764 (High) detected in linuxlinux-4.1.52 | Mend: dependency security vulnerability | ## CVE-2022-22764 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Mozilla developers Paul Adenot and the Mozilla Fuzzing Team reported memory safety bugs present in Firefox 96 and Firefox ESR 91.5. Some of these bugs showed evidence of memory corruption and we presume that with enough effort some of these could have been exploited to run arbitrary code. This vulnerability affects Firefox < 97, Thunderbird < 91.6, and Firefox ESR < 91.6.
<p>Publish Date: 2022-12-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22764>CVE-2022-22764</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-22764">https://nvd.nist.gov/vuln/detail/CVE-2022-22764</a></p>
<p>Release Date: 2022-01-07</p>
<p>Fix Resolution: linux-libc-headers - 5.14;linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-22764 (High) detected in linuxlinux-4.1.52 - ## CVE-2022-22764 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Mozilla developers Paul Adenot and the Mozilla Fuzzing Team reported memory safety bugs present in Firefox 96 and Firefox ESR 91.5. Some of these bugs showed evidence of memory corruption and we presume that with enough effort some of these could have been exploited to run arbitrary code. This vulnerability affects Firefox < 97, Thunderbird < 91.6, and Firefox ESR < 91.6.
<p>Publish Date: 2022-12-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22764>CVE-2022-22764</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-22764">https://nvd.nist.gov/vuln/detail/CVE-2022-22764</a></p>
<p>Release Date: 2022-01-07</p>
<p>Fix Resolution: linux-libc-headers - 5.14;linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files vulnerability details mozilla developers paul adenot and the mozilla fuzzing team reported memory safety bugs present in firefox and firefox esr some of these bugs showed evidence of memory corruption and we presume that with enough effort some of these could have been exploited to run arbitrary code this vulnerability affects firefox thunderbird and firefox esr publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux libc headers linux yocto gitautoinc gitautoinc step up your open source security game with mend | 0 |
32,686 | 8,921,078,995 | IssuesEvent | 2019-01-21 09:08:32 | neovim/neovim | https://api.github.com/repos/neovim/neovim | closed | (clang >= 6.0 bug) Annoying warnings for isnan(), fpclassify(), et al | blocked:external build help wanted | - `nvim --version`: ef4feab0e75be
- Vim (version: 8.0.1565) behaves differently? No warnings, but I didn't check if the same functions were used
- Operating system/version: arch linux
- Terminal name/version: pangoterm
- `$TERM`: xterm
### Steps to reproduce using `nvim -u NORC`
```
rm -rf build && CMAKE_EXTRA_FLAGS="-DCMAKE_C_COMPILER=clang -DCLANG_ASAN_UBSAN=1" make -j4
```
### Actual behaviour
```
[197/284] Building C object src/nvim/CMakeFiles/nvim.dir/eval/encode.c.o
In file included from ../src/nvim/eval/encode.c:455:
../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
ersion]
TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:455:
../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
version]
TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:493:
../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
ersion]
TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:493:
../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
version]
TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:762:
../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
ersion]
TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:762:
../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
version]
TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
6 warnings generated.
[261/284] Building C object src/nvim/CMakeFiles/nvim.dir/strings.c.o
../src/nvim/strings.c:1223:23: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion]
if (isinf((double)f)
~~~~~~^~~~~~~~~~
/usr/include/math.h:472:46: note: expanded from macro 'isinf'
# define isinf(x) __MATH_TG ((x), __isinf, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
../src/nvim/strings.c:1230:30: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion]
} else if (isnan(f)) {
~~~~~~^~
/usr/include/math.h:455:46: note: expanded from macro 'isnan'
# define isnan(x) __MATH_TG ((x), __isnan, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
2 warnings generated.
```
### Expected behaviour
No warnings. `maths.h` uses gcc magic to avoid this warning when compiling with gcc 4.4+, which doesn't work with clang (unless C11 is active, but using a different C version than the project's chosen one is probably a bad idea).
Anyone thinking of a better workaround than moving float code to a special `-Wno-conversion` c file?
| 1.0 | (clang >= 6.0 bug) Annoying warnings for isnan(), fpclassify(), et al - - `nvim --version`: ef4feab0e75be
- Vim (version: 8.0.1565) behaves differently? No warnings, but I didn't check if the same functions were used
- Operating system/version: arch linux
- Terminal name/version: pangoterm
- `$TERM`: xterm
### Steps to reproduce using `nvim -u NORC`
```
rm -rf build && CMAKE_EXTRA_FLAGS="-DCMAKE_C_COMPILER=clang -DCLANG_ASAN_UBSAN=1" make -j4
```
### Actual behaviour
```
[197/284] Building C object src/nvim/CMakeFiles/nvim.dir/eval/encode.c.o
In file included from ../src/nvim/eval/encode.c:455:
../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
ersion]
TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:455:
../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
version]
TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:493:
../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
ersion]
TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:493:
../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
version]
TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:330:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:762:
../src/nvim/eval/typval_encode.c.h:330:7: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wconv
ersion]
TYPVAL_ENCODE_CONV_FLOAT(tv, tv->vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
In file included from ../src/nvim/eval/encode.c:762:
../src/nvim/eval/typval_encode.c.h:491:13: warning: implicit conversion loses floating-point precision: 'const float_T' (aka 'const double') to 'float' [-Wcon
version]
TYPVAL_ENCODE_CONV_FLOAT(tv, val_di->di_tv.vval.v_float);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../src/nvim/eval/encode.c:534:26: note: expanded from macro 'TYPVAL_ENCODE_CONV_FLOAT'
switch (fpclassify(flt_)) { \
~~~~~~~~~~~^~~~~
/usr/include/math.h:415:56: note: expanded from macro 'fpclassify'
# define fpclassify(x) __MATH_TG ((x), __fpclassify, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
6 warnings generated.
[261/284] Building C object src/nvim/CMakeFiles/nvim.dir/strings.c.o
../src/nvim/strings.c:1223:23: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion]
if (isinf((double)f)
~~~~~~^~~~~~~~~~
/usr/include/math.h:472:46: note: expanded from macro 'isinf'
# define isinf(x) __MATH_TG ((x), __isinf, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
../src/nvim/strings.c:1230:30: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wconversion]
} else if (isnan(f)) {
~~~~~~^~
/usr/include/math.h:455:46: note: expanded from macro 'isnan'
# define isnan(x) __MATH_TG ((x), __isnan, (x))
~~~~~~~~~~~~~~~~~~~~~~~~~~^~~
/usr/include/math.h:370:16: note: expanded from macro '__MATH_TG'
? FUNC ## f ARGS \
~~~~~~~~~ ^~~~
2 warnings generated.
```
### Expected behaviour
No warnings. `maths.h` uses gcc magic to avoid this warning when compiling with gcc 4.4+, which doesn't work with clang (unless C11 is active, but using a different C version than the project's chosen one is probably a bad idea).
Anyone thinking of a better workaround than moving float code to a special `-Wno-conversion` c file?
| non_main | clang bug annoying warnings for isnan fpclassify et al nvim version vim version behaves differently no warnings but i didn t check if the same functions were used operating system version arch linux terminal name version pangoterm term xterm steps to reproduce using nvim u norc rm rf build cmake extra flags dcmake c compiler clang dclang asan ubsan make actual behaviour building c object src nvim cmakefiles nvim dir eval encode c o in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wconv ersion typval encode conv float tv tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args in file included from src nvim eval encode c src nvim eval typval encode c h warning implicit conversion loses floating point precision const float t aka const double to float wcon version typval encode conv float tv val di di tv vval v float src nvim eval encode c note expanded from macro typval encode conv float switch fpclassify flt usr include math h note expanded from macro fpclassify define fpclassify x math tg x fpclassify x usr include math h note expanded from macro math tg func f args warnings generated building c object src nvim cmakefiles nvim dir strings c o src nvim strings c warning implicit conversion loses floating point precision double to float if isinf double f usr include math h note expanded from macro isinf define isinf x math tg x isinf x usr include math h note expanded from macro math tg func f args src nvim strings c warning implicit conversion loses floating point precision double to float else if isnan f usr include math h note expanded from macro isnan define isnan x math tg x isnan x usr include math h note expanded from macro math tg func f args warnings generated expected behaviour no warnings maths h uses gcc magic to avoid this warning when compiling with gcc which doesn t work with clang unless is active but using a different c version than the project s chosen one is probably a bad idea anyone thinking of a better workaround than moving float code to a special wno conversion c file | 0 |
2,654 | 9,083,455,570 | IssuesEvent | 2019-02-17 20:37:25 | pound-python/infobob | https://api.github.com/repos/pound-python/infobob | closed | Remove unused commands | maintainability | `infobob_exec`, `infobob_codepad`, and `infobob_print` can be removed, they haven't been active in a while, and it'd doubtful they even work still. | True | Remove unused commands - `infobob_exec`, `infobob_codepad`, and `infobob_print` can be removed, they haven't been active in a while, and it'd doubtful they even work still. | main | remove unused commands infobob exec infobob codepad and infobob print can be removed they haven t been active in a while and it d doubtful they even work still | 1 |
428 | 3,516,654,646 | IssuesEvent | 2016-01-12 01:03:29 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | closed | Java cask fails to install | awaiting maintainer feedback bug cask | ### Description of issue
Java cask will not install.
### Output of `brew cask install java --verbose`
```
$ brew cask install java --verbose
==> Caveats
This Cask makes minor modifications to the JRE to prevent issues with
packaged applications, as discussed here:
https://bugs.eclipse.org/bugs/show_bug.cgi?id=411361
If your Java application still asks for JRE installation, you might need
to reboot or logout/login.
Installing this Cask means you have AGREED to the Oracle Binary Code
License Agreement for Java SE at
http://www.oracle.com/technetwork/java/javase/terms/license/index.html
==> Downloading http://download.oracle.com/otn-pub/java/jdk/8u66-b17/jdk-8u66-macosx-x64.dmg
Already downloaded: /Library/Caches/Homebrew/java-1.8.0_66-b17.dmg
Error: Permission denied - (/Library/Caches/Homebrew/java-1.8.0_66-b17.dmg, /Library/Caches/Homebrew/Casks/java-1.8.0_66-b17.dmg)
Most likely, this means you have an outdated version of homebrew-cask. Please run:
brew uninstall --force brew-cask; brew untap phinze/cask; brew update; brew cleanup; brew cask cleanup
If this doesn’t fix the problem, please report this bug:
https://github.com/caskroom/homebrew-cask#reporting-bugs
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:349:in `symlink'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:349:in `block in ln_s'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:1574:in `fu_each_src_dest0'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:347:in `ln_s'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:370:in `ln_sf'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/download.rb:45:in `create_cache_symlink'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/download.rb:15:in `perform'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/installer.rb:94:in `download'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/installer.rb:68:in `install'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:20:in `block in install_casks'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:17:in `each'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:17:in `install_casks'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:6:in `run'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli.rb:78:in `run_command'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli.rb:116:in `process'
/usr/local/Library/Taps/caskroom/homebrew-cask/cmd/brew-cask.rb:26:in `<top (required)>'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:55:in `require'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:55:in `require'
/usr/local/Library/brew.rb:69:in `require?'
/usr/local/Library/brew.rb:149:in `<main>'
Error: Kernel.exit
```
### Output of `brew doctor`
```
$ brew doctor
Your system is ready to brew.
```
### Output of `brew cask doctor`
```
$ brew cask doctor
==> OS X Release:
10.11
==> OS X Release with Patchlevel:
10.11.1
==> Hardware Architecture:
intel-64
==> Ruby Version:
2.0.0-p645
==> Ruby Path:
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/bin/ruby
==> Homebrew Version:
Homebrew 0.9.5 (git revision 3f86; last commit 2016-01-07)
==> Homebrew Executable Path:
/usr/local/bin/brew
==> Homebrew Cellar Path:
/usr/local/Cellar
==> Homebrew Repository Path:
/usr/local
==> Homebrew Origin:
https://github.com/Homebrew/homebrew
==> Homebrew-cask Version:
0.60.0 (git revision a595; last commit 41 minutes ago)
==> Homebrew-cask Install Location:
<NONE>
==> Homebrew-cask Staging Location:
/opt/homebrew-cask/Caskroom
==> Homebrew-cask Cached Downloads:
/Library/Caches/Homebrew
/Library/Caches/Homebrew/Casks
0 files
0.00 megs
==> Homebrew-cask Default Tap Path:
/usr/local/Library/Taps/caskroom/homebrew-cask
==> Homebrew-cask Alternate Cask Taps:
<NONE>
==> Homebrew-cask Default Tap Cask Count:
2941
==> Contents of $LOAD_PATH:
/usr/local/Library/Taps/caskroom/homebrew-cask/lib
/usr/local/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin15
/Library/Ruby/Site/2.0.0/universal-darwin15
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin15
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin15
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin15
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin15
==> Contents of $RUBYLIB Environment Variable:
<NONE>
==> Contents of $RUBYOPT Environment Variable:
<NONE>
==> Contents of $RUBYPATH Environment Variable:
<NONE>
==> Contents of $RBENV_VERSION Environment Variable:
<NONE>
==> Contents of $CHRUBY_VERSION Environment Variable:
<NONE>
==> Contents of $GEM_HOME Environment Variable:
<NONE>
==> Contents of $GEM_PATH Environment Variable:
<NONE>
==> Contents of $BUNDLE_PATH Environment Variable:
<NONE>
==> Contents of $PATH Environment Variable:
PATH="/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/Library/Taps/caskroom/homebrew-cask/cmd:/usr/local/Library/ENV/scm"
==> Contents of $SHELL Environment Variable:
SHELL="/bin/bash"
==> Contents of Locale Environment Variables:
LANG="en_US.UTF-8"
==> Running As Privileged User:
No
```
| True | Java cask fails to install - ### Description of issue
Java cask will not install.
### Output of `brew cask install java --verbose`
```
$ brew cask install java --verbose
==> Caveats
This Cask makes minor modifications to the JRE to prevent issues with
packaged applications, as discussed here:
https://bugs.eclipse.org/bugs/show_bug.cgi?id=411361
If your Java application still asks for JRE installation, you might need
to reboot or logout/login.
Installing this Cask means you have AGREED to the Oracle Binary Code
License Agreement for Java SE at
http://www.oracle.com/technetwork/java/javase/terms/license/index.html
==> Downloading http://download.oracle.com/otn-pub/java/jdk/8u66-b17/jdk-8u66-macosx-x64.dmg
Already downloaded: /Library/Caches/Homebrew/java-1.8.0_66-b17.dmg
Error: Permission denied - (/Library/Caches/Homebrew/java-1.8.0_66-b17.dmg, /Library/Caches/Homebrew/Casks/java-1.8.0_66-b17.dmg)
Most likely, this means you have an outdated version of homebrew-cask. Please run:
brew uninstall --force brew-cask; brew untap phinze/cask; brew update; brew cleanup; brew cask cleanup
If this doesn’t fix the problem, please report this bug:
https://github.com/caskroom/homebrew-cask#reporting-bugs
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:349:in `symlink'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:349:in `block in ln_s'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:1574:in `fu_each_src_dest0'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:347:in `ln_s'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/fileutils.rb:370:in `ln_sf'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/download.rb:45:in `create_cache_symlink'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/download.rb:15:in `perform'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/installer.rb:94:in `download'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/installer.rb:68:in `install'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:20:in `block in install_casks'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:17:in `each'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:17:in `install_casks'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli/install.rb:6:in `run'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli.rb:78:in `run_command'
/usr/local/Library/Taps/caskroom/homebrew-cask/lib/hbc/cli.rb:116:in `process'
/usr/local/Library/Taps/caskroom/homebrew-cask/cmd/brew-cask.rb:26:in `<top (required)>'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:55:in `require'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:55:in `require'
/usr/local/Library/brew.rb:69:in `require?'
/usr/local/Library/brew.rb:149:in `<main>'
Error: Kernel.exit
```
### Output of `brew doctor`
```
$ brew doctor
Your system is ready to brew.
```
### Output of `brew cask doctor`
```
$ brew cask doctor
==> OS X Release:
10.11
==> OS X Release with Patchlevel:
10.11.1
==> Hardware Architecture:
intel-64
==> Ruby Version:
2.0.0-p645
==> Ruby Path:
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/bin/ruby
==> Homebrew Version:
Homebrew 0.9.5 (git revision 3f86; last commit 2016-01-07)
==> Homebrew Executable Path:
/usr/local/bin/brew
==> Homebrew Cellar Path:
/usr/local/Cellar
==> Homebrew Repository Path:
/usr/local
==> Homebrew Origin:
https://github.com/Homebrew/homebrew
==> Homebrew-cask Version:
0.60.0 (git revision a595; last commit 41 minutes ago)
==> Homebrew-cask Install Location:
<NONE>
==> Homebrew-cask Staging Location:
/opt/homebrew-cask/Caskroom
==> Homebrew-cask Cached Downloads:
/Library/Caches/Homebrew
/Library/Caches/Homebrew/Casks
0 files
0.00 megs
==> Homebrew-cask Default Tap Path:
/usr/local/Library/Taps/caskroom/homebrew-cask
==> Homebrew-cask Alternate Cask Taps:
<NONE>
==> Homebrew-cask Default Tap Cask Count:
2941
==> Contents of $LOAD_PATH:
/usr/local/Library/Taps/caskroom/homebrew-cask/lib
/usr/local/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin15
/Library/Ruby/Site/2.0.0/universal-darwin15
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin15
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin15
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin15
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin15
==> Contents of $RUBYLIB Environment Variable:
<NONE>
==> Contents of $RUBYOPT Environment Variable:
<NONE>
==> Contents of $RUBYPATH Environment Variable:
<NONE>
==> Contents of $RBENV_VERSION Environment Variable:
<NONE>
==> Contents of $CHRUBY_VERSION Environment Variable:
<NONE>
==> Contents of $GEM_HOME Environment Variable:
<NONE>
==> Contents of $GEM_PATH Environment Variable:
<NONE>
==> Contents of $BUNDLE_PATH Environment Variable:
<NONE>
==> Contents of $PATH Environment Variable:
PATH="/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/Library/Taps/caskroom/homebrew-cask/cmd:/usr/local/Library/ENV/scm"
==> Contents of $SHELL Environment Variable:
SHELL="/bin/bash"
==> Contents of Locale Environment Variables:
LANG="en_US.UTF-8"
==> Running As Privileged User:
No
```
| main | java cask fails to install description of issue java cask will not install output of brew cask install java verbose brew cask install java verbose caveats this cask makes minor modifications to the jre to prevent issues with packaged applications as discussed here if your java application still asks for jre installation you might need to reboot or logout login installing this cask means you have agreed to the oracle binary code license agreement for java se at downloading already downloaded library caches homebrew java dmg error permission denied library caches homebrew java dmg library caches homebrew casks java dmg most likely this means you have an outdated version of homebrew cask please run brew uninstall force brew cask brew untap phinze cask brew update brew cleanup brew cask cleanup if this doesn’t fix the problem please report this bug system library frameworks ruby framework versions usr lib ruby fileutils rb in symlink system library frameworks ruby framework versions usr lib ruby fileutils rb in block in ln s system library frameworks ruby framework versions usr lib ruby fileutils rb in fu each src system library frameworks ruby framework versions usr lib ruby fileutils rb in ln s system library frameworks ruby framework versions usr lib ruby fileutils rb in ln sf usr local library taps caskroom homebrew cask lib hbc download rb in create cache symlink usr local library taps caskroom homebrew cask lib hbc download rb in perform usr local library taps caskroom homebrew cask lib hbc installer rb in download usr local library taps caskroom homebrew cask lib hbc installer rb in install usr local library taps caskroom homebrew cask lib hbc cli install rb in block in install casks usr local library taps caskroom homebrew cask lib hbc cli install rb in each usr local library taps caskroom homebrew cask lib hbc cli install rb in install casks usr local library taps caskroom homebrew cask lib hbc cli install rb in run usr local library taps caskroom homebrew cask lib hbc cli rb in run command usr local library taps caskroom homebrew cask lib hbc cli rb in process usr local library taps caskroom homebrew cask cmd brew cask rb in system library frameworks ruby framework versions usr lib ruby rubygems core ext kernel require rb in require system library frameworks ruby framework versions usr lib ruby rubygems core ext kernel require rb in require usr local library brew rb in require usr local library brew rb in error kernel exit output of brew doctor brew doctor your system is ready to brew output of brew cask doctor brew cask doctor os x release os x release with patchlevel hardware architecture intel ruby version ruby path system library frameworks ruby framework versions usr bin ruby homebrew version homebrew git revision last commit homebrew executable path usr local bin brew homebrew cellar path usr local cellar homebrew repository path usr local homebrew origin homebrew cask version git revision last commit minutes ago homebrew cask install location homebrew cask staging location opt homebrew cask caskroom homebrew cask cached downloads library caches homebrew library caches homebrew casks files megs homebrew cask default tap path usr local library taps caskroom homebrew cask homebrew cask alternate cask taps homebrew cask default tap cask count contents of load path usr local library taps caskroom homebrew cask lib usr local library homebrew library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal contents of rubylib environment variable contents of rubyopt environment variable contents of rubypath environment variable contents of rbenv version environment variable contents of chruby version environment variable contents of gem home environment variable contents of gem path environment variable contents of bundle path environment variable contents of path environment variable path usr local bin usr bin bin usr sbin sbin usr local library taps caskroom homebrew cask cmd usr local library env scm contents of shell environment variable shell bin bash contents of locale environment variables lang en us utf running as privileged user no | 1 |
101,110 | 30,876,042,897 | IssuesEvent | 2023-08-03 14:21:31 | helidon-io/helidon | https://api.github.com/repos/helidon-io/helidon | opened | [4.x] `Prototype.BuilderInterceptor` was renamed to `Prototype.BuilderDecorator`; the builder README and `@Prototype.Blueprint` attributes are now inconsistent with the change | 4.x builder | ## Environment Details
* Helidon Version: 4.x
* Helidon SE or Helidon MP
* JDK version:
* OS:
* Docker version (if applicable):
----------
## Problem Description
A recent PR #7281 renamed `Prototype.BuilderInterceptor` to `Prototype.BuilderDecorator`, but:
* the builder `README.md` was not updated
* the `@Prototype.Blueprint` annotation still uses the `builderInterceptor` attribute name.
| 1.0 | [4.x] `Prototype.BuilderInterceptor` was renamed to `Prototype.BuilderDecorator`; the builder README and `@Prototype.Blueprint` attributes are now inconsistent with the change - ## Environment Details
* Helidon Version: 4.x
* Helidon SE or Helidon MP
* JDK version:
* OS:
* Docker version (if applicable):
----------
## Problem Description
A recent PR #7281 renamed `Prototype.BuilderInterceptor` to `Prototype.BuilderDecorator`, but:
* the builder `README.md` was not updated
* the `@Prototype.Blueprint` annotation still uses the `builderInterceptor` attribute name.
| non_main | prototype builderinterceptor was renamed to prototype builderdecorator the builder readme and prototype blueprint attributes are now inconsistent with the change environment details helidon version x helidon se or helidon mp jdk version os docker version if applicable problem description a recent pr renamed prototype builderinterceptor to prototype builderdecorator but the builder readme md was not updated the prototype blueprint annotation still uses the builderinterceptor attribute name | 0 |
45,917 | 13,055,821,972 | IssuesEvent | 2020-07-30 02:50:20 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | libarchive review (Trac #261) | Incomplete Migration Migrated from Trac combo core defect | Migrated from https://code.icecube.wisc.edu/ticket/261
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:57",
"description": "See: #IT282",
"reporter": "nega",
"cc": "",
"resolution": "worksforme",
"_ts": "1416713877165085",
"component": "combo core",
"summary": "libarchive review",
"priority": "normal",
"keywords": "libarchive",
"time": "2011-05-11T20:39:48",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| 1.0 | libarchive review (Trac #261) - Migrated from https://code.icecube.wisc.edu/ticket/261
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:57",
"description": "See: #IT282",
"reporter": "nega",
"cc": "",
"resolution": "worksforme",
"_ts": "1416713877165085",
"component": "combo core",
"summary": "libarchive review",
"priority": "normal",
"keywords": "libarchive",
"time": "2011-05-11T20:39:48",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| non_main | libarchive review trac migrated from json status closed changetime description see reporter nega cc resolution worksforme ts component combo core summary libarchive review priority normal keywords libarchive time milestone owner olivas type defect | 0 |
75,271 | 7,467,696,132 | IssuesEvent | 2018-04-02 16:15:37 | GoogleCloudPlatform/forseti-security | https://api.github.com/repos/GoogleCloudPlatform/forseti-security | closed | Enable the IAM and firewall notification by default. | module: notifier priority: p1 release-testing: 2.0 RC2 | This is currently disabled. But after removing the notification from the scanner, this needs to be enabled.
https://github.com/GoogleCloudPlatform/forseti-security/blob/2.0-dev/configs/server/forseti_conf_server.yaml.in#L85 | 1.0 | Enable the IAM and firewall notification by default. - This is currently disabled. But after removing the notification from the scanner, this needs to be enabled.
https://github.com/GoogleCloudPlatform/forseti-security/blob/2.0-dev/configs/server/forseti_conf_server.yaml.in#L85 | non_main | enable the iam and firewall notification by default this is currently disabled but after removing the notification from the scanner this needs to be enabled | 0 |
296,732 | 22,321,314,747 | IssuesEvent | 2022-06-14 06:43:34 | elastic/package-spec | https://api.github.com/repos/elastic/package-spec | opened | Create implementation guide for the spec | documentation | The package spec contains descriptions of resources that can be installed in the Elastic Stack. Some resources need special management that is not always trivial ([example](https://github.com/elastic/package-spec/pull/307/files#r895515041)). It may be good to maintain in the spec a reference documentation for clients that implement the package spec, as Fleet.
This documentation may contain a list of general practices for the resources defined, as well as resource-specific considerations.
We may need to target different clients, e.g. installation use cases like in fleet vs. inspection use cases as in elastic-package or the registries. | 1.0 | Create implementation guide for the spec - The package spec contains descriptions of resources that can be installed in the Elastic Stack. Some resources need special management that is not always trivial ([example](https://github.com/elastic/package-spec/pull/307/files#r895515041)). It may be good to maintain in the spec a reference documentation for clients that implement the package spec, as Fleet.
This documentation may contain a list of general practices for the resources defined, as well as resource-specific considerations.
We may need to target different clients, e.g. installation use cases like in fleet vs. inspection use cases as in elastic-package or the registries. | non_main | create implementation guide for the spec the package spec contains descriptions of resources that can be installed in the elastic stack some resources need special management that is not always trivial it may be good to maintain in the spec a reference documentation for clients that implement the package spec as fleet this documentation may contain a list of general practices for the resources defined as well as resource specific considerations we may need to target different clients e g installation use cases like in fleet vs inspection use cases as in elastic package or the registries | 0 |
101,828 | 4,141,752,446 | IssuesEvent | 2016-06-14 07:20:38 | qspin/qtaste | https://api.github.com/repos/qspin/qtaste | closed | QTaste doesn't generate all Test API documentation anymore | category:kernel priority:low status:To be validated type:update | In QTaste 2.1.0 and current 2.2.0-SNAPSHOT, the Test API documentation generated in /target/TestAPI-doc/ doesn't include the documentation of the other Test API on which the Test API depends anymore.
E.g. at BMS level, I had the documentation of BMS, BDS, BSS, ICOMP, QTaste-Toolbox, etc..., now I only have BMS documentation.
There is a maven inheritance configuration issue in qtaste-testapi-parent:
the configuration of maven-dependency-plugin is not inherited anymore!
Also even if I override the configuration myself, JavaGUI is not included because the package name doesn't match the includes:
<includes>**/qtaste/testapi/api/*.java</includes>
must be changed to:
<includes>**/qtaste/**/testapi/api/*.java</includes> | 1.0 | QTaste doesn't generate all Test API documentation anymore - In QTaste 2.1.0 and current 2.2.0-SNAPSHOT, the Test API documentation generated in /target/TestAPI-doc/ doesn't include the documentation of the other Test API on which the Test API depends anymore.
E.g. at BMS level, I had the documentation of BMS, BDS, BSS, ICOMP, QTaste-Toolbox, etc..., now I only have BMS documentation.
There is a maven inheritance configuration issue in qtaste-testapi-parent:
the configuration of maven-dependency-plugin is not inherited anymore!
Also even if I override the configuration myself, JavaGUI is not included because the package name doesn't match the includes:
<includes>**/qtaste/testapi/api/*.java</includes>
must be changed to:
<includes>**/qtaste/**/testapi/api/*.java</includes> | non_main | qtaste doesn t generate all test api documentation anymore in qtaste and current snapshot the test api documentation generated in target testapi doc doesn t include the documentation of the other test api on which the test api depends anymore e g at bms level i had the documentation of bms bds bss icomp qtaste toolbox etc now i only have bms documentation there is a maven inheritance configuration issue in qtaste testapi parent the configuration of maven dependency plugin is not inherited anymore also even if i override the configuration myself javagui is not included because the package name doesn t match the includes lt includes gt qtaste testapi api java lt includes gt must be changed to lt includes gt qtaste testapi api java lt includes gt | 0 |
1,306 | 5,546,361,385 | IssuesEvent | 2017-03-23 00:35:46 | OpenLightingProject/ola | https://api.github.com/repos/OpenLightingProject/ola | opened | Packing structs on Windows | Difficulty-Easy Language-C++ Maintainability OpSys-Windows | Do it the same way Jack does:
https://github.com/jackaudio/jack2/blob/master/common/jack/systemdeps.h
So we can then add that code to here:
https://github.com/OpenLightingProject/ola/blob/master/include/ola/base/Macro.h#L139
And use it whenever we're packing structs with PRE_PACK and POST_PACK. | True | Packing structs on Windows - Do it the same way Jack does:
https://github.com/jackaudio/jack2/blob/master/common/jack/systemdeps.h
So we can then add that code to here:
https://github.com/OpenLightingProject/ola/blob/master/include/ola/base/Macro.h#L139
And use it whenever we're packing structs with PRE_PACK and POST_PACK. | main | packing structs on windows do it the same way jack does so we can then add that code to here and use it whenever we re packing structs with pre pack and post pack | 1 |
1,898 | 6,577,549,525 | IssuesEvent | 2017-09-12 01:41:36 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | while running a copy module task, ansible requested this info passed to the list. | affects_2.0 bug_report waiting_on_maintainer | ##### Issue Type:
- Bug Report
##### Plugin Name:
<!-- Name of the plugin/module/task -->
##### Ansible Version:
```
23:06 $ ansible --version
ansible 2.0.0.2
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### Ansible Configuration:
Dynamic inventory from dopy
```
[ssh_connection]
scp_if_ssh=True
ssh_args= -o ForwardAgent=yes
```
##### Environment:
OSX 10.10 host
Ubuntu 14.04 remote
##### Summary:
While running playbook, ansible reported back
```
[WARNING]: Calculating checksum failed unusually, please report this to the list so it can be fixed command: rc=flag; [ -r
/etc/apt/apt.conf.d/10periodic ] || rc=2; [ -f /etc/apt/apt.conf.d/10periodic ] || rc=1; [ -d /etc/apt/apt.conf.d/10periodic ] && rc=3; python -V
2>/dev/null || rc=4; [ x"$rc" != "xflag" ] && echo "${rc} "/etc/apt/apt.conf.d/10periodic && exit 0; (python -c 'import hashlib; BLOCKSIZE =
65536; hasher = hashlib.sha1(); afile = open("'/etc/apt/apt.conf.d/10periodic'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0:
hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (python -c 'import sha; BLOCKSIZE
= 65536; hasher = sha.sha(); afile = open("'/etc/apt/apt.conf.d/10periodic'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0:
hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (echo '0
'/etc/apt/apt.conf.d/10periodic) ---- output: {'stdout_lines': [], 'stdout': u'', 'stderr': u'', 'rc': 255} ----
```
##### Steps To Reproduce:
Unable to reproduce. Reporting because ansible said to.
```
- name: Adjust APT update intervals
copy:
src: config/apt_periodic
dest: /etc/apt/apt.conf.d/10periodic
```
| True | while running a copy module task, ansible requested this info passed to the list. - ##### Issue Type:
- Bug Report
##### Plugin Name:
<!-- Name of the plugin/module/task -->
##### Ansible Version:
```
23:06 $ ansible --version
ansible 2.0.0.2
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### Ansible Configuration:
Dynamic inventory from dopy
```
[ssh_connection]
scp_if_ssh=True
ssh_args= -o ForwardAgent=yes
```
##### Environment:
OSX 10.10 host
Ubuntu 14.04 remote
##### Summary:
While running playbook, ansible reported back
```
[WARNING]: Calculating checksum failed unusually, please report this to the list so it can be fixed command: rc=flag; [ -r
/etc/apt/apt.conf.d/10periodic ] || rc=2; [ -f /etc/apt/apt.conf.d/10periodic ] || rc=1; [ -d /etc/apt/apt.conf.d/10periodic ] && rc=3; python -V
2>/dev/null || rc=4; [ x"$rc" != "xflag" ] && echo "${rc} "/etc/apt/apt.conf.d/10periodic && exit 0; (python -c 'import hashlib; BLOCKSIZE =
65536; hasher = hashlib.sha1(); afile = open("'/etc/apt/apt.conf.d/10periodic'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0:
hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (python -c 'import sha; BLOCKSIZE
= 65536; hasher = sha.sha(); afile = open("'/etc/apt/apt.conf.d/10periodic'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0:
hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (echo '0
'/etc/apt/apt.conf.d/10periodic) ---- output: {'stdout_lines': [], 'stdout': u'', 'stderr': u'', 'rc': 255} ----
```
##### Steps To Reproduce:
Unable to reproduce. Reporting because ansible said to.
```
- name: Adjust APT update intervals
copy:
src: config/apt_periodic
dest: /etc/apt/apt.conf.d/10periodic
```
| main | while running a copy module task ansible requested this info passed to the list issue type bug report plugin name ansible version ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides ansible configuration dynamic inventory from dopy scp if ssh true ssh args o forwardagent yes environment osx host ubuntu remote summary while running playbook ansible reported back calculating checksum failed unusually please report this to the list so it can be fixed command rc flag r etc apt apt conf d rc rc rc python v dev null rc echo rc etc apt apt conf d exit python c import hashlib blocksize hasher hashlib afile open etc apt apt conf d rb buf afile read blocksize while len buf hasher update buf buf afile read blocksize afile close print hasher hexdigest dev null python c import sha blocksize hasher sha sha afile open etc apt apt conf d rb buf afile read blocksize while len buf hasher update buf buf afile read blocksize afile close print hasher hexdigest dev null echo etc apt apt conf d output stdout lines stdout u stderr u rc steps to reproduce unable to reproduce reporting because ansible said to name adjust apt update intervals copy src config apt periodic dest etc apt apt conf d | 1 |
82,161 | 3,603,658,295 | IssuesEvent | 2016-02-03 19:52:42 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | opened | Make Kubernetes more bulletproof to gcloud failure | area/cluster-lifecycle priority/P1 team/control-plane team/gke | #20306 was one instance of this, but @brendandburns says there are multiple places in kube-up and elsewhere that we need to also subject to the same hygiene.
| 1.0 | Make Kubernetes more bulletproof to gcloud failure - #20306 was one instance of this, but @brendandburns says there are multiple places in kube-up and elsewhere that we need to also subject to the same hygiene.
| non_main | make kubernetes more bulletproof to gcloud failure was one instance of this but brendandburns says there are multiple places in kube up and elsewhere that we need to also subject to the same hygiene | 0 |
426,249 | 12,369,822,939 | IssuesEvent | 2020-05-18 15:49:22 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | Add force IPV4 field to cloud view | Category: Sunstone Priority: Low Sponsored Status: Accepted Type: Feature | **Description**
The force IPV4 input text doesn't appear in the cloud view.
**Use case**
Users can force an IPV4.
**Interface Changes**
Sunstone
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| 1.0 | Add force IPV4 field to cloud view - **Description**
The force IPV4 input text doesn't appear in the cloud view.
**Use case**
Users can force an IPV4.
**Interface Changes**
Sunstone
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| non_main | add force field to cloud view description the force input text doesn t appear in the cloud view use case users can force an interface changes sunstone progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches | 0 |
51,293 | 7,694,084,490 | IssuesEvent | 2018-05-18 07:27:16 | pombase/website | https://api.github.com/repos/pombase/website | closed | add a link to help section from | documentation gene page |
Would it be useful to have a direct help link
https://www.pombase.org/documentation/advanced-search
from here
https://www.pombase.org/query | 1.0 | add a link to help section from -
Would it be useful to have a direct help link
https://www.pombase.org/documentation/advanced-search
from here
https://www.pombase.org/query | non_main | add a link to help section from would it be useful to have a direct help link from here | 0 |
829,414 | 31,878,881,502 | IssuesEvent | 2023-09-16 05:55:41 | roq-trading/roq-issues | https://api.github.com/repos/roq-trading/roq-issues | closed | [roq-cme] TCP fragmentation could cause freeze | bug high priority support | iLink SBE parsing incorrectly assumed messages were received in full.
The socket read code was based on MDP SBE parsing which is based on UDP where datagrams are received in full (or nothing at all). Fragmentation is not possible.
However, iLink is based on TCP where messages are allowed to be fragmented (by the TCP stream protocol).
Symptoms:
* Gateway freeze
Reason:
Message parsing happens inside a loop that processes the read stream buffer.
When the parser failed (due to message length) it would return 0 (bytes) to indicate that the read buffer did not yet have enough data.
Unfortunately, the loop would only check on read buffer being fully processed by checking when the read buffer became empty.
This resulted in an infinite loop.
Solution:
Break the loop when not enough data is ready to process the next message. | 1.0 | [roq-cme] TCP fragmentation could cause freeze - iLink SBE parsing incorrectly assumed messages were received in full.
The socket read code was based on MDP SBE parsing which is based on UDP where datagrams are received in full (or nothing at all). Fragmentation is not possible.
However, iLink is based on TCP where messages are allowed to be fragmented (by the TCP stream protocol).
Symptoms:
* Gateway freeze
Reason:
Message parsing happens inside a loop that processes the read stream buffer.
When the parser failed (due to message length) it would return 0 (bytes) to indicate that the read buffer did not yet have enough data.
Unfortunately, the loop would only check on read buffer being fully processed by checking when the read buffer became empty.
This resulted in an infinite loop.
Solution:
Break the loop when not enough data is ready to process the next message. | non_main | tcp fragmentation could cause freeze ilink sbe parsing incorrectly assumed messages were received in full the socket read code was based on mdp sbe parsing which is based on udp where datagrams are received in full or nothing at all fragmentation is not possible however ilink is based on tcp where messages are allowed to be fragmented by the tcp stream protocol symptoms gateway freeze reason message parsing happens inside a loop that processes the read stream buffer when the parser failed due to message length it would return bytes to indicate that the read buffer did not yet have enough data unfortunately the loop would only check on read buffer being fully processed by checking when the read buffer became empty this resulted in an infinite loop solution break the loop when not enough data is ready to process the next message | 0 |
778,159 | 27,305,560,064 | IssuesEvent | 2023-02-24 07:51:41 | openforis/arena | https://api.github.com/repos/openforis/arena | closed | Copy/Clone records from cycle to another | Priority_1 | We assume that schema in the source and target cycles are identical.
If these are not identical, a direct cloning will be not possible. | 1.0 | Copy/Clone records from cycle to another - We assume that schema in the source and target cycles are identical.
If these are not identical, a direct cloning will be not possible. | non_main | copy clone records from cycle to another we assume that schema in the source and target cycles are identical if these are not identical a direct cloning will be not possible | 0 |
275,693 | 20,954,628,442 | IssuesEvent | 2022-03-27 00:20:15 | numpy/numpy | https://api.github.com/repos/numpy/numpy | closed | DOC: Reference Gitpod as simple alternative in building from source guide | 04 - Documentation sprint | ### Issue with current documentation:
I spent an hour trying and failing to get numpy to build from source.
I followed the documentation but failed to discover that I could just use Gitpod.
I only became aware of this as @mattip suggested it to me in a comment https://github.com/numpy/numpy/issues/20677#issuecomment-1002744924
### Idea or request for content:
The discoverability of the very helpful Gitpod setup could be improved by including a reference to it in this doc: https://numpy.org/doc/stable/dev/development_environment.html
I overlooked it here, but spent most of the time in the other doc, that's maybe why I missed it. Since Gitpod works so well and is not necessarily something one would expect to exist, having more pointers to it could help. | 1.0 | DOC: Reference Gitpod as simple alternative in building from source guide - ### Issue with current documentation:
I spent an hour trying and failing to get numpy to build from source.
I followed the documentation but failed to discover that I could just use Gitpod.
I only became aware of this as @mattip suggested it to me in a comment https://github.com/numpy/numpy/issues/20677#issuecomment-1002744924
### Idea or request for content:
The discoverability of the very helpful Gitpod setup could be improved by including a reference to it in this doc: https://numpy.org/doc/stable/dev/development_environment.html
I overlooked it here, but spent most of the time in the other doc, that's maybe why I missed it. Since Gitpod works so well and is not necessarily something one would expect to exist, having more pointers to it could help. | non_main | doc reference gitpod as simple alternative in building from source guide issue with current documentation i spent an hour trying and failing to get numpy to build from source i followed the documentation but failed to discover that i could just use gitpod i only became aware of this as mattip suggested it to me in a comment idea or request for content the discoverability of the very helpful gitpod setup could be improved by including a reference to it in this doc i overlooked it here but spent most of the time in the other doc that s maybe why i missed it since gitpod works so well and is not necessarily something one would expect to exist having more pointers to it could help | 0 |
281,658 | 8,697,938,874 | IssuesEvent | 2018-12-04 21:45:09 | mplusmuseum/mplusmuseum-collections-explorer | https://api.github.com/repos/mplusmuseum/mplusmuseum-collections-explorer | closed | Search - feature: Inline Search Filter 'Pills' | priority-low question | Convert a comma-separated list into these 'pill' UI elements. Clicking the 'X' should clear them.
@revdancatt do we know if this will be an "or" search or an "and" search?

| 1.0 | Search - feature: Inline Search Filter 'Pills' - Convert a comma-separated list into these 'pill' UI elements. Clicking the 'X' should clear them.
@revdancatt do we know if this will be an "or" search or an "and" search?

| non_main | search feature inline search filter pills convert a comma separated list into these pill ui elements clicking the x should clear them revdancatt do we know if this will be an or search or an and search | 0 |
3,860 | 17,025,697,920 | IssuesEvent | 2021-07-03 13:05:36 | CyanCoding/Project-Tracker | https://api.github.com/repos/CyanCoding/Project-Tracker | closed | Change public readonly variables to resources | backend maintainability | In the code we have a lot of static resources (such as URLs, GUIDs, file locations, etc) that never change. These are used at various points around the program. It would be great if we could move from having a ton of global variables that are always passed as parameters and move to using these resources. | True | Change public readonly variables to resources - In the code we have a lot of static resources (such as URLs, GUIDs, file locations, etc) that never change. These are used at various points around the program. It would be great if we could move from having a ton of global variables that are always passed as parameters and move to using these resources. | main | change public readonly variables to resources in the code we have a lot of static resources such as urls guids file locations etc that never change these are used at various points around the program it would be great if we could move from having a ton of global variables that are always passed as parameters and move to using these resources | 1 |
2,482 | 8,639,916,595 | IssuesEvent | 2018-11-23 22:41:10 | F5OEO/rpitx | https://api.github.com/repos/F5OEO/rpitx | closed | RPITX : transmitting from TCP streaming | V1 related (not maintained) | Good morning,
I'm facing a problem in transmitting using RPITX from TCP streaming (generated by gnuradio).
The scenario is the following :
a) Raspbian Stretch 4.14.44+ #1117 on RaspberryPI2B+ armv6l GNU/Linux
b) rpitx Version 0.2 compiled Jun 4 2018
c) rpitx test script transmitting correctly on 10m : /rpitx -m IQ -i /home/pi/pat/ssbIQ.wav -f 28150 -l
d) Gnuradio modulating the same sampleaudio.wav file in SSB / USB and sending it in complex format on TCP sink port (client mode). Waterfall shows modulated signal.
e) data streaming received by netcat and sent to rpitx as follow :
while true; do (nc -l 8011; dd if=/dev/zero bs=4096 count=30); done | sudo rpitx -i - -m IQ -f 28150
The result is that rpitx stays idle on the calibration phase and nothing is transmitted on air.
rpitx Version 0.2 compiled Jun 4 2018 (F5OEO Evariste) running on Frequency=0.000000 SampleRate=48000 Master PLL = 1000000000
memory: 512 MB
processor: Broadcom BCM2835
i2cDevice: /dev/i2c-1
model: Model B+
manufacturer: Sony
pcb revision: 1
warranty void: no
revision: 0010
peripheral base: 0x20000000
Jessie
Using mbox device /dev/vcio.
3616000 Size NUM PAGES 883 PAGE_SIZE 4096
MASH 3 Freq PLL# 6
Calibrate : ppm=-10 DMA 160ns:1361ns
... and no more.
What I'm doing wrong ?
Any help appreciated
Thank-you
Best Regards, 73
Ugo Poddine IU1IPB
| True | RPITX : transmitting from TCP streaming - Good morning,
I'm facing a problem in transmitting using RPITX from TCP streaming (generated by gnuradio).
The scenario is the following :
a) Raspbian Stretch 4.14.44+ #1117 on RaspberryPI2B+ armv6l GNU/Linux
b) rpitx Version 0.2 compiled Jun 4 2018
c) rpitx test script transmitting correctly on 10m : /rpitx -m IQ -i /home/pi/pat/ssbIQ.wav -f 28150 -l
d) Gnuradio modulating the same sampleaudio.wav file in SSB / USB and sending it in complex format on TCP sink port (client mode). Waterfall shows modulated signal.
e) data streaming received by netcat and sent to rpitx as follow :
while true; do (nc -l 8011; dd if=/dev/zero bs=4096 count=30); done | sudo rpitx -i - -m IQ -f 28150
The result is that rpitx stays idle on the calibration phase and nothing is transmitted on air.
rpitx Version 0.2 compiled Jun 4 2018 (F5OEO Evariste) running on Frequency=0.000000 SampleRate=48000 Master PLL = 1000000000
memory: 512 MB
processor: Broadcom BCM2835
i2cDevice: /dev/i2c-1
model: Model B+
manufacturer: Sony
pcb revision: 1
warranty void: no
revision: 0010
peripheral base: 0x20000000
Jessie
Using mbox device /dev/vcio.
3616000 Size NUM PAGES 883 PAGE_SIZE 4096
MASH 3 Freq PLL# 6
Calibrate : ppm=-10 DMA 160ns:1361ns
... and no more.
What I'm doing wrong ?
Any help appreciated
Thank-you
Best Regards, 73
Ugo Poddine IU1IPB
| main | rpitx transmitting from tcp streaming good morning i m facing a problem in transmitting using rpitx from tcp streaming generated by gnuradio the scenario is the following a raspbian stretch on gnu linux b rpitx version compiled jun c rpitx test script transmitting correctly on rpitx m iq i home pi pat ssbiq wav f l d gnuradio modulating the same sampleaudio wav file in ssb usb and sending it in complex format on tcp sink port client mode waterfall shows modulated signal e data streaming received by netcat and sent to rpitx as follow while true do nc l dd if dev zero bs count done sudo rpitx i m iq f the result is that rpitx stays idle on the calibration phase and nothing is transmitted on air rpitx version compiled jun evariste running on frequency samplerate master pll memory mb processor broadcom dev model model b manufacturer sony pcb revision warranty void no revision peripheral base jessie using mbox device dev vcio size num pages page size mash freq pll calibrate ppm dma and no more what i m doing wrong any help appreciated thank you best regards ugo poddine | 1 |
3,600 | 14,539,876,498 | IssuesEvent | 2020-12-15 12:31:49 | pace/bricks | https://api.github.com/repos/pace/bricks | opened | Revisit ordering of default RoundTripper Chain | S::Ready T::Maintainance | The default RoundTripper Chain ordering might not make sense right now, because certain more "important" roundtrippers are executed later in the stack, e.g., request id roundtripper, and other potentially too early, e.g. dump roundtripper. | True | Revisit ordering of default RoundTripper Chain - The default RoundTripper Chain ordering might not make sense right now, because certain more "important" roundtrippers are executed later in the stack, e.g., request id roundtripper, and other potentially too early, e.g. dump roundtripper. | main | revisit ordering of default roundtripper chain the default roundtripper chain ordering might not make sense right now because certain more important roundtrippers are executed later in the stack e g request id roundtripper and other potentially too early e g dump roundtripper | 1 |
4,602 | 23,849,803,478 | IssuesEvent | 2022-09-06 16:49:56 | ocsf/ocsf-schema | https://api.github.com/repos/ocsf/ocsf-schema | closed | All first release objects merged into main | maintainers | First release Objects, checkbox checked when merged into main.
- [ ] API Details
- [ ] Attack
- [ ] Authorization Information
- [ ] CVSS Scores
- [ ] Client
- [ ] Cloud
- [ ] Compliance Details
- [ ] Container
- [ ] DNS Answer
- [ ] DNS Query
- [ ] Device
- [ ] Device Hardware Info
- [ ] Digital Certificate
- [ ] Digital Signature
- [ ] Display
- [ ] Distributed Computing Environment/Remote Procedure Call (DCE/RPC)
- [ ] Endpoint
- [ ] Enrichment
- [ ] Event Origin
- [ ] Event Source
- [ ] Extended User
- [ ] Feature
- [ ] File
- [ ] Finding Details
- [ ] Fingerprint
- [ ] Geo Location
- [ ] Group
- [ ] HTTP Cookie
- [ ] HTTP Header
- [ ] HTTP Request
- [ ] HTTP Response
- [ ] Identity
- [ ] Identity Provider
- [ ] Image
- [ ] Job
- [ ] Kernel Resource
- [ ] Keyboard Information
- [ ] Malware
- [ ] Managed Entity
- [ ] Metadata
- [ ] Module
- [ ] Network Connection Information
- [ ] Network Endpoint
- [ ] Network Interface
- [ ] Network Proxy
- [ ] Network Traffic
- [ ] OS
- [ ] Object
- [ ] Observable
- [ ] Peripheral Device
- [ ] Process
- [ ] Product
- [ ] Registry Key
- [ ] Registry Value
- [ ] Related Findings
- [ ] Remediation Details
- [ ] Request Elements
- [ ] Resource
- [ ] Response Elements
- [ ] Rule
- [ ] Server
- [ ] Service
- [ ] Session
- [ ] Subject Alternative Name
- [ ] TLS Extension
- [ ] Transport Layer Security (TLS)
- [ ] Uniform Resource Locator
- [ ] User
- [ ] Vulnerability Details
| True | All first release objects merged into main - First release Objects, checkbox checked when merged into main.
- [ ] API Details
- [ ] Attack
- [ ] Authorization Information
- [ ] CVSS Scores
- [ ] Client
- [ ] Cloud
- [ ] Compliance Details
- [ ] Container
- [ ] DNS Answer
- [ ] DNS Query
- [ ] Device
- [ ] Device Hardware Info
- [ ] Digital Certificate
- [ ] Digital Signature
- [ ] Display
- [ ] Distributed Computing Environment/Remote Procedure Call (DCE/RPC)
- [ ] Endpoint
- [ ] Enrichment
- [ ] Event Origin
- [ ] Event Source
- [ ] Extended User
- [ ] Feature
- [ ] File
- [ ] Finding Details
- [ ] Fingerprint
- [ ] Geo Location
- [ ] Group
- [ ] HTTP Cookie
- [ ] HTTP Header
- [ ] HTTP Request
- [ ] HTTP Response
- [ ] Identity
- [ ] Identity Provider
- [ ] Image
- [ ] Job
- [ ] Kernel Resource
- [ ] Keyboard Information
- [ ] Malware
- [ ] Managed Entity
- [ ] Metadata
- [ ] Module
- [ ] Network Connection Information
- [ ] Network Endpoint
- [ ] Network Interface
- [ ] Network Proxy
- [ ] Network Traffic
- [ ] OS
- [ ] Object
- [ ] Observable
- [ ] Peripheral Device
- [ ] Process
- [ ] Product
- [ ] Registry Key
- [ ] Registry Value
- [ ] Related Findings
- [ ] Remediation Details
- [ ] Request Elements
- [ ] Resource
- [ ] Response Elements
- [ ] Rule
- [ ] Server
- [ ] Service
- [ ] Session
- [ ] Subject Alternative Name
- [ ] TLS Extension
- [ ] Transport Layer Security (TLS)
- [ ] Uniform Resource Locator
- [ ] User
- [ ] Vulnerability Details
| main | all first release objects merged into main first release objects checkbox checked when merged into main api details attack authorization information cvss scores client cloud compliance details container dns answer dns query device device hardware info digital certificate digital signature display distributed computing environment remote procedure call dce rpc endpoint enrichment event origin event source extended user feature file finding details fingerprint geo location group http cookie http header http request http response identity identity provider image job kernel resource keyboard information malware managed entity metadata module network connection information network endpoint network interface network proxy network traffic os object observable peripheral device process product registry key registry value related findings remediation details request elements resource response elements rule server service session subject alternative name tls extension transport layer security tls uniform resource locator user vulnerability details | 1 |
231,861 | 25,552,040,897 | IssuesEvent | 2022-11-30 01:08:59 | npenin/thewheel-dotnet | https://api.github.com/repos/npenin/thewheel-dotnet | reopened | newtonsoft.json.9.0.1.nupkg: 1 vulnerabilities (highest severity is: 7.5) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>newtonsoft.json.9.0.1.nupkg</b></p></summary>
<p>Json.NET is a popular high-performance JSON framework for .NET</p>
<p>Library home page: <a href="https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg">https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg</a></p>
<p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg,/ckages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/8d3ffe3535dffc84b035cd193db429b0d8b00401">8d3ffe3535dffc84b035cd193db429b0d8b00401</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (newtonsoft.json.9.0.1.nupkg version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [WS-2022-0161](https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | newtonsoft.json.9.0.1.nupkg | Direct | Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2022-0161</summary>
### Vulnerable Library - <b>newtonsoft.json.9.0.1.nupkg</b></p>
<p>Json.NET is a popular high-performance JSON framework for .NET</p>
<p>Library home page: <a href="https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg">https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg</a></p>
<p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg,/ckages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **newtonsoft.json.9.0.1.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/8d3ffe3535dffc84b035cd193db429b0d8b00401">8d3ffe3535dffc84b035cd193db429b0d8b00401</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Improper Handling of Exceptional Conditions in Newtonsoft.Json.
Newtonsoft.Json prior to version 13.0.1 is vulnerable to Insecure Defaults due to improper handling of StackOverFlow exception (SOE) whenever nested expressions are being processed. Exploiting this vulnerability results in Denial Of Service (DoS), and it is exploitable when an attacker sends 5 requests that cause SOE in time frame of 5 minutes. This vulnerability affects Internet Information Services (IIS) Applications.
<p>Publish Date: 2022-06-22
<p>URL: <a href=https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66>WS-2022-0161</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-06-22</p>
<p>Fix Resolution: Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | newtonsoft.json.9.0.1.nupkg: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>newtonsoft.json.9.0.1.nupkg</b></p></summary>
<p>Json.NET is a popular high-performance JSON framework for .NET</p>
<p>Library home page: <a href="https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg">https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg</a></p>
<p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg,/ckages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/8d3ffe3535dffc84b035cd193db429b0d8b00401">8d3ffe3535dffc84b035cd193db429b0d8b00401</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (newtonsoft.json.9.0.1.nupkg version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [WS-2022-0161](https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | newtonsoft.json.9.0.1.nupkg | Direct | Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2022-0161</summary>
### Vulnerable Library - <b>newtonsoft.json.9.0.1.nupkg</b></p>
<p>Json.NET is a popular high-performance JSON framework for .NET</p>
<p>Library home page: <a href="https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg">https://api.nuget.org/packages/newtonsoft.json.9.0.1.nupkg</a></p>
<p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg,/ckages/newtonsoft.json/9.0.1/newtonsoft.json.9.0.1.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **newtonsoft.json.9.0.1.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/8d3ffe3535dffc84b035cd193db429b0d8b00401">8d3ffe3535dffc84b035cd193db429b0d8b00401</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Improper Handling of Exceptional Conditions in Newtonsoft.Json.
Newtonsoft.Json prior to version 13.0.1 is vulnerable to Insecure Defaults due to improper handling of StackOverFlow exception (SOE) whenever nested expressions are being processed. Exploiting this vulnerability results in Denial Of Service (DoS), and it is exploitable when an attacker sends 5 requests that cause SOE in time frame of 5 minutes. This vulnerability affects Internet Information Services (IIS) Applications.
<p>Publish Date: 2022-06-22
<p>URL: <a href=https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66>WS-2022-0161</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-06-22</p>
<p>Fix Resolution: Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_main | newtonsoft json nupkg vulnerabilities highest severity is vulnerable library newtonsoft json nupkg json net is a popular high performance json framework for net library home page a href path to dependency file thewheel tests thewheel tests csproj path to vulnerable library home wss scanner nuget packages newtonsoft json newtonsoft json nupkg ckages newtonsoft json newtonsoft json nupkg found in head commit a href vulnerabilities cve severity cvss dependency type fixed in newtonsoft json nupkg version remediation available high newtonsoft json nupkg direct newtonsoft json microsoft extensions apidescription server details ws vulnerable library newtonsoft json nupkg json net is a popular high performance json framework for net library home page a href path to dependency file thewheel tests thewheel tests csproj path to vulnerable library home wss scanner nuget packages newtonsoft json newtonsoft json nupkg ckages newtonsoft json newtonsoft json nupkg dependency hierarchy x newtonsoft json nupkg vulnerable library found in head commit a href found in base branch main vulnerability details improper handling of exceptional conditions in newtonsoft json newtonsoft json prior to version is vulnerable to insecure defaults due to improper handling of stackoverflow exception soe whenever nested expressions are being processed exploiting this vulnerability results in denial of service dos and it is exploitable when an attacker sends requests that cause soe in time frame of minutes this vulnerability affects internet information services iis applications publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution newtonsoft json microsoft extensions apidescription server step up your open source security game with mend | 0 |
56,661 | 15,266,678,273 | IssuesEvent | 2021-02-22 09:07:11 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | DataScroller with SelectBooleanCheckbox: Initial checked value is overwritten for loaded chunks during first decode | defect | **Describe the defect**
If I create a datascroller with a list of items, that have a boolean value that is initially true and bound to a p:selectBooleanCheckbox within the datascroller, the checkbox is checked for entries that are initially dispayed (ok) but for additionally loaded elements due to scrolling the checkbox is not checked anymore.
**Reproducer**
[primefaces-test.zip](https://github.com/primefaces/primefaces/files/6003830/primefaces-test.zip)
**Environment:**
- PF Version: _10.0.0.RC1_
- JSF + version: Mojarra 2.2.8
- Affected browsers: ALL
**To Reproduce**
Steps to reproduce the behavior:
1. Go to test page in attached primefaces-test
2. Scroll down until additional entries are loaded
4. See error (additional entry index>5 have an unchecked checkbox)
**Expected behavior**
All checkboxes should be checked according to data in Model.
Possible Reason for that:
The problem is related to the fact that in **SelectBooleanCheckboxRenderer.decode** and submittedValue of null is interpreted as "false" and set on the underlying model. For lazy loaded chunk this seems to be executed during the ajax loadChunks call and since the new elements were just yet added their is not value in the requestparameter map for them.
Other Input Forms (e.g inputtext) for fine because their just skip the updatemodel phase if submited value == null.
**Example XHTML**
```html
<h:form id="frmTest">
<p:dataScroller var="_item" value="#{testView.model}" chunkSize="5" >
<div style="height:300px">
<p:inputText value="#{_item.text}" />
<p:selectBooleanCheckbox value="#{_item.check}"/> This checkbox should always be checked, but as soon as new chunk is loaded it is not.
</div>
</p:dataScroller>
</h:form>
```
**Example Bean**
```java
@Data
@Named
@ViewScoped
public class TestView implements Serializable {
private static final long serialVersionUID = 1L;
private ListDataModel<Item> model = new ListDataModel<>();
@Data
public static class Item implements Serializable {
private static final long serialVersionUID = 1L;
private String text;
private boolean check=true;
public Item(String text) {
super();
this.text = text;
}
}
@PostConstruct
public void init() {
List<Item> list = new ArrayList<>();
for (int i =0; i<50; i++) {
list.add(new Item("Text "+i));
}
model = new ListDataModel<>(list);
}
public ListDataModel<Item> getModel() {
return model;
}
}
```
| 1.0 | DataScroller with SelectBooleanCheckbox: Initial checked value is overwritten for loaded chunks during first decode - **Describe the defect**
If I create a datascroller with a list of items, that have a boolean value that is initially true and bound to a p:selectBooleanCheckbox within the datascroller, the checkbox is checked for entries that are initially dispayed (ok) but for additionally loaded elements due to scrolling the checkbox is not checked anymore.
**Reproducer**
[primefaces-test.zip](https://github.com/primefaces/primefaces/files/6003830/primefaces-test.zip)
**Environment:**
- PF Version: _10.0.0.RC1_
- JSF + version: Mojarra 2.2.8
- Affected browsers: ALL
**To Reproduce**
Steps to reproduce the behavior:
1. Go to test page in attached primefaces-test
2. Scroll down until additional entries are loaded
4. See error (additional entry index>5 have an unchecked checkbox)
**Expected behavior**
All checkboxes should be checked according to data in Model.
Possible Reason for that:
The problem is related to the fact that in **SelectBooleanCheckboxRenderer.decode** and submittedValue of null is interpreted as "false" and set on the underlying model. For lazy loaded chunk this seems to be executed during the ajax loadChunks call and since the new elements were just yet added their is not value in the requestparameter map for them.
Other Input Forms (e.g inputtext) for fine because their just skip the updatemodel phase if submited value == null.
**Example XHTML**
```html
<h:form id="frmTest">
<p:dataScroller var="_item" value="#{testView.model}" chunkSize="5" >
<div style="height:300px">
<p:inputText value="#{_item.text}" />
<p:selectBooleanCheckbox value="#{_item.check}"/> This checkbox should always be checked, but as soon as new chunk is loaded it is not.
</div>
</p:dataScroller>
</h:form>
```
**Example Bean**
```java
@Data
@Named
@ViewScoped
public class TestView implements Serializable {
private static final long serialVersionUID = 1L;
private ListDataModel<Item> model = new ListDataModel<>();
@Data
public static class Item implements Serializable {
private static final long serialVersionUID = 1L;
private String text;
private boolean check=true;
public Item(String text) {
super();
this.text = text;
}
}
@PostConstruct
public void init() {
List<Item> list = new ArrayList<>();
for (int i =0; i<50; i++) {
list.add(new Item("Text "+i));
}
model = new ListDataModel<>(list);
}
public ListDataModel<Item> getModel() {
return model;
}
}
```
| non_main | datascroller with selectbooleancheckbox initial checked value is overwritten for loaded chunks during first decode describe the defect if i create a datascroller with a list of items that have a boolean value that is initially true and bound to a p selectbooleancheckbox within the datascroller the checkbox is checked for entries that are initially dispayed ok but for additionally loaded elements due to scrolling the checkbox is not checked anymore reproducer environment pf version jsf version mojarra affected browsers all to reproduce steps to reproduce the behavior go to test page in attached primefaces test scroll down until additional entries are loaded see error additional entry index have an unchecked checkbox expected behavior all checkboxes should be checked according to data in model possible reason for that the problem is related to the fact that in selectbooleancheckboxrenderer decode and submittedvalue of null is interpreted as false and set on the underlying model for lazy loaded chunk this seems to be executed during the ajax loadchunks call and since the new elements were just yet added their is not value in the requestparameter map for them other input forms e g inputtext for fine because their just skip the updatemodel phase if submited value null example xhtml html this checkbox should always be checked but as soon as new chunk is loaded it is not example bean java data named viewscoped public class testview implements serializable private static final long serialversionuid private listdatamodel model new listdatamodel data public static class item implements serializable private static final long serialversionuid private string text private boolean check true public item string text super this text text postconstruct public void init list list new arraylist for int i i i list add new item text i model new listdatamodel list public listdatamodel getmodel return model | 0 |
117,120 | 11,945,718,945 | IssuesEvent | 2020-04-03 06:34:08 | heidicrq/ped | https://api.github.com/repos/heidicrq/ped | opened | Adding a task - contradiction between the UG and the error message | severity.Medium type.DocumentationBug | Can’t add task without deadline even though deadline is said to be optional in the UG. However error message suggests to the user that deadline is compulsory.

| 1.0 | Adding a task - contradiction between the UG and the error message - Can’t add task without deadline even though deadline is said to be optional in the UG. However error message suggests to the user that deadline is compulsory.

| non_main | adding a task contradiction between the ug and the error message can’t add task without deadline even though deadline is said to be optional in the ug however error message suggests to the user that deadline is compulsory | 0 |
679,139 | 23,222,209,359 | IssuesEvent | 2022-08-02 19:23:44 | dominicm00/ham | https://api.github.com/repos/dominicm00/ham | closed | Refactor memory and lifetime management | high priority large tracker refactor | Ham data classes derive from `Referenceable` as a replacement to smart pointers, but have a critical flaw; `Referenceable` objects cannot be const if their reference count is to be modified. This greatly limits the use of const throughout the codebase despite the fact that most Ham data is immutable after parsing, and causes difficult to resolve conflicts between const and non-const interfaces.
To resolve this, we should move to a more standard memory and lifetime management system, including:
- Using `shared_ptr` instead of `Referenceable`
- Removing manual memory management (e.g. the custom String classes)
- Throwing exceptions when constructors fail
PRs:
- [x] #66
- [ ] #67 | 1.0 | Refactor memory and lifetime management - Ham data classes derive from `Referenceable` as a replacement to smart pointers, but have a critical flaw; `Referenceable` objects cannot be const if their reference count is to be modified. This greatly limits the use of const throughout the codebase despite the fact that most Ham data is immutable after parsing, and causes difficult to resolve conflicts between const and non-const interfaces.
To resolve this, we should move to a more standard memory and lifetime management system, including:
- Using `shared_ptr` instead of `Referenceable`
- Removing manual memory management (e.g. the custom String classes)
- Throwing exceptions when constructors fail
PRs:
- [x] #66
- [ ] #67 | non_main | refactor memory and lifetime management ham data classes derive from referenceable as a replacement to smart pointers but have a critical flaw referenceable objects cannot be const if their reference count is to be modified this greatly limits the use of const throughout the codebase despite the fact that most ham data is immutable after parsing and causes difficult to resolve conflicts between const and non const interfaces to resolve this we should move to a more standard memory and lifetime management system including using shared ptr instead of referenceable removing manual memory management e g the custom string classes throwing exceptions when constructors fail prs | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.