Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
754,683 | 26,398,359,989 | IssuesEvent | 2023-01-12 21:46:49 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] "ERROR: Invalid argument: Invalid column number 2" after adding a new column | kind/bug area/ysql priority/medium status/awaiting-triage | Jira Link: [DB-3756](https://yugabyte.atlassian.net/browse/DB-3756)
### Description
Steps to reproduce:
1) Connect to one host, say 127.0.0.1:
```
yugabyte=# CREATE TABLE t (id int);
yugabyte=# INSERT INTO t VALUES (1), (2);
```
This populates the table caches for table "t" in TServer 1
End this connection
2) Connect to another host: ysqlsh -h 127.0.0.2
```
yugabyte=# ALTER TABLE t ADD COLUMN col1 TEXT;
```
3) Start a new connection to TServer1: ysqlsh -h 127.0.0.1
```
yugabyte=# INSERT INTO t VALUES (1);
ERROR: Invalid argument: Invalid column number 2
yugabyte=# INSERT INTO t VALUES (1);
ERROR: Invalid argument: Invalid column number 2
yugabyte=# INSERT INTO t VALUES (1);
ERROR: Invalid argument: Invalid column number 2
yugabyte=# INSERT INTO t VALUES (1);
```
This is because the new connection has the latest catalog version. Hence there is no cache refresh instigated by this connection. However TServer1 does not have the latest schema version in its table cache. The cached schema version does not have the new column, and thus the above error is thrown.
4) Start another connection in some other node (127.0.0.2 or 127.0.0.3). This succeeds:
```
yugabyte=# INSERT INTO t VALUES (1);
INSERT 0 1
``` | 1.0 | [YSQL] "ERROR: Invalid argument: Invalid column number 2" after adding a new column - Jira Link: [DB-3756](https://yugabyte.atlassian.net/browse/DB-3756)
### Description
Steps to reproduce:
1) Connect to one host, say 127.0.0.1:
```
yugabyte=# CREATE TABLE t (id int);
yugabyte=# INSERT INTO t VALUES (1), (2);
```
This populates the table caches for table "t" in TServer 1
End this connection
2) Connect to another host: ysqlsh -h 127.0.0.2
```
yugabyte=# ALTER TABLE t ADD COLUMN col1 TEXT;
```
3) Start a new connection to TServer1: ysqlsh -h 127.0.0.1
```
yugabyte=# INSERT INTO t VALUES (1);
ERROR: Invalid argument: Invalid column number 2
yugabyte=# INSERT INTO t VALUES (1);
ERROR: Invalid argument: Invalid column number 2
yugabyte=# INSERT INTO t VALUES (1);
ERROR: Invalid argument: Invalid column number 2
yugabyte=# INSERT INTO t VALUES (1);
```
This is because the new connection has the latest catalog version. Hence there is no cache refresh instigated by this connection. However TServer1 does not have the latest schema version in its table cache. The cached schema version does not have the new column, and thus the above error is thrown.
4) Start another connection in some other node (127.0.0.2 or 127.0.0.3). This succeeds:
```
yugabyte=# INSERT INTO t VALUES (1);
INSERT 0 1
``` | priority | error invalid argument invalid column number after adding a new column jira link description steps to reproduce connect to one host say yugabyte create table t id int yugabyte insert into t values this populates the table caches for table t in tserver end this connection connect to another host ysqlsh h yugabyte alter table t add column text start a new connection to ysqlsh h yugabyte insert into t values error invalid argument invalid column number yugabyte insert into t values error invalid argument invalid column number yugabyte insert into t values error invalid argument invalid column number yugabyte insert into t values this is because the new connection has the latest catalog version hence there is no cache refresh instigated by this connection however does not have the latest schema version in its table cache the cached schema version does not have the new column and thus the above error is thrown start another connection in some other node or this succeeds yugabyte insert into t values insert | 1 |
567,249 | 16,851,639,274 | IssuesEvent | 2021-06-20 16:25:21 | GastonGit/Hot-Twitch-Clips | https://api.github.com/repos/GastonGit/Hot-Twitch-Clips | closed | TMI.JS :: DISCONNECT :: | bug high priority | Problem(https://github.com/GastonGit/Hot-Twitch-Clips/issues/126) still exists after fix with disconnects happening with no reason attached to disconnect. | 1.0 | TMI.JS :: DISCONNECT :: - Problem(https://github.com/GastonGit/Hot-Twitch-Clips/issues/126) still exists after fix with disconnects happening with no reason attached to disconnect. | priority | tmi js disconnect problem still exists after fix with disconnects happening with no reason attached to disconnect | 1 |
114,533 | 9,741,910,592 | IssuesEvent | 2019-06-02 12:59:31 | aiidateam/aiida_core | https://api.github.com/repos/aiidateam/aiida_core | closed | Random failing tests `verdi quicksetup` on SqlAlchemy python 3 | aiida-core 1.x priority/quality-of-life topic/testing topic/verdi type/bug | I have noticed this test failing multiple times the recent days, always on python 3 and SqlAlchemy. Not sure where it is coming from:
```
======================================================================
FAIL: test_quicksetup_from_config_file (aiida.backends.tests.cmdline.commands.test_setup.TestVerdiSetup)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/travis/build/aiidateam/aiida_core/aiida/backends/tests/utils/configuration.py", line 105, in decorated_function
function(*args, **kwargs)
File "/home/travis/build/aiidateam/aiida_core/aiida/backends/tests/cmdline/commands/test_setup.py", line 90, in test_quicksetup_from_config_file
self.assertClickResultNoException(result)
File "/home/travis/build/aiidateam/aiida_core/aiida/backends/testbase.py", line 187, in assertClickResultNoException
self.assertIsNone(cli_result.exception, ''.join(traceback.format_exception(*cli_result.exc_info)))
AssertionError: SystemExit(<ExitCode.CRITICAL: 1>,) is not None : Traceback (most recent call last):
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 213, in ensure_connection
self.connect()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 189, in connect
self.connection = self.get_new_connection(conn_params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/postgresql/base.py", line 176, in get_new_connection
connection = Database.connect(**conn_params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/psycopg2/__init__.py", line 126, in connect
conn = _connect(dsn, connection_factory=connection_factory, **kwasync)
psycopg2.OperationalError: FATAL: role "user" does not exist
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/travis/build/aiidateam/aiida_core/aiida/cmdline/commands/cmd_setup.py", line 72, in setup
backend.migrate()
File "/home/travis/build/aiidateam/aiida_core/aiida/orm/implementation/django/backend.py", line 53, in migrate
migrate_database()
File "/home/travis/build/aiidateam/aiida_core/aiida/backends/djsite/utils.py", line 54, in migrate_database
call_command('migrate')
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/core/management/__init__.py", line 131, in call_command
return command.execute(*args, **defaults)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/core/management/base.py", line 330, in execute
output = self.handle(*args, **options)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/core/management/commands/migrate.py", line 83, in handle
executor = MigrationExecutor(connection, self.migration_progress_callback)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/executor.py", line 20, in __init__
self.loader = MigrationLoader(self.connection)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/loader.py", line 52, in __init__
self.build_graph()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/loader.py", line 210, in build_graph
self.applied_migrations = recorder.applied_migrations()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/recorder.py", line 65, in applied_migrations
self.ensure_schema()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/recorder.py", line 52, in ensure_schema
if self.Migration._meta.db_table in self.connection.introspection.table_names(self.connection.cursor()):
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 254, in cursor
return self._cursor()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 229, in _cursor
self.ensure_connection()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 213, in ensure_connection
self.connect()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/utils.py", line 94, in __exit__
six.reraise(dj_exc_type, dj_exc_value, traceback)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/utils/six.py", line 685, in reraise
raise value.with_traceback(tb)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 213, in ensure_connection
self.connect()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 189, in connect
self.connection = self.get_new_connection(conn_params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/postgresql/base.py", line 176, in get_new_connection
connection = Database.connect(**conn_params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/psycopg2/__init__.py", line 126, in connect
conn = _connect(dsn, connection_factory=connection_factory, **kwasync)
django.db.utils.OperationalError: FATAL: role "user" does not exist
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/testing.py", line 326, in invoke
cli.main(args=args or (), prog_name=prog_name, **extra)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/core.py", line 717, in main
rv = self.invoke(ctx)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/core.py", line 956, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/core.py", line 555, in invoke
return callback(*args, **kwargs)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/decorators.py", line 17, in new_func
return f(get_current_context(), *args, **kwargs)
File "/home/travis/build/aiidateam/aiida_core/aiida/cmdline/commands/cmd_setup.py", line 173, in quicksetup
ctx.invoke(setup, **setup_parameters)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/core.py", line 555, in invoke
return callback(*args, **kwargs)
File "/home/travis/build/aiidateam/aiida_core/aiida/cmdline/commands/cmd_setup.py", line 75, in setup
'database migration failed, probably because connection details are incorrect:\n{}'.format(exception))
File "/home/travis/build/aiidateam/aiida_core/aiida/cmdline/utils/echo.py", line 114, in echo_critical
sys.exit(ExitCode.CRITICAL)
SystemExit: ExitCode.CRITICAL
``` | 1.0 | Random failing tests `verdi quicksetup` on SqlAlchemy python 3 - I have noticed this test failing multiple times the recent days, always on python 3 and SqlAlchemy. Not sure where it is coming from:
```
======================================================================
FAIL: test_quicksetup_from_config_file (aiida.backends.tests.cmdline.commands.test_setup.TestVerdiSetup)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/travis/build/aiidateam/aiida_core/aiida/backends/tests/utils/configuration.py", line 105, in decorated_function
function(*args, **kwargs)
File "/home/travis/build/aiidateam/aiida_core/aiida/backends/tests/cmdline/commands/test_setup.py", line 90, in test_quicksetup_from_config_file
self.assertClickResultNoException(result)
File "/home/travis/build/aiidateam/aiida_core/aiida/backends/testbase.py", line 187, in assertClickResultNoException
self.assertIsNone(cli_result.exception, ''.join(traceback.format_exception(*cli_result.exc_info)))
AssertionError: SystemExit(<ExitCode.CRITICAL: 1>,) is not None : Traceback (most recent call last):
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 213, in ensure_connection
self.connect()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 189, in connect
self.connection = self.get_new_connection(conn_params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/postgresql/base.py", line 176, in get_new_connection
connection = Database.connect(**conn_params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/psycopg2/__init__.py", line 126, in connect
conn = _connect(dsn, connection_factory=connection_factory, **kwasync)
psycopg2.OperationalError: FATAL: role "user" does not exist
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/travis/build/aiidateam/aiida_core/aiida/cmdline/commands/cmd_setup.py", line 72, in setup
backend.migrate()
File "/home/travis/build/aiidateam/aiida_core/aiida/orm/implementation/django/backend.py", line 53, in migrate
migrate_database()
File "/home/travis/build/aiidateam/aiida_core/aiida/backends/djsite/utils.py", line 54, in migrate_database
call_command('migrate')
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/core/management/__init__.py", line 131, in call_command
return command.execute(*args, **defaults)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/core/management/base.py", line 330, in execute
output = self.handle(*args, **options)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/core/management/commands/migrate.py", line 83, in handle
executor = MigrationExecutor(connection, self.migration_progress_callback)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/executor.py", line 20, in __init__
self.loader = MigrationLoader(self.connection)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/loader.py", line 52, in __init__
self.build_graph()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/loader.py", line 210, in build_graph
self.applied_migrations = recorder.applied_migrations()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/recorder.py", line 65, in applied_migrations
self.ensure_schema()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/migrations/recorder.py", line 52, in ensure_schema
if self.Migration._meta.db_table in self.connection.introspection.table_names(self.connection.cursor()):
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 254, in cursor
return self._cursor()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 229, in _cursor
self.ensure_connection()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 213, in ensure_connection
self.connect()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/utils.py", line 94, in __exit__
six.reraise(dj_exc_type, dj_exc_value, traceback)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/utils/six.py", line 685, in reraise
raise value.with_traceback(tb)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 213, in ensure_connection
self.connect()
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/base/base.py", line 189, in connect
self.connection = self.get_new_connection(conn_params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/django/db/backends/postgresql/base.py", line 176, in get_new_connection
connection = Database.connect(**conn_params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/psycopg2/__init__.py", line 126, in connect
conn = _connect(dsn, connection_factory=connection_factory, **kwasync)
django.db.utils.OperationalError: FATAL: role "user" does not exist
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/testing.py", line 326, in invoke
cli.main(args=args or (), prog_name=prog_name, **extra)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/core.py", line 717, in main
rv = self.invoke(ctx)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/core.py", line 956, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/core.py", line 555, in invoke
return callback(*args, **kwargs)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/decorators.py", line 17, in new_func
return f(get_current_context(), *args, **kwargs)
File "/home/travis/build/aiidateam/aiida_core/aiida/cmdline/commands/cmd_setup.py", line 173, in quicksetup
ctx.invoke(setup, **setup_parameters)
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/click/core.py", line 555, in invoke
return callback(*args, **kwargs)
File "/home/travis/build/aiidateam/aiida_core/aiida/cmdline/commands/cmd_setup.py", line 75, in setup
'database migration failed, probably because connection details are incorrect:\n{}'.format(exception))
File "/home/travis/build/aiidateam/aiida_core/aiida/cmdline/utils/echo.py", line 114, in echo_critical
sys.exit(ExitCode.CRITICAL)
SystemExit: ExitCode.CRITICAL
``` | non_priority | random failing tests verdi quicksetup on sqlalchemy python i have noticed this test failing multiple times the recent days always on python and sqlalchemy not sure where it is coming from fail test quicksetup from config file aiida backends tests cmdline commands test setup testverdisetup traceback most recent call last file home travis build aiidateam aiida core aiida backends tests utils configuration py line in decorated function function args kwargs file home travis build aiidateam aiida core aiida backends tests cmdline commands test setup py line in test quicksetup from config file self assertclickresultnoexception result file home travis build aiidateam aiida core aiida backends testbase py line in assertclickresultnoexception self assertisnone cli result exception join traceback format exception cli result exc info assertionerror systemexit is not none traceback most recent call last file home travis virtualenv lib site packages django db backends base base py line in ensure connection self connect file home travis virtualenv lib site packages django db backends base base py line in connect self connection self get new connection conn params file home travis virtualenv lib site packages django db backends postgresql base py line in get new connection connection database connect conn params file home travis virtualenv lib site packages init py line in connect conn connect dsn connection factory connection factory kwasync operationalerror fatal role user does not exist the above exception was the direct cause of the following exception traceback most recent call last file home travis build aiidateam aiida core aiida cmdline commands cmd setup py line in setup backend migrate file home travis build aiidateam aiida core aiida orm implementation django backend py line in migrate migrate database file home travis build aiidateam aiida core aiida backends djsite utils py line in migrate database call command migrate file home travis virtualenv lib site packages django core management init py line in call command return command execute args defaults file home travis virtualenv lib site packages django core management base py line in execute output self handle args options file home travis virtualenv lib site packages django core management commands migrate py line in handle executor migrationexecutor connection self migration progress callback file home travis virtualenv lib site packages django db migrations executor py line in init self loader migrationloader self connection file home travis virtualenv lib site packages django db migrations loader py line in init self build graph file home travis virtualenv lib site packages django db migrations loader py line in build graph self applied migrations recorder applied migrations file home travis virtualenv lib site packages django db migrations recorder py line in applied migrations self ensure schema file home travis virtualenv lib site packages django db migrations recorder py line in ensure schema if self migration meta db table in self connection introspection table names self connection cursor file home travis virtualenv lib site packages django db backends base base py line in cursor return self cursor file home travis virtualenv lib site packages django db backends base base py line in cursor self ensure connection file home travis virtualenv lib site packages django db backends base base py line in ensure connection self connect file home travis virtualenv lib site packages django db utils py line in exit six reraise dj exc type dj exc value traceback file home travis virtualenv lib site packages django utils six py line in reraise raise value with traceback tb file home travis virtualenv lib site packages django db backends base base py line in ensure connection self connect file home travis virtualenv lib site packages django db backends base base py line in connect self connection self get new connection conn params file home travis virtualenv lib site packages django db backends postgresql base py line in get new connection connection database connect conn params file home travis virtualenv lib site packages init py line in connect conn connect dsn connection factory connection factory kwasync django db utils operationalerror fatal role user does not exist during handling of the above exception another exception occurred traceback most recent call last file home travis virtualenv lib site packages click testing py line in invoke cli main args args or prog name prog name extra file home travis virtualenv lib site packages click core py line in main rv self invoke ctx file home travis virtualenv lib site packages click core py line in invoke return ctx invoke self callback ctx params file home travis virtualenv lib site packages click core py line in invoke return callback args kwargs file home travis virtualenv lib site packages click decorators py line in new func return f get current context args kwargs file home travis build aiidateam aiida core aiida cmdline commands cmd setup py line in quicksetup ctx invoke setup setup parameters file home travis virtualenv lib site packages click core py line in invoke return callback args kwargs file home travis build aiidateam aiida core aiida cmdline commands cmd setup py line in setup database migration failed probably because connection details are incorrect n format exception file home travis build aiidateam aiida core aiida cmdline utils echo py line in echo critical sys exit exitcode critical systemexit exitcode critical | 0 |
76,619 | 15,496,152,308 | IssuesEvent | 2021-03-11 02:09:21 | jinuem/React-Type-Script-Starter | https://api.github.com/repos/jinuem/React-Type-Script-Starter | opened | CVE-2020-7608 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-2.4.1.tgz</b>, <b>yargs-parser-4.2.1.tgz</b>, <b>yargs-parser-5.0.0.tgz</b>, <b>yargs-parser-7.0.0.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-2.4.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p>
<p>Path to dependency file: /React-Type-Script-Starter/package.json</p>
<p>Path to vulnerable library: React-Type-Script-Starter/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-ts-2.5.0.tgz (Root Library)
- cli-highlight-1.1.4.tgz
- yargs-4.8.1.tgz
- :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-4.2.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-4.2.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-4.2.1.tgz</a></p>
<p>Path to dependency file: /React-Type-Script-Starter/package.json</p>
<p>Path to vulnerable library: React-Type-Script-Starter/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-ts-2.5.0.tgz (Root Library)
- webpack-2.6.0.tgz
- yargs-6.6.0.tgz
- :x: **yargs-parser-4.2.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-5.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz</a></p>
<p>Path to dependency file: /React-Type-Script-Starter/package.json</p>
<p>Path to vulnerable library: React-Type-Script-Starter/node_modules/jest/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-ts-2.5.0.tgz (Root Library)
- jest-20.0.3.tgz
- jest-cli-20.0.4.tgz
- jest-runtime-20.0.4.tgz
- yargs-7.1.0.tgz
- :x: **yargs-parser-5.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p>
<p>Path to dependency file: /React-Type-Script-Starter/package.json</p>
<p>Path to vulnerable library: React-Type-Script-Starter/node_modules/ts-jest/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-ts-2.5.0.tgz (Root Library)
- ts-jest-20.0.7.tgz
- yargs-8.0.2.tgz
- :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7608 (Medium) detected in multiple libraries - ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-2.4.1.tgz</b>, <b>yargs-parser-4.2.1.tgz</b>, <b>yargs-parser-5.0.0.tgz</b>, <b>yargs-parser-7.0.0.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-2.4.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p>
<p>Path to dependency file: /React-Type-Script-Starter/package.json</p>
<p>Path to vulnerable library: React-Type-Script-Starter/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-ts-2.5.0.tgz (Root Library)
- cli-highlight-1.1.4.tgz
- yargs-4.8.1.tgz
- :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-4.2.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-4.2.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-4.2.1.tgz</a></p>
<p>Path to dependency file: /React-Type-Script-Starter/package.json</p>
<p>Path to vulnerable library: React-Type-Script-Starter/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-ts-2.5.0.tgz (Root Library)
- webpack-2.6.0.tgz
- yargs-6.6.0.tgz
- :x: **yargs-parser-4.2.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-5.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz</a></p>
<p>Path to dependency file: /React-Type-Script-Starter/package.json</p>
<p>Path to vulnerable library: React-Type-Script-Starter/node_modules/jest/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-ts-2.5.0.tgz (Root Library)
- jest-20.0.3.tgz
- jest-cli-20.0.4.tgz
- jest-runtime-20.0.4.tgz
- yargs-7.1.0.tgz
- :x: **yargs-parser-5.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p>
<p>Path to dependency file: /React-Type-Script-Starter/package.json</p>
<p>Path to vulnerable library: React-Type-Script-Starter/node_modules/ts-jest/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-ts-2.5.0.tgz (Root Library)
- ts-jest-20.0.7.tgz
- yargs-8.0.2.tgz
- :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file react type script starter package json path to vulnerable library react type script starter node modules yargs parser package json dependency hierarchy react scripts ts tgz root library cli highlight tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file react type script starter package json path to vulnerable library react type script starter node modules webpack dev server node modules yargs parser package json dependency hierarchy react scripts ts tgz root library webpack tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file react type script starter package json path to vulnerable library react type script starter node modules jest node modules yargs parser package json dependency hierarchy react scripts ts tgz root library jest tgz jest cli tgz jest runtime tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file react type script starter package json path to vulnerable library react type script starter node modules ts jest node modules yargs parser package json dependency hierarchy react scripts ts tgz root library ts jest tgz yargs tgz x yargs parser tgz vulnerable library vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
249,702 | 7,964,659,940 | IssuesEvent | 2018-07-13 22:38:25 | huridocs/uwazi | https://api.github.com/repos/huridocs/uwazi | closed | Display a tool-tip asking user to change the default admin password | Priority: Low | Users aren't changing the default password for admin. We could add a tool-tip for instances with default password so they are inclined to do so. | 1.0 | Display a tool-tip asking user to change the default admin password - Users aren't changing the default password for admin. We could add a tool-tip for instances with default password so they are inclined to do so. | priority | display a tool tip asking user to change the default admin password users aren t changing the default password for admin we could add a tool tip for instances with default password so they are inclined to do so | 1 |
171,398 | 27,112,795,828 | IssuesEvent | 2023-02-15 16:24:18 | hackforla/expunge-assist | https://api.github.com/repos/hackforla/expunge-assist | closed | [Re-organize Figma: Archive] Archive FAQ page | role: design priority: medium role: UX content writing size: 1pt | ### Overview
Extension of #807
### Action Items
- [x] Read through comments and determine if they can be marked as "resolved"
- [x] For comments that cannot be marked as resolved, add a box on the side of the wireframe where the comment is as and copy and paste the comment/thread of comments to the side.
- [x] Draw an arrow to the area the comment was referencing
- [x] Once this is done to all the comments on the page, copy all the content on the "FAQ" and pasting it under "IUT2 (2022)" page under the "Archive" section on Figma
- [x] Select the FAQ page content and make it into a "section"
- [x] Name it "FAQ page"
- [x] Let team know when complete
### Resources/Instructions
Reach out to @anitadesigns or @SamHyler for questions
| 1.0 | [Re-organize Figma: Archive] Archive FAQ page - ### Overview
Extension of #807
### Action Items
- [x] Read through comments and determine if they can be marked as "resolved"
- [x] For comments that cannot be marked as resolved, add a box on the side of the wireframe where the comment is as and copy and paste the comment/thread of comments to the side.
- [x] Draw an arrow to the area the comment was referencing
- [x] Once this is done to all the comments on the page, copy all the content on the "FAQ" and pasting it under "IUT2 (2022)" page under the "Archive" section on Figma
- [x] Select the FAQ page content and make it into a "section"
- [x] Name it "FAQ page"
- [x] Let team know when complete
### Resources/Instructions
Reach out to @anitadesigns or @SamHyler for questions
| non_priority | archive faq page overview extension of action items read through comments and determine if they can be marked as resolved for comments that cannot be marked as resolved add a box on the side of the wireframe where the comment is as and copy and paste the comment thread of comments to the side draw an arrow to the area the comment was referencing once this is done to all the comments on the page copy all the content on the faq and pasting it under page under the archive section on figma select the faq page content and make it into a section name it faq page let team know when complete resources instructions reach out to anitadesigns or samhyler for questions | 0 |
72,050 | 3,371,408,300 | IssuesEvent | 2015-11-23 19:00:13 | aaroneiche/do-want | https://api.github.com/repos/aaroneiche/do-want | closed | Implement OAuth | enhancement imported Priority-Low | _Original author: aaron.ei...@gmail.com (September 06, 2012 16:41:34)_
Do-Want should support OAuth authentication through popular services like google, facebook, and twitter.
_Original issue: http://code.google.com/p/do-want/issues/detail?id=16_ | 1.0 | Implement OAuth - _Original author: aaron.ei...@gmail.com (September 06, 2012 16:41:34)_
Do-Want should support OAuth authentication through popular services like google, facebook, and twitter.
_Original issue: http://code.google.com/p/do-want/issues/detail?id=16_ | priority | implement oauth original author aaron ei gmail com september do want should support oauth authentication through popular services like google facebook and twitter original issue | 1 |
35,801 | 2,793,080,605 | IssuesEvent | 2015-05-11 08:31:47 | handsontable/handsontable | https://api.github.com/repos/handsontable/handsontable | closed | Creating ContextMenu items on demand pr cell | Feature Guess: a day or more Priority: normal | Is there any way to dynamically set the items of a context menu based on the cell that is right clicked? I need to show a different set of context menu items for each column, and ideally for each cell.
In my case, I need access to the underlying data of the current row and the column of the clicked cell to determine how to fill the context menu. Disabling any items that are invalid for the current context won't be a good solution for me, because it would mean that the context menu would contain a lot of disabled items.
If filling the context menu based on the clicked cell is not possible, would it be possible to fill it based on which column the clicked cell belongs to?
Thanks! | 1.0 | Creating ContextMenu items on demand pr cell - Is there any way to dynamically set the items of a context menu based on the cell that is right clicked? I need to show a different set of context menu items for each column, and ideally for each cell.
In my case, I need access to the underlying data of the current row and the column of the clicked cell to determine how to fill the context menu. Disabling any items that are invalid for the current context won't be a good solution for me, because it would mean that the context menu would contain a lot of disabled items.
If filling the context menu based on the clicked cell is not possible, would it be possible to fill it based on which column the clicked cell belongs to?
Thanks! | priority | creating contextmenu items on demand pr cell is there any way to dynamically set the items of a context menu based on the cell that is right clicked i need to show a different set of context menu items for each column and ideally for each cell in my case i need access to the underlying data of the current row and the column of the clicked cell to determine how to fill the context menu disabling any items that are invalid for the current context won t be a good solution for me because it would mean that the context menu would contain a lot of disabled items if filling the context menu based on the clicked cell is not possible would it be possible to fill it based on which column the clicked cell belongs to thanks | 1 |
202,608 | 23,077,569,693 | IssuesEvent | 2022-07-26 02:12:45 | YJSoft/macos-plist | https://api.github.com/repos/YJSoft/macos-plist | opened | CVE-2021-35065 (High) detected in glob-parent-5.1.2.tgz | security vulnerability | ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- mocha-8.3.2.tgz (Root Library)
- chokidar-3.5.1.tgz
- :x: **glob-parent-5.1.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (mocha): 9.1.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-35065 (High) detected in glob-parent-5.1.2.tgz - ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- mocha-8.3.2.tgz (Root Library)
- chokidar-3.5.1.tgz
- :x: **glob-parent-5.1.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (mocha): 9.1.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in glob parent tgz cve high severity vulnerability vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy mocha tgz root library chokidar tgz x glob parent tgz vulnerable library found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent direct dependency fix resolution mocha step up your open source security game with mend | 0 |
239,396 | 19,862,317,367 | IssuesEvent | 2022-01-22 02:39:03 | Exiled-Team/EXILED | https://api.github.com/repos/Exiled-Team/EXILED | closed | [BUG] Scp-049 can revive SCP (like SCP-049-2) | requires-testing | **Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. Have an Dead SCP
2. Respawn it
3. And you have the problem
**Expected behavior**
SCP-049 can revive SCP
**Server logs**
Please include a pastebin of your localadmin log file (or both MA_log and SCP_log files if you use MultiAdmin) from the time in which the bug occured
it's don't have error
**EXILED Version ("latest" is not a version):**
5.0.0
**Results of `show plugins` command in console:**
**Additional context**
Was here before of the new update of SL | 1.0 | [BUG] Scp-049 can revive SCP (like SCP-049-2) - **Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. Have an Dead SCP
2. Respawn it
3. And you have the problem
**Expected behavior**
SCP-049 can revive SCP
**Server logs**
Please include a pastebin of your localadmin log file (or both MA_log and SCP_log files if you use MultiAdmin) from the time in which the bug occured
it's don't have error
**EXILED Version ("latest" is not a version):**
5.0.0
**Results of `show plugins` command in console:**
**Additional context**
Was here before of the new update of SL | non_priority | scp can revive scp like scp describe the bug a clear and concise description of what the bug is to reproduce steps to reproduce the behavior have an dead scp respawn it and you have the problem expected behavior scp can revive scp server logs please include a pastebin of your localadmin log file or both ma log and scp log files if you use multiadmin from the time in which the bug occured it s don t have error exiled version latest is not a version results of show plugins command in console additional context was here before of the new update of sl | 0 |
252,198 | 27,233,114,173 | IssuesEvent | 2023-02-21 14:34:57 | billmcchesney1/linkerd2 | https://api.github.com/repos/billmcchesney1/linkerd2 | opened | CVE-2021-23440 (High) detected in set-value-2.0.1.tgz | security vulnerability | ## CVE-2021-23440 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-value-2.0.1.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz</a></p>
<p>Path to dependency file: /web/app/package.json</p>
<p>Path to vulnerable library: /web/app/node_modules/set-value/package.json</p>
<p>
Dependency Hierarchy:
- webpack-4.44.1.tgz (Root Library)
- micromatch-3.1.10.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- :x: **set-value-2.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/linkerd2/commit/2ba890cd094fe9f32531e2f3491a16207038088f">2ba890cd094fe9f32531e2f3491a16207038088f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays.
Mend Note: After conducting further research, Mend has determined that all versions of set-value up to version 4.0.0 are vulnerable to CVE-2021-23440.
<p>Publish Date: 2021-09-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23440>CVE-2021-23440</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-09-12</p>
<p>Fix Resolution: set-value - 4.0.1
</p>
</p>
</details>
<p></p>
| True | CVE-2021-23440 (High) detected in set-value-2.0.1.tgz - ## CVE-2021-23440 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-value-2.0.1.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz</a></p>
<p>Path to dependency file: /web/app/package.json</p>
<p>Path to vulnerable library: /web/app/node_modules/set-value/package.json</p>
<p>
Dependency Hierarchy:
- webpack-4.44.1.tgz (Root Library)
- micromatch-3.1.10.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- :x: **set-value-2.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/linkerd2/commit/2ba890cd094fe9f32531e2f3491a16207038088f">2ba890cd094fe9f32531e2f3491a16207038088f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays.
Mend Note: After conducting further research, Mend has determined that all versions of set-value up to version 4.0.0 are vulnerable to CVE-2021-23440.
<p>Publish Date: 2021-09-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23440>CVE-2021-23440</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-09-12</p>
<p>Fix Resolution: set-value - 4.0.1
</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in set value tgz cve high severity vulnerability vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file web app package json path to vulnerable library web app node modules set value package json dependency hierarchy webpack tgz root library micromatch tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library found in head commit a href found in base branch main vulnerability details this affects the package set value before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays mend note after conducting further research mend has determined that all versions of set value up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution set value | 0 |
124,354 | 4,912,468,036 | IssuesEvent | 2016-11-23 09:11:41 | bespokeinteractive/mchapp | https://api.github.com/repos/bespokeinteractive/mchapp | opened | ANC clinic: There's an error when trying to update tetanus vaccine given to an ANC patient | bug High Priority | ANC clinic: There's an error when trying to update tetanus vaccine given to an ANC patient
Task
-----------
-Ensure user is able to update Tetanus administered to a patient

| 1.0 | ANC clinic: There's an error when trying to update tetanus vaccine given to an ANC patient - ANC clinic: There's an error when trying to update tetanus vaccine given to an ANC patient
Task
-----------
-Ensure user is able to update Tetanus administered to a patient

| priority | anc clinic there s an error when trying to update tetanus vaccine given to an anc patient anc clinic there s an error when trying to update tetanus vaccine given to an anc patient task ensure user is able to update tetanus administered to a patient | 1 |
819,866 | 30,753,558,464 | IssuesEvent | 2023-07-28 22:06:04 | Memmy-App/memmy | https://api.github.com/repos/Memmy-App/memmy | closed | [Bug] : Removed “sort by new [posts]” | bug medium priority visual | ### Check for open issues
- [X] I have verified that another issue for this is not open, or it has been closed and has not been fixed.
### Minimal reproducible example
Latest update removed “sort by new [posts]” and replaced it with “sort by new comments”
### Expected Behavior
I’d still like the option to sort by new posts.
### Version
0.5.0.10
### App Store or TestFlight?
TestFlight
### iOS Version
16.x
### Device
iPhone 14 Pro | 1.0 | [Bug] : Removed “sort by new [posts]” - ### Check for open issues
- [X] I have verified that another issue for this is not open, or it has been closed and has not been fixed.
### Minimal reproducible example
Latest update removed “sort by new [posts]” and replaced it with “sort by new comments”
### Expected Behavior
I’d still like the option to sort by new posts.
### Version
0.5.0.10
### App Store or TestFlight?
TestFlight
### iOS Version
16.x
### Device
iPhone 14 Pro | priority | removed “sort by new ” check for open issues i have verified that another issue for this is not open or it has been closed and has not been fixed minimal reproducible example latest update removed “sort by new ” and replaced it with “sort by new comments” expected behavior i’d still like the option to sort by new posts version app store or testflight testflight ios version x device iphone pro | 1 |
69,015 | 22,056,615,320 | IssuesEvent | 2022-05-30 13:26:15 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | opened | ClientMessage frame read for a message may be broken in case of large messages | Type: Defect All Languages Should Check to-jira | Please see [this](https://github.com/hazelcast/hazelcast/blob/v5.1.1/hazelcast/src/main/java/com/hazelcast/client/impl/protocol/ClientMessageReader.java#L68) line of code.
The ClientMessageReader is basically copying data from one buffer to another buffer.
The source buffer (src in the code) is bytes received directly from the client socket connection.
The destination buffer is an application byte[] , which is wrapped inside an object called ClientMessage.
(Its actually slight more complex ... ClientMessage has multiple buffers called Frames, and within ClientMessage there is a linked list of Frames, each Frame has its own byte[])
There is an important variable called readOffset; this denotes the offset in the current destination buffer where we should put the next set of bytes.
-1 is the initial state meaning no data has been copied yet. The reason that readOffset is a member variable is because the class is stateful and the main readFrom method in the class is called in a loop.
This is important to understand , so here is an example ... we might invoke the readFrom method with 25,000 bytes in the src buffer. Within the readFrom method, we might derive the overall expected length of the frame is 75,000 bytes. The code will still copy the available 25,000 bytes from the src into the destination buffer, and it will set the readOffset to 25,000. The method will then complete and return false (false == 'incomplete').
Then on the next iteration/call to readFrom , we might have 30,000 new bytes in the src. Again it will copy those bytes into the destination (starting at the readOffset 25,000), then it will bump the readOffset pointer forward to 55,000, and so on and so on ... until it receives the full frame of 75,000 bytes.
Once a frame is full (i.e. we have copied the number of bytes specified in the frame header) , the code will start populating a new frame.
Once a frame is full and it is flagged as IS_FINAL_FLAG = true .. that whole ClientMessage is complete, and the readFrom will return true .
Setup for error:
We have large messages being sent from client to server therefore they tend to span fairly large number of tcp packets.
We also have data being sent from cloud to onprem ... so sometimes those tcp packets can be a few micros apart.
This leads to plenty of occurrences where readFrom is invoked when the src buffer does not have the entire dataset (frame) available. And thus our larger messages are often built up iteratively as described above. That generally works well ... apart from 1 case ..
Sometimes when readFrom is called, the src buffer has the data needed to complete the destination message except the last few bytes of the final frame (5 bytes say).
So, we process as described above what we have, and create an almost complete ClientMessage in memory in ClientMessageReader. Then micros later the remaining data becomes available in the src buffer, and we can append that by another call to readFrom. So that is invoked .. but it hits this check ...
int remaining = src.remaining();
if (remaining < SIZE_OF_FRAME_LENGTH_AND_FLAGS) {
// we don't have even the frame length and flags ready
return false;
}
which won't read the src buffer if its size is less that 6 bytes.
So, then we can't complete the frame / message because the residual part is too small to pass through the above check.
The connection then stays in that state, until at some point later, something else is received in the src buffer (typically a hearbeat). At that point we get through this guard because the buffer exceeds the minimum size, and we can process the 5 bytes that have been sat there waiting (and the heartbeat).
I think the above check should only occur when we are at the beginning of frames i.e. just moved down to iinside the
if (readOffset == -1) {
block. | 1.0 | ClientMessage frame read for a message may be broken in case of large messages - Please see [this](https://github.com/hazelcast/hazelcast/blob/v5.1.1/hazelcast/src/main/java/com/hazelcast/client/impl/protocol/ClientMessageReader.java#L68) line of code.
The ClientMessageReader is basically copying data from one buffer to another buffer.
The source buffer (src in the code) is bytes received directly from the client socket connection.
The destination buffer is an application byte[] , which is wrapped inside an object called ClientMessage.
(Its actually slight more complex ... ClientMessage has multiple buffers called Frames, and within ClientMessage there is a linked list of Frames, each Frame has its own byte[])
There is an important variable called readOffset; this denotes the offset in the current destination buffer where we should put the next set of bytes.
-1 is the initial state meaning no data has been copied yet. The reason that readOffset is a member variable is because the class is stateful and the main readFrom method in the class is called in a loop.
This is important to understand , so here is an example ... we might invoke the readFrom method with 25,000 bytes in the src buffer. Within the readFrom method, we might derive the overall expected length of the frame is 75,000 bytes. The code will still copy the available 25,000 bytes from the src into the destination buffer, and it will set the readOffset to 25,000. The method will then complete and return false (false == 'incomplete').
Then on the next iteration/call to readFrom , we might have 30,000 new bytes in the src. Again it will copy those bytes into the destination (starting at the readOffset 25,000), then it will bump the readOffset pointer forward to 55,000, and so on and so on ... until it receives the full frame of 75,000 bytes.
Once a frame is full (i.e. we have copied the number of bytes specified in the frame header) , the code will start populating a new frame.
Once a frame is full and it is flagged as IS_FINAL_FLAG = true .. that whole ClientMessage is complete, and the readFrom will return true .
Setup for error:
We have large messages being sent from client to server therefore they tend to span fairly large number of tcp packets.
We also have data being sent from cloud to onprem ... so sometimes those tcp packets can be a few micros apart.
This leads to plenty of occurrences where readFrom is invoked when the src buffer does not have the entire dataset (frame) available. And thus our larger messages are often built up iteratively as described above. That generally works well ... apart from 1 case ..
Sometimes when readFrom is called, the src buffer has the data needed to complete the destination message except the last few bytes of the final frame (5 bytes say).
So, we process as described above what we have, and create an almost complete ClientMessage in memory in ClientMessageReader. Then micros later the remaining data becomes available in the src buffer, and we can append that by another call to readFrom. So that is invoked .. but it hits this check ...
int remaining = src.remaining();
if (remaining < SIZE_OF_FRAME_LENGTH_AND_FLAGS) {
// we don't have even the frame length and flags ready
return false;
}
which won't read the src buffer if its size is less that 6 bytes.
So, then we can't complete the frame / message because the residual part is too small to pass through the above check.
The connection then stays in that state, until at some point later, something else is received in the src buffer (typically a hearbeat). At that point we get through this guard because the buffer exceeds the minimum size, and we can process the 5 bytes that have been sat there waiting (and the heartbeat).
I think the above check should only occur when we are at the beginning of frames i.e. just moved down to iinside the
if (readOffset == -1) {
block. | non_priority | clientmessage frame read for a message may be broken in case of large messages please see line of code the clientmessagereader is basically copying data from one buffer to another buffer the source buffer src in the code is bytes received directly from the client socket connection the destination buffer is an application byte which is wrapped inside an object called clientmessage its actually slight more complex clientmessage has multiple buffers called frames and within clientmessage there is a linked list of frames each frame has its own byte there is an important variable called readoffset this denotes the offset in the current destination buffer where we should put the next set of bytes is the initial state meaning no data has been copied yet the reason that readoffset is a member variable is because the class is stateful and the main readfrom method in the class is called in a loop this is important to understand so here is an example we might invoke the readfrom method with bytes in the src buffer within the readfrom method we might derive the overall expected length of the frame is bytes the code will still copy the available bytes from the src into the destination buffer and it will set the readoffset to the method will then complete and return false false incomplete then on the next iteration call to readfrom we might have new bytes in the src again it will copy those bytes into the destination starting at the readoffset then it will bump the readoffset pointer forward to and so on and so on until it receives the full frame of bytes once a frame is full i e we have copied the number of bytes specified in the frame header the code will start populating a new frame once a frame is full and it is flagged as is final flag true that whole clientmessage is complete and the readfrom will return true setup for error we have large messages being sent from client to server therefore they tend to span fairly large number of tcp packets we also have data being sent from cloud to onprem so sometimes those tcp packets can be a few micros apart this leads to plenty of occurrences where readfrom is invoked when the src buffer does not have the entire dataset frame available and thus our larger messages are often built up iteratively as described above that generally works well apart from case sometimes when readfrom is called the src buffer has the data needed to complete the destination message except the last few bytes of the final frame bytes say so we process as described above what we have and create an almost complete clientmessage in memory in clientmessagereader then micros later the remaining data becomes available in the src buffer and we can append that by another call to readfrom so that is invoked but it hits this check int remaining src remaining if remaining size of frame length and flags we don t have even the frame length and flags ready return false which won t read the src buffer if its size is less that bytes so then we can t complete the frame message because the residual part is too small to pass through the above check the connection then stays in that state until at some point later something else is received in the src buffer typically a hearbeat at that point we get through this guard because the buffer exceeds the minimum size and we can process the bytes that have been sat there waiting and the heartbeat i think the above check should only occur when we are at the beginning of frames i e just moved down to iinside the if readoffset block | 0 |
628,466 | 19,986,647,596 | IssuesEvent | 2022-01-30 19:11:46 | CodersCamp2021-HK/CodersCamp2021.Project.React | https://api.github.com/repos/CodersCamp2021-HK/CodersCamp2021.Project.React | opened | Poprawić routing | priority: high 🟠 type: ♻️ refactor | Teraz jak gh otwieramy to jest, że strona nie istnieje, trzeba dodać prefix do każdego routingu | 1.0 | Poprawić routing - Teraz jak gh otwieramy to jest, że strona nie istnieje, trzeba dodać prefix do każdego routingu | priority | poprawić routing teraz jak gh otwieramy to jest że strona nie istnieje trzeba dodać prefix do każdego routingu | 1 |
94,339 | 11,863,873,051 | IssuesEvent | 2020-03-25 20:33:04 | RoboJackets/igvc-software | https://api.github.com/repos/RoboJackets/igvc-software | opened | Joystick changes | level ➤ medium type ➤ design document type ➤ new feature type ➤ refactor | Our current joystick node's control style, ie. tank drive, is hard to drive around with. It could do some upgrading and improvements.
AC: Better joystick node. | 1.0 | Joystick changes - Our current joystick node's control style, ie. tank drive, is hard to drive around with. It could do some upgrading and improvements.
AC: Better joystick node. | non_priority | joystick changes our current joystick node s control style ie tank drive is hard to drive around with it could do some upgrading and improvements ac better joystick node | 0 |
137,782 | 18,764,965,415 | IssuesEvent | 2021-11-05 21:53:45 | zbmowrey/zbmowrey-com | https://api.github.com/repos/zbmowrey/zbmowrey-com | closed | CVE-2018-20676 (Medium) detected in bootstrap-3.3.1.min.js - autoclosed | security vulnerability | ## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.1.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.1/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.1/js/bootstrap.min.js</a></p>
<p>Path to dependency file: zbmowrey-com/public/vendor/simple-line-icons/scripts/cheatsheet.template.html</p>
<p>Path to vulnerable library: /public/vendor/simple-line-icons/scripts/cheatsheet.template.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/zbmowrey/zbmowrey-com/commit/aa5d5a83db626878704cb84d38f651c02c9b038a">aa5d5a83db626878704cb84d38f651c02c9b038a</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-20676 (Medium) detected in bootstrap-3.3.1.min.js - autoclosed - ## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.1.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.1/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.1/js/bootstrap.min.js</a></p>
<p>Path to dependency file: zbmowrey-com/public/vendor/simple-line-icons/scripts/cheatsheet.template.html</p>
<p>Path to vulnerable library: /public/vendor/simple-line-icons/scripts/cheatsheet.template.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/zbmowrey/zbmowrey-com/commit/aa5d5a83db626878704cb84d38f651c02c9b038a">aa5d5a83db626878704cb84d38f651c02c9b038a</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in bootstrap min js autoclosed cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file zbmowrey com public vendor simple line icons scripts cheatsheet template html path to vulnerable library public vendor simple line icons scripts cheatsheet template html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap step up your open source security game with whitesource | 0 |
44,513 | 13,057,546,121 | IssuesEvent | 2020-07-30 07:32:29 | stefanfreitag/anti-spaghetti-workshop | https://api.github.com/repos/stefanfreitag/anti-spaghetti-workshop | opened | CVE-2018-19839 (Medium) detected in node-sass-4.14.1.tgz, CSS::Sass-v3.4.11 | security vulnerability | ## CVE-2018-19839 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/anti-spaghetti-workshop/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/anti-spaghetti-workshop/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- gulp-sass-4.1.0.tgz (Root Library)
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/stefanfreitag/anti-spaghetti-workshop/commit/f56495930cd3fc69bb3dcc6531eb9bfb4def2eb1">f56495930cd3fc69bb3dcc6531eb9bfb4def2eb1</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass prior to 3.5.5, the function handle_error in sass_context.cpp allows attackers to cause a denial-of-service resulting from a heap-based buffer over-read via a crafted sass file.
<p>Publish Date: 2018-12-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19839>CVE-2018-19839</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19839">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19839</a></p>
<p>Release Date: 2018-12-04</p>
<p>Fix Resolution: Libsass:3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-19839 (Medium) detected in node-sass-4.14.1.tgz, CSS::Sass-v3.4.11 - ## CVE-2018-19839 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/anti-spaghetti-workshop/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/anti-spaghetti-workshop/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- gulp-sass-4.1.0.tgz (Root Library)
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/stefanfreitag/anti-spaghetti-workshop/commit/f56495930cd3fc69bb3dcc6531eb9bfb4def2eb1">f56495930cd3fc69bb3dcc6531eb9bfb4def2eb1</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass prior to 3.5.5, the function handle_error in sass_context.cpp allows attackers to cause a denial-of-service resulting from a heap-based buffer over-read via a crafted sass file.
<p>Publish Date: 2018-12-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19839>CVE-2018-19839</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19839">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19839</a></p>
<p>Release Date: 2018-12-04</p>
<p>Fix Resolution: Libsass:3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in node sass tgz css sass cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm anti spaghetti workshop package json path to vulnerable library tmp ws scm anti spaghetti workshop node modules node sass package json dependency hierarchy gulp sass tgz root library x node sass tgz vulnerable library found in head commit a href vulnerability details in libsass prior to the function handle error in sass context cpp allows attackers to cause a denial of service resulting from a heap based buffer over read via a crafted sass file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
176,463 | 6,559,882,231 | IssuesEvent | 2017-09-07 07:00:17 | gdgphilippines/devfest | https://api.github.com/repos/gdgphilippines/devfest | opened | Change loading of content for each page | enhancement Priority Medium ready | In relation to gdgphilippines/devfestphfiles-2017
Need to change the parsing of files | 1.0 | Change loading of content for each page - In relation to gdgphilippines/devfestphfiles-2017
Need to change the parsing of files | priority | change loading of content for each page in relation to gdgphilippines devfestphfiles need to change the parsing of files | 1 |
708,255 | 24,335,462,602 | IssuesEvent | 2022-10-01 02:46:44 | AY2223S1-CS2103T-W12-3/tp | https://api.github.com/repos/AY2223S1-CS2103T-W12-3/tp | opened | Refactoring packages | priority.High type.Chore | Rename the packages listed below
- [ ] Seedu.Address -> RC4HDB
- [ ] Parser class -> Parser class + CommandParser class | 1.0 | Refactoring packages - Rename the packages listed below
- [ ] Seedu.Address -> RC4HDB
- [ ] Parser class -> Parser class + CommandParser class | priority | refactoring packages rename the packages listed below seedu address parser class parser class commandparser class | 1 |
20,953 | 2,632,574,930 | IssuesEvent | 2015-03-08 08:31:46 | prestaconcept/open-source-management | https://api.github.com/repos/prestaconcept/open-source-management | closed | [Deployment] Handle migration on install | priority-minor | When you have to install an application with migration on a new server, migration will fail cause shema is already uptodate.
Use the command to fill migration table | 1.0 | [Deployment] Handle migration on install - When you have to install an application with migration on a new server, migration will fail cause shema is already uptodate.
Use the command to fill migration table | priority | handle migration on install when you have to install an application with migration on a new server migration will fail cause shema is already uptodate use the command to fill migration table | 1 |
377,115 | 11,163,994,480 | IssuesEvent | 2019-12-27 02:20:13 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Unable to add "Missing" resources from a table when unowned | Medium Priority | Found an issue where I could not add the missing resources to any table after the job has started working if I am not the owner of it. In this case, I attempted to make a mint using an Unowned Anvil. I had the bricks and lumber handy, but because I couldn't hold both at once I had to start with one. Afterwards, however, I could not proceed to add it. No error message, it just doesn't accept it. I am unsure if the behavior is the same if another player owns it, as this was only observed for unowned - created by another player outside of their property. | 1.0 | Unable to add "Missing" resources from a table when unowned - Found an issue where I could not add the missing resources to any table after the job has started working if I am not the owner of it. In this case, I attempted to make a mint using an Unowned Anvil. I had the bricks and lumber handy, but because I couldn't hold both at once I had to start with one. Afterwards, however, I could not proceed to add it. No error message, it just doesn't accept it. I am unsure if the behavior is the same if another player owns it, as this was only observed for unowned - created by another player outside of their property. | priority | unable to add missing resources from a table when unowned found an issue where i could not add the missing resources to any table after the job has started working if i am not the owner of it in this case i attempted to make a mint using an unowned anvil i had the bricks and lumber handy but because i couldn t hold both at once i had to start with one afterwards however i could not proceed to add it no error message it just doesn t accept it i am unsure if the behavior is the same if another player owns it as this was only observed for unowned created by another player outside of their property | 1 |
401,555 | 11,795,117,290 | IssuesEvent | 2020-03-18 08:17:40 | thaliawww/concrexit | https://api.github.com/repos/thaliawww/concrexit | closed | Sync mailinglists using Google Admin SDK | feature priority: medium | In GitLab by @se-bastiaan on Oct 7, 2019, 13:02
### One-sentence description
Sync mailinglists using Google Admin SDK
### Motivation
Because we switch to GSuite
### Desired functionality
Sync the mailinglists to
### Suggested implementation
We currently have a syncmailinglists scripts in the serverconfig repo that uses the mailinglists API. Instead of syncing to Mailman we want it to sync to GSuite. And by replacing the script with a management command we can move this code to concrexit.
You probably want to make sure this has 100% test coverage. | 1.0 | Sync mailinglists using Google Admin SDK - In GitLab by @se-bastiaan on Oct 7, 2019, 13:02
### One-sentence description
Sync mailinglists using Google Admin SDK
### Motivation
Because we switch to GSuite
### Desired functionality
Sync the mailinglists to
### Suggested implementation
We currently have a syncmailinglists scripts in the serverconfig repo that uses the mailinglists API. Instead of syncing to Mailman we want it to sync to GSuite. And by replacing the script with a management command we can move this code to concrexit.
You probably want to make sure this has 100% test coverage. | priority | sync mailinglists using google admin sdk in gitlab by se bastiaan on oct one sentence description sync mailinglists using google admin sdk motivation because we switch to gsuite desired functionality sync the mailinglists to suggested implementation we currently have a syncmailinglists scripts in the serverconfig repo that uses the mailinglists api instead of syncing to mailman we want it to sync to gsuite and by replacing the script with a management command we can move this code to concrexit you probably want to make sure this has test coverage | 1 |
778,842 | 27,331,370,835 | IssuesEvent | 2023-02-25 17:19:26 | EnsembleUI/ensemble | https://api.github.com/repos/EnsembleUI/ensemble | closed | Markdown links color to match primary color | enhancement HIGHEST Priority | Currently links in a markdown widget are blue. They should match the theme's primary color.
```yaml
View:
Column:
children:
- Markdown:
text: Here is an [inline links](https://ensembleui.com)
``` | 1.0 | Markdown links color to match primary color - Currently links in a markdown widget are blue. They should match the theme's primary color.
```yaml
View:
Column:
children:
- Markdown:
text: Here is an [inline links](https://ensembleui.com)
``` | priority | markdown links color to match primary color currently links in a markdown widget are blue they should match the theme s primary color yaml view column children markdown text here is an | 1 |
48,277 | 7,398,674,675 | IssuesEvent | 2018-03-19 07:34:05 | PaddlePaddle/Paddle | https://api.github.com/repos/PaddlePaddle/Paddle | opened | API doc problems of reshape | PythonAPI documentation | #### 针对以下问题修改时请协调组织其他内容语言
#### 单词拼写、遗漏和语法等问题未列出,修改时请注意
*请参考[API注释标准](https://github.com/PaddlePaddle/Paddle/blob/develop/doc/fluid/dev/api_doc_std_cn.md)!*
------
源文件:https://github.com/PaddlePaddle/Paddle/blob/develop/python/paddle/fluid/layers/ops.py
- reshape
http://www.paddlepaddle.org/docs/develop/api/en/fluid/layers.html#reshape
- 缺 Examples
- 表达须完善
- 格式问题(数据类型介绍等) | 1.0 | API doc problems of reshape - #### 针对以下问题修改时请协调组织其他内容语言
#### 单词拼写、遗漏和语法等问题未列出,修改时请注意
*请参考[API注释标准](https://github.com/PaddlePaddle/Paddle/blob/develop/doc/fluid/dev/api_doc_std_cn.md)!*
------
源文件:https://github.com/PaddlePaddle/Paddle/blob/develop/python/paddle/fluid/layers/ops.py
- reshape
http://www.paddlepaddle.org/docs/develop/api/en/fluid/layers.html#reshape
- 缺 Examples
- 表达须完善
- 格式问题(数据类型介绍等) | non_priority | api doc problems of reshape 针对以下问题修改时请协调组织其他内容语言 单词拼写、遗漏和语法等问题未列出,修改时请注意 请参考 源文件: reshape 缺 examples 表达须完善 格式问题(数据类型介绍等) | 0 |
822,123 | 30,854,183,933 | IssuesEvent | 2023-08-02 19:09:46 | calcom/cal.com | https://api.github.com/repos/calcom/cal.com | closed | [CAL-2120] Org owner/admin can't see the teams created by other owner/admin | 🐛 bug linear High priority organizations | He should be able to see all teams so that he can add himself to any team he wants.
He shouldn't be able to see a team's event-types unless he has added him to the team.
<sub>From [SyncLinear.com](https://synclinear.com) | [CAL-2120](https://linear.app/calcom/issue/CAL-2120/org-owneradmin-cant-see-the-teams-created-by-other-owneradmin)</sub> | 1.0 | [CAL-2120] Org owner/admin can't see the teams created by other owner/admin - He should be able to see all teams so that he can add himself to any team he wants.
He shouldn't be able to see a team's event-types unless he has added him to the team.
<sub>From [SyncLinear.com](https://synclinear.com) | [CAL-2120](https://linear.app/calcom/issue/CAL-2120/org-owneradmin-cant-see-the-teams-created-by-other-owneradmin)</sub> | priority | org owner admin can t see the teams created by other owner admin he should be able to see all teams so that he can add himself to any team he wants he shouldn t be able to see a team s event types unless he has added him to the team from | 1 |
350,781 | 10,508,486,341 | IssuesEvent | 2019-09-27 08:46:08 | woocommerce/woocommerce-gateway-paypal-express-checkout | https://api.github.com/repos/woocommerce/woocommerce-gateway-paypal-express-checkout | closed | Synchronise Renewals Not Supported ATM | Priority: Low [Type] Enhancement | Hello,
I am delighted to see that PayPal Subscriptions for download/virtual products are now supported. However, I noticed that apparently Synchronise renewals are not a thing with PayPal? Is this something that can be added?
I tried this with the same product and switching the setting to sync to 1st of the months and not sync at all and in the checkout process, it offered PayPal only when Sync Renewals were turned off for the subscription I was trying to make.
PayPal CHeckout Version: 1.6.16
WooCommerce Version: 3.6.5
WooCommerce Subscriptions Version: 2.5.7
Thanks to all putting in the effort for the Subscription support in the first place,
Chris | 1.0 | Synchronise Renewals Not Supported ATM - Hello,
I am delighted to see that PayPal Subscriptions for download/virtual products are now supported. However, I noticed that apparently Synchronise renewals are not a thing with PayPal? Is this something that can be added?
I tried this with the same product and switching the setting to sync to 1st of the months and not sync at all and in the checkout process, it offered PayPal only when Sync Renewals were turned off for the subscription I was trying to make.
PayPal CHeckout Version: 1.6.16
WooCommerce Version: 3.6.5
WooCommerce Subscriptions Version: 2.5.7
Thanks to all putting in the effort for the Subscription support in the first place,
Chris | priority | synchronise renewals not supported atm hello i am delighted to see that paypal subscriptions for download virtual products are now supported however i noticed that apparently synchronise renewals are not a thing with paypal is this something that can be added i tried this with the same product and switching the setting to sync to of the months and not sync at all and in the checkout process it offered paypal only when sync renewals were turned off for the subscription i was trying to make paypal checkout version woocommerce version woocommerce subscriptions version thanks to all putting in the effort for the subscription support in the first place chris | 1 |
71,773 | 18,872,083,421 | IssuesEvent | 2021-11-13 11:08:42 | test-st-petersburg/DocTemplates | https://api.github.com/repos/test-st-petersburg/DocTemplates | opened | Невозможно установить требуемые nuget модули | bug build dependencies | ## Требуемое поведение
`Invoke-PSDepend` должен выполняться без ошибок.
## Текущее поведение
При установке на "голую" машину (на которой не был сконфигурирован nuget) не один из модулей nuget не удаётся установить.
## Возможное решение
- [ ] необходимо добавить репозиторий в конфигурацию nuget либо явно в командную строку
- [ ] при ошибке восстановления модулей nuget процесс сборки должен прерываться. Сейчас это не так. | 1.0 | Невозможно установить требуемые nuget модули - ## Требуемое поведение
`Invoke-PSDepend` должен выполняться без ошибок.
## Текущее поведение
При установке на "голую" машину (на которой не был сконфигурирован nuget) не один из модулей nuget не удаётся установить.
## Возможное решение
- [ ] необходимо добавить репозиторий в конфигурацию nuget либо явно в командную строку
- [ ] при ошибке восстановления модулей nuget процесс сборки должен прерываться. Сейчас это не так. | non_priority | невозможно установить требуемые nuget модули требуемое поведение invoke psdepend должен выполняться без ошибок текущее поведение при установке на голую машину на которой не был сконфигурирован nuget не один из модулей nuget не удаётся установить возможное решение необходимо добавить репозиторий в конфигурацию nuget либо явно в командную строку при ошибке восстановления модулей nuget процесс сборки должен прерываться сейчас это не так | 0 |
68,670 | 7,107,108,862 | IssuesEvent | 2018-01-16 18:48:52 | webcompat/webcompat.com | https://api.github.com/repos/webcompat/webcompat.com | closed | Add optional "browsers" argument for functional tests | in progress type: testing | I had already done that while trying to get tests to run in Safari, but didn't merge it in since those tests failed. I think this is helpful, specially when working with the tests, troubleshooting, etc.
My idea is to add the `browsers` argument and if the user doesn't specify anything, just run all the available ones.
It would look like this: `npm run test:js browsers=safari,chrome,firefox` | 1.0 | Add optional "browsers" argument for functional tests - I had already done that while trying to get tests to run in Safari, but didn't merge it in since those tests failed. I think this is helpful, specially when working with the tests, troubleshooting, etc.
My idea is to add the `browsers` argument and if the user doesn't specify anything, just run all the available ones.
It would look like this: `npm run test:js browsers=safari,chrome,firefox` | non_priority | add optional browsers argument for functional tests i had already done that while trying to get tests to run in safari but didn t merge it in since those tests failed i think this is helpful specially when working with the tests troubleshooting etc my idea is to add the browsers argument and if the user doesn t specify anything just run all the available ones it would look like this npm run test js browsers safari chrome firefox | 0 |
356,722 | 25,176,249,748 | IssuesEvent | 2022-11-11 09:31:07 | seelengxd/pe | https://api.github.com/repos/seelengxd/pe | opened | Use cases: wrong extension numbering? | severity.VeryLow type.DocumentationBug | Selecting lesson slot extension (DG - Use Cases)

<!--session: 1668153014144-6643e22d-f7c4-4d37-ba3e-b9b7512b6d35-->
<!--Version: Web v3.4.4--> | 1.0 | Use cases: wrong extension numbering? - Selecting lesson slot extension (DG - Use Cases)

<!--session: 1668153014144-6643e22d-f7c4-4d37-ba3e-b9b7512b6d35-->
<!--Version: Web v3.4.4--> | non_priority | use cases wrong extension numbering selecting lesson slot extension dg use cases | 0 |
524,212 | 15,208,989,908 | IssuesEvent | 2021-02-17 04:08:41 | brave/qa-resources | https://api.github.com/repos/brave/qa-resources | closed | iOS 1.23 release | iOS 12 iOS 13 iOS 14 priority/P1 | ### <img src="https://www.rebron.org/wordpress/wp-content/uploads/2019/06/ios-1.png">`iOS 1.23 Release`
#### Release Date/Target:
* Release Date: **`February 12, 2021`**
#### Summary:
Includes the following features/fixes:
* Security fixes
* Shields fixes
* Brave-Today Enhancements
* UI/UX Enhancements
**Milestone:**
* https://github.com/brave/brave-ios/milestone/52
**Current Progress:**
* iPhone --> **`Completed`**
* Tablet --> **`Completed`**
**Manual Passes:**
* iOS 14 --> https://github.com/brave/brave-ios/issues/3284
* iOS 13--> https://github.com/brave/brave-ios/issues/3283
* iOS 12 --> https://github.com/brave/brave-ios/issues/3282 | 1.0 | iOS 1.23 release - ### <img src="https://www.rebron.org/wordpress/wp-content/uploads/2019/06/ios-1.png">`iOS 1.23 Release`
#### Release Date/Target:
* Release Date: **`February 12, 2021`**
#### Summary:
Includes the following features/fixes:
* Security fixes
* Shields fixes
* Brave-Today Enhancements
* UI/UX Enhancements
**Milestone:**
* https://github.com/brave/brave-ios/milestone/52
**Current Progress:**
* iPhone --> **`Completed`**
* Tablet --> **`Completed`**
**Manual Passes:**
* iOS 14 --> https://github.com/brave/brave-ios/issues/3284
* iOS 13--> https://github.com/brave/brave-ios/issues/3283
* iOS 12 --> https://github.com/brave/brave-ios/issues/3282 | priority | ios release img src release release date target release date february summary includes the following features fixes security fixes shields fixes brave today enhancements ui ux enhancements milestone current progress iphone completed tablet completed manual passes ios ios ios | 1 |
238,454 | 19,721,397,130 | IssuesEvent | 2022-01-13 15:41:43 | 389ds/389-ds-base | https://api.github.com/repos/389ds/389-ds-base | closed | ns-slapd crash in CI test plugins test_dna_interval_with_different_values | closed: duplicate CI test needs triage priority_high | **Issue Description**
ns-slapd crash in CI test plugins test_dna_interval_with_different_values
because the operation modifiers is corrupted when trying to log the operation
According to the operation we are handling modify operation on uid=interval3,ou=People,dc=example,dc=com replace uidNumber: -1
Crash is systematic both with bdb and mdb ( maybe a Bug in normalize_mods2bvals ? )
**Package Version and Platform:**
- Platform: fedora 35:
- Package and version: I reproduced it with [Issue 4939 PR](https://github.com/389ds/389-ds-base/pull/5071) that said it probably occurs on master too
**Steps to Reproduce**
Set up 1minuteTips VM with fedora 35
Once test is complete log with ssh and run:
```
pip install -U setuptools
pip install -U pytest
pip install -U slugify
sysctl -w fs.suid_dumpable=1
cd /mnt/tests/rhds/tests/upstream/ds/dirsrvtests/tests/suites/plugins
#export NSSLAPD_DB_LIB=mdb
DEBUGGING=1 py.test --log-level=debug -vvvvv $PWD -k test_dna_interval_with_different_values
coredumpctl debug
```
**Expected results**
Test should be pass and coredumpctl should not report any core file
( Currently test is PASS but a core file is reported)
**Additional context**
```
Core was generated by `/usr/sbin/ns-slapd -D /etc/dirsrv/slapd-standalone1 -i /run/dirsrv/slapd-standa'.
Program terminated with signal SIGSEGV, Segmentation fault.
#0 write_audit_file (logtype=logtype@entry=4, optype=<optimized out>, dn=dn@entry=0x7f7575f82e90 "uid=interval3,ou=People,dc=example,dc=com", change=change@entry=0x7f7575fc3600, flag=flag@entry=0, curtime=curtime@entry=1642082606, rc=1, sourcelog=8) at ldap/servers/slapd/auditlog.c:233
233 int operationtype = mods[j]->mod_op & ~LDAP_MOD_BVALUES;
[Current thread is 1 (Thread 0x7f7579352640 (LWP 323802))]
(gdb) where
#0 write_audit_file
(logtype=logtype@entry=4, optype=<optimized out>, dn=dn@entry=0x7f7575f82e90 "uid=interval3,ou=People,dc=example,dc=com", change=change@entry=0x7f7575fc3600, flag=flag@entry=0, curtime=curtime@entry=1642082606, rc=1, sourcelog=8) at ldap/servers/slapd/auditlog.c:233
#1 0x00007f7592edab13 in write_auditfail_log_entry (pb=<optimized out>) at ldap/servers/slapd/auditlog.c:166
#2 0x00007f7592f24181 in op_shared_modify (pb=pb@entry=0x7f758d7f2540, pw_change=pw_change@entry=0, old_pw=0x0) at ldap/servers/slapd/modify.c:1047
#3 0x00007f7592f24b8d in do_modify (pb=pb@entry=0x7f758d7f2540) at ldap/servers/slapd/modify.c:377
#4 0x000055b4e5349ce9 in connection_dispatch_operation (pb=0x7f758d7f2540, op=<optimized out>, conn=<optimized out>) at ldap/servers/slapd/connection.c:648
#5 connection_threadmain (arg=<optimized out>) at ldap/servers/slapd/connection.c:1800
#6 0x00007f75929a8524 in _pt_root (arg=0x7f758edb8140) at ../../../../nspr/pr/src/pthreads/ptthread.c:201
#7 0x00007f7592c69a87 in start_thread (arg=<optimized out>) at pthread_create.c:435
#8 0x00007f7592cee640 in clone3 () at ../sysdeps/unix/sysv/linux/x86_64/clone3.S:81
(gdb) up 2
#2 0x00007f7592f24181 in op_shared_modify (pb=pb@entry=0x7f758d7f2540, pw_change=pw_change@entry=0, old_pw=0x0) at ldap/servers/slapd/modify.c:1047
1047 write_auditfail_log_entry(pb); /* Record the operation in the audit log */
(gdb) p *pb->pb_op->o_params.p
Attempt to take contents of a non-pointer value.
(gdb) p pb->pb_op->o_params.p
$1 = {p_add = {target_entry = 0x7f7575fc3600, parentuniqueid = 0x0}, p_bind = {bind_method = 140142467364352, bind_creds = 0x0, bind_saslmechanism = 0x0, bind_ret_saslcreds = 0x0}, p_compare = {compare_ava = {
ava_type = 0x7f7575fc3600 "\202", ava_value = {bv_len = 0, bv_val = 0x0}, ava_private = 0x0}}, p_modify = {modify_mods = 0x7f7575fc3600}, p_modrdn = {modrdn_newrdn = 0x7f7575fc3600 "\202",
modrdn_deloldrdn = 0, modrdn_newsuperior_address = {udn = 0x0, uniqueid = 0x0, sdn = 0x0}, modrdn_mods = 0x0}, p_search = {search_scope = 1979463168, search_deref = 32629, search_sizelimit = 0,
search_timelimit = 0, search_filter = 0x0, search_strfilter = 0x0, search_attrs = 0x0, search_attrsonly = 0, search_is_and = 0, search_gerattrs = 0x0}, p_abandon = {abandon_targetmsgid = 1979463168},
p_extended = {exop_oid = 0x7f7575fc3600 "\202", exop_value = 0x0}}
(gdb) p pb->pb_op->o_params.p.p_modify.modify_mods
$2 = (LDAPMod **) 0x7f7575fc3600
(gdb) p pb->pb_op->o_params.p.p_modify.modify_mods[0][0]
Cannot access memory at address 0x7f7500000082
```
Note: Although the test is PASS this bug may cause failure to other CI tests (because the instance is down while it is expected up) | 1.0 | ns-slapd crash in CI test plugins test_dna_interval_with_different_values - **Issue Description**
ns-slapd crash in CI test plugins test_dna_interval_with_different_values
because the operation modifiers is corrupted when trying to log the operation
According to the operation we are handling modify operation on uid=interval3,ou=People,dc=example,dc=com replace uidNumber: -1
Crash is systematic both with bdb and mdb ( maybe a Bug in normalize_mods2bvals ? )
**Package Version and Platform:**
- Platform: fedora 35:
- Package and version: I reproduced it with [Issue 4939 PR](https://github.com/389ds/389-ds-base/pull/5071) that said it probably occurs on master too
**Steps to Reproduce**
Set up 1minuteTips VM with fedora 35
Once test is complete log with ssh and run:
```
pip install -U setuptools
pip install -U pytest
pip install -U slugify
sysctl -w fs.suid_dumpable=1
cd /mnt/tests/rhds/tests/upstream/ds/dirsrvtests/tests/suites/plugins
#export NSSLAPD_DB_LIB=mdb
DEBUGGING=1 py.test --log-level=debug -vvvvv $PWD -k test_dna_interval_with_different_values
coredumpctl debug
```
**Expected results**
Test should be pass and coredumpctl should not report any core file
( Currently test is PASS but a core file is reported)
**Additional context**
```
Core was generated by `/usr/sbin/ns-slapd -D /etc/dirsrv/slapd-standalone1 -i /run/dirsrv/slapd-standa'.
Program terminated with signal SIGSEGV, Segmentation fault.
#0 write_audit_file (logtype=logtype@entry=4, optype=<optimized out>, dn=dn@entry=0x7f7575f82e90 "uid=interval3,ou=People,dc=example,dc=com", change=change@entry=0x7f7575fc3600, flag=flag@entry=0, curtime=curtime@entry=1642082606, rc=1, sourcelog=8) at ldap/servers/slapd/auditlog.c:233
233 int operationtype = mods[j]->mod_op & ~LDAP_MOD_BVALUES;
[Current thread is 1 (Thread 0x7f7579352640 (LWP 323802))]
(gdb) where
#0 write_audit_file
(logtype=logtype@entry=4, optype=<optimized out>, dn=dn@entry=0x7f7575f82e90 "uid=interval3,ou=People,dc=example,dc=com", change=change@entry=0x7f7575fc3600, flag=flag@entry=0, curtime=curtime@entry=1642082606, rc=1, sourcelog=8) at ldap/servers/slapd/auditlog.c:233
#1 0x00007f7592edab13 in write_auditfail_log_entry (pb=<optimized out>) at ldap/servers/slapd/auditlog.c:166
#2 0x00007f7592f24181 in op_shared_modify (pb=pb@entry=0x7f758d7f2540, pw_change=pw_change@entry=0, old_pw=0x0) at ldap/servers/slapd/modify.c:1047
#3 0x00007f7592f24b8d in do_modify (pb=pb@entry=0x7f758d7f2540) at ldap/servers/slapd/modify.c:377
#4 0x000055b4e5349ce9 in connection_dispatch_operation (pb=0x7f758d7f2540, op=<optimized out>, conn=<optimized out>) at ldap/servers/slapd/connection.c:648
#5 connection_threadmain (arg=<optimized out>) at ldap/servers/slapd/connection.c:1800
#6 0x00007f75929a8524 in _pt_root (arg=0x7f758edb8140) at ../../../../nspr/pr/src/pthreads/ptthread.c:201
#7 0x00007f7592c69a87 in start_thread (arg=<optimized out>) at pthread_create.c:435
#8 0x00007f7592cee640 in clone3 () at ../sysdeps/unix/sysv/linux/x86_64/clone3.S:81
(gdb) up 2
#2 0x00007f7592f24181 in op_shared_modify (pb=pb@entry=0x7f758d7f2540, pw_change=pw_change@entry=0, old_pw=0x0) at ldap/servers/slapd/modify.c:1047
1047 write_auditfail_log_entry(pb); /* Record the operation in the audit log */
(gdb) p *pb->pb_op->o_params.p
Attempt to take contents of a non-pointer value.
(gdb) p pb->pb_op->o_params.p
$1 = {p_add = {target_entry = 0x7f7575fc3600, parentuniqueid = 0x0}, p_bind = {bind_method = 140142467364352, bind_creds = 0x0, bind_saslmechanism = 0x0, bind_ret_saslcreds = 0x0}, p_compare = {compare_ava = {
ava_type = 0x7f7575fc3600 "\202", ava_value = {bv_len = 0, bv_val = 0x0}, ava_private = 0x0}}, p_modify = {modify_mods = 0x7f7575fc3600}, p_modrdn = {modrdn_newrdn = 0x7f7575fc3600 "\202",
modrdn_deloldrdn = 0, modrdn_newsuperior_address = {udn = 0x0, uniqueid = 0x0, sdn = 0x0}, modrdn_mods = 0x0}, p_search = {search_scope = 1979463168, search_deref = 32629, search_sizelimit = 0,
search_timelimit = 0, search_filter = 0x0, search_strfilter = 0x0, search_attrs = 0x0, search_attrsonly = 0, search_is_and = 0, search_gerattrs = 0x0}, p_abandon = {abandon_targetmsgid = 1979463168},
p_extended = {exop_oid = 0x7f7575fc3600 "\202", exop_value = 0x0}}
(gdb) p pb->pb_op->o_params.p.p_modify.modify_mods
$2 = (LDAPMod **) 0x7f7575fc3600
(gdb) p pb->pb_op->o_params.p.p_modify.modify_mods[0][0]
Cannot access memory at address 0x7f7500000082
```
Note: Although the test is PASS this bug may cause failure to other CI tests (because the instance is down while it is expected up) | non_priority | ns slapd crash in ci test plugins test dna interval with different values issue description ns slapd crash in ci test plugins test dna interval with different values because the operation modifiers is corrupted when trying to log the operation according to the operation we are handling modify operation on uid ou people dc example dc com replace uidnumber crash is systematic both with bdb and mdb maybe a bug in normalize package version and platform platform fedora package and version i reproduced it with that said it probably occurs on master too steps to reproduce set up vm with fedora once test is complete log with ssh and run pip install u setuptools pip install u pytest pip install u slugify sysctl w fs suid dumpable cd mnt tests rhds tests upstream ds dirsrvtests tests suites plugins export nsslapd db lib mdb debugging py test log level debug vvvvv pwd k test dna interval with different values coredumpctl debug expected results test should be pass and coredumpctl should not report any core file currently test is pass but a core file is reported additional context core was generated by usr sbin ns slapd d etc dirsrv slapd i run dirsrv slapd standa program terminated with signal sigsegv segmentation fault write audit file logtype logtype entry optype dn dn entry uid ou people dc example dc com change change entry flag flag entry curtime curtime entry rc sourcelog at ldap servers slapd auditlog c int operationtype mods mod op ldap mod bvalues gdb where write audit file logtype logtype entry optype dn dn entry uid ou people dc example dc com change change entry flag flag entry curtime curtime entry rc sourcelog at ldap servers slapd auditlog c in write auditfail log entry pb at ldap servers slapd auditlog c in op shared modify pb pb entry pw change pw change entry old pw at ldap servers slapd modify c in do modify pb pb entry at ldap servers slapd modify c in connection dispatch operation pb op conn at ldap servers slapd connection c connection threadmain arg at ldap servers slapd connection c in pt root arg at nspr pr src pthreads ptthread c in start thread arg at pthread create c in at sysdeps unix sysv linux s gdb up in op shared modify pb pb entry pw change pw change entry old pw at ldap servers slapd modify c write auditfail log entry pb record the operation in the audit log gdb p pb pb op o params p attempt to take contents of a non pointer value gdb p pb pb op o params p p add target entry parentuniqueid p bind bind method bind creds bind saslmechanism bind ret saslcreds p compare compare ava ava type ava value bv len bv val ava private p modify modify mods p modrdn modrdn newrdn modrdn deloldrdn modrdn newsuperior address udn uniqueid sdn modrdn mods p search search scope search deref search sizelimit search timelimit search filter search strfilter search attrs search attrsonly search is and search gerattrs p abandon abandon targetmsgid p extended exop oid exop value gdb p pb pb op o params p p modify modify mods ldapmod gdb p pb pb op o params p p modify modify mods cannot access memory at address note although the test is pass this bug may cause failure to other ci tests because the instance is down while it is expected up | 0 |
680,944 | 23,291,273,102 | IssuesEvent | 2022-08-05 23:34:12 | GoogleContainerTools/skaffold | https://api.github.com/repos/GoogleContainerTools/skaffold | closed | [v2][verify] skaffold verify docker.io images that have a "/" in them - ex: "alpine/curl:latest" | kind/bug priority/p1 2.0.0 area/verify | Currently `skaffold verify` is not able to properly parse skaffold docker.io images (with the repo ommitted) that have a "/" in them - ex: "alpine/curl:latest". For example:
working `verify` stanza
```
- name: alpine-wget
container:
name: alpine-wget
image: alpine:3.15.4
command: ["/bin/sh"]
args: ["-c", "wget http://www.google.com"]
```
not working `verify` stanza
```
- name: alpine-curl
container:
name: alpine-curl
image: alpine/curl:3.15.4
args: ["http://www.google.com"]
```
error output:
```
aprindle@aprindle ~/verify-test $ skaffold verify -a build.artifacts
Tags used in verification:
- integration-test-container -> gcr.io/aprindle-test-cluster/integration-test-container:latest@sha256:6d6da2378765cd9dda71cbd20f3cf5818c92d49ab98a2554de12d034613dfa6a
- metrics-test-container -> gcr.io/aprindle-test-cluster/metrics-test-container:latest@sha256:3fbce881177ead1c2ae00d58974fd6959c648d7691593f6448892c04139355f7
[integration-test-container] Integration Test 1/4 Running ...
[integration-test-container] Integration Test 1/4 Passed!
[integration-test-container] Integration Test 2/4 Running...!
[integration-test-container] Integration Test 2/4 Passed!
[integration-test-container] Integration Test 3/4 Running...!
[integration-test-container] Integration Test 3/4 Passed!
[integration-test-container] Integration Test 4/4 Running...!
[integration-test-container] Integration Test 4/4 Passed!
[metrics-test-container] Metrics test in progress...
[metrics-test-container] Metrics test passed!
Error response from daemon: manifest for alpine/curl:3.15.4 not found: manifest unknown: manifest unknown
``` | 1.0 | [v2][verify] skaffold verify docker.io images that have a "/" in them - ex: "alpine/curl:latest" - Currently `skaffold verify` is not able to properly parse skaffold docker.io images (with the repo ommitted) that have a "/" in them - ex: "alpine/curl:latest". For example:
working `verify` stanza
```
- name: alpine-wget
container:
name: alpine-wget
image: alpine:3.15.4
command: ["/bin/sh"]
args: ["-c", "wget http://www.google.com"]
```
not working `verify` stanza
```
- name: alpine-curl
container:
name: alpine-curl
image: alpine/curl:3.15.4
args: ["http://www.google.com"]
```
error output:
```
aprindle@aprindle ~/verify-test $ skaffold verify -a build.artifacts
Tags used in verification:
- integration-test-container -> gcr.io/aprindle-test-cluster/integration-test-container:latest@sha256:6d6da2378765cd9dda71cbd20f3cf5818c92d49ab98a2554de12d034613dfa6a
- metrics-test-container -> gcr.io/aprindle-test-cluster/metrics-test-container:latest@sha256:3fbce881177ead1c2ae00d58974fd6959c648d7691593f6448892c04139355f7
[integration-test-container] Integration Test 1/4 Running ...
[integration-test-container] Integration Test 1/4 Passed!
[integration-test-container] Integration Test 2/4 Running...!
[integration-test-container] Integration Test 2/4 Passed!
[integration-test-container] Integration Test 3/4 Running...!
[integration-test-container] Integration Test 3/4 Passed!
[integration-test-container] Integration Test 4/4 Running...!
[integration-test-container] Integration Test 4/4 Passed!
[metrics-test-container] Metrics test in progress...
[metrics-test-container] Metrics test passed!
Error response from daemon: manifest for alpine/curl:3.15.4 not found: manifest unknown: manifest unknown
``` | priority | skaffold verify docker io images that have a in them ex alpine curl latest currently skaffold verify is not able to properly parse skaffold docker io images with the repo ommitted that have a in them ex alpine curl latest for example working verify stanza name alpine wget container name alpine wget image alpine command args not working verify stanza name alpine curl container name alpine curl image alpine curl args error output aprindle aprindle verify test skaffold verify a build artifacts tags used in verification integration test container gcr io aprindle test cluster integration test container latest metrics test container gcr io aprindle test cluster metrics test container latest integration test running integration test passed integration test running integration test passed integration test running integration test passed integration test running integration test passed metrics test in progress metrics test passed error response from daemon manifest for alpine curl not found manifest unknown manifest unknown | 1 |
48,828 | 12,247,171,740 | IssuesEvent | 2020-05-05 15:29:57 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Hitting restore timeouts with NuGet | area-Infrastructure blocking-clean-ci blocking-official-build untriaged | Hitting timeout issues with nuget.org.
```
Restore completed in 5.75 sec for /Users/runner/runners/2.166.2/work/1/s/src/VS.Web.CG.Msbuild/VS.Web.CG.Msbuild.csproj.
Retrying 'FindPackagesByIdAsync' for source 'https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json'.
The HTTP request to 'GET https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json' has timed out after 100000ms.
Retrying 'FindPackagesByIdAsync' for source 'https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json'.
The HTTP request to 'GET https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json' has timed out after 100000ms.
/Users/runner/runners/2.166.2/work/1/s/.dotnet/sdk/5.0.100-preview.2.20165.1/NuGet.targets(128,5): error : Failed to retrieve information about 'Microsoft.CodeAnalysis.Common' from remote source 'https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json'. [/Users/runner/runners/2.166.2/work/1/s/Scaffolding.sln]
/Users/runner/runners/2.166.2/work/1/s/.dotnet/sdk/5.0.100-preview.2.20165.1/NuGet.targets(128,5): error : The HTTP request to 'GET https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json' has timed out after 100000ms. [/Users/runner/runners/2.166.2/work/1/s/Scaffolding.sln]
##[error].dotnet/sdk/5.0.100-preview.2.20165.1/NuGet.targets(128,5): error : Failed to retrieve information about 'Microsoft.CodeAnalysis.Common' from remote source 'https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json'.
The HTTP request to 'GET https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json' has timed out after 100000ms.
```
This has not yet hit our CI but is already impacting a number of PRs
Core-eng Issue https://github.com/dotnet/core-eng/issues/9634
<!-- runfo report start -->
|Definition|Build|Kind|Job Name|
|---|---|---|---|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604353](https://dev.azure.com/dnceng/public/_build/results?buildId=604353)|[PR 35021](https://github.com/dotnet/runtime/pull/35021)|Libraries Build tvOS x64 Debug|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[604340](https://dev.azure.com/dnceng/public/_build/results?buildId=604340)|[PR 20903](https://github.com/dotnet/aspnetcore/pull/20903)|Test: macOS 10.13|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build tvOS x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS arm debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|CoreCLR Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|CoreCLR Product Build OSX x64 checked|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build tvOS x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono LLVM Product Build OSX x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build OSX x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS arm release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build tvOS arm64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS arm64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS arm64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Libraries Build OSX x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono LLVM Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build tvOS arm64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|Libraries Build tvOS x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|CoreCLR Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|Libraries Build iOS arm64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|Libraries Build tvOS arm64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|Libraries Build iOS x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604243](https://dev.azure.com/dnceng/public/_build/results?buildId=604243)|[PR 34933](https://github.com/dotnet/runtime/pull/34933)|Libraries Build tvOS x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604243](https://dev.azure.com/dnceng/public/_build/results?buildId=604243)|[PR 34933](https://github.com/dotnet/runtime/pull/34933)|Libraries Build iOS arm64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604243](https://dev.azure.com/dnceng/public/_build/results?buildId=604243)|[PR 34933](https://github.com/dotnet/runtime/pull/34933)|Libraries Build tvOS arm64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604243](https://dev.azure.com/dnceng/public/_build/results?buildId=604243)|[PR 34933](https://github.com/dotnet/runtime/pull/34933)|Libraries Build iOS x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604212](https://dev.azure.com/dnceng/public/_build/results?buildId=604212)|[PR 34825](https://github.com/dotnet/runtime/pull/34825)|Installer Build and Test tvOS_x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604116](https://dev.azure.com/dnceng/public/_build/results?buildId=604116)|[PR 34767](https://github.com/dotnet/runtime/pull/34767)|Libraries Build iOS arm64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604116](https://dev.azure.com/dnceng/public/_build/results?buildId=604116)|[PR 34767](https://github.com/dotnet/runtime/pull/34767)|Libraries Build tvOS arm64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[603910](https://dev.azure.com/dnceng/public/_build/results?buildId=603910)|[PR 35010](https://github.com/dotnet/runtime/pull/35010)|Libraries Build tvOS x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[603244](https://dev.azure.com/dnceng/public/_build/results?buildId=603244)|[PR 35040](https://github.com/dotnet/runtime/pull/35040)|Libraries Test Run release mono OSX x64 Debug|
Most [recent](https://dev.azure.com/dnceng/public/_build/results?buildId=604353) failure 4/16/2020 1:42:00 PM
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
| 1.0 | Hitting restore timeouts with NuGet - Hitting timeout issues with nuget.org.
```
Restore completed in 5.75 sec for /Users/runner/runners/2.166.2/work/1/s/src/VS.Web.CG.Msbuild/VS.Web.CG.Msbuild.csproj.
Retrying 'FindPackagesByIdAsync' for source 'https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json'.
The HTTP request to 'GET https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json' has timed out after 100000ms.
Retrying 'FindPackagesByIdAsync' for source 'https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json'.
The HTTP request to 'GET https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json' has timed out after 100000ms.
/Users/runner/runners/2.166.2/work/1/s/.dotnet/sdk/5.0.100-preview.2.20165.1/NuGet.targets(128,5): error : Failed to retrieve information about 'Microsoft.CodeAnalysis.Common' from remote source 'https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json'. [/Users/runner/runners/2.166.2/work/1/s/Scaffolding.sln]
/Users/runner/runners/2.166.2/work/1/s/.dotnet/sdk/5.0.100-preview.2.20165.1/NuGet.targets(128,5): error : The HTTP request to 'GET https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json' has timed out after 100000ms. [/Users/runner/runners/2.166.2/work/1/s/Scaffolding.sln]
##[error].dotnet/sdk/5.0.100-preview.2.20165.1/NuGet.targets(128,5): error : Failed to retrieve information about 'Microsoft.CodeAnalysis.Common' from remote source 'https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json'.
The HTTP request to 'GET https://api.nuget.org/v3-flatcontainer/microsoft.codeanalysis.common/index.json' has timed out after 100000ms.
```
This has not yet hit our CI but is already impacting a number of PRs
Core-eng Issue https://github.com/dotnet/core-eng/issues/9634
<!-- runfo report start -->
|Definition|Build|Kind|Job Name|
|---|---|---|---|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604353](https://dev.azure.com/dnceng/public/_build/results?buildId=604353)|[PR 35021](https://github.com/dotnet/runtime/pull/35021)|Libraries Build tvOS x64 Debug|
|[aspnetcore-ci](https://dnceng.visualstudio.com/public/_build?definitionId=278)|[604340](https://dev.azure.com/dnceng/public/_build/results?buildId=604340)|[PR 20903](https://github.com/dotnet/aspnetcore/pull/20903)|Test: macOS 10.13|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build tvOS x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS arm debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|CoreCLR Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|CoreCLR Product Build OSX x64 checked|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build tvOS x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono LLVM Product Build OSX x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build OSX x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS arm release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build tvOS arm64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS arm64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS arm64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Libraries Build OSX x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono LLVM Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build tvOS arm64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604268](https://dev.azure.com/dnceng/public/_build/results?buildId=604268)|[PR 35030](https://github.com/dotnet/runtime/pull/35030)|Mono Product Build iOS x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|Libraries Build tvOS x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|CoreCLR Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|Libraries Build iOS arm64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|Libraries Build tvOS arm64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604247](https://dev.azure.com/dnceng/public/_build/results?buildId=604247)|[PR 34934](https://github.com/dotnet/runtime/pull/34934)|Libraries Build iOS x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604243](https://dev.azure.com/dnceng/public/_build/results?buildId=604243)|[PR 34933](https://github.com/dotnet/runtime/pull/34933)|Libraries Build tvOS x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604243](https://dev.azure.com/dnceng/public/_build/results?buildId=604243)|[PR 34933](https://github.com/dotnet/runtime/pull/34933)|Libraries Build iOS arm64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604243](https://dev.azure.com/dnceng/public/_build/results?buildId=604243)|[PR 34933](https://github.com/dotnet/runtime/pull/34933)|Libraries Build tvOS arm64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604243](https://dev.azure.com/dnceng/public/_build/results?buildId=604243)|[PR 34933](https://github.com/dotnet/runtime/pull/34933)|Libraries Build iOS x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604212](https://dev.azure.com/dnceng/public/_build/results?buildId=604212)|[PR 34825](https://github.com/dotnet/runtime/pull/34825)|Installer Build and Test tvOS_x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604116](https://dev.azure.com/dnceng/public/_build/results?buildId=604116)|[PR 34767](https://github.com/dotnet/runtime/pull/34767)|Libraries Build iOS arm64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[604116](https://dev.azure.com/dnceng/public/_build/results?buildId=604116)|[PR 34767](https://github.com/dotnet/runtime/pull/34767)|Libraries Build tvOS arm64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[603910](https://dev.azure.com/dnceng/public/_build/results?buildId=603910)|[PR 35010](https://github.com/dotnet/runtime/pull/35010)|Libraries Build tvOS x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[603244](https://dev.azure.com/dnceng/public/_build/results?buildId=603244)|[PR 35040](https://github.com/dotnet/runtime/pull/35040)|Libraries Test Run release mono OSX x64 Debug|
Most [recent](https://dev.azure.com/dnceng/public/_build/results?buildId=604353) failure 4/16/2020 1:42:00 PM
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
<!-- runfo report end -->
| non_priority | hitting restore timeouts with nuget hitting timeout issues with nuget org restore completed in sec for users runner runners work s src vs web cg msbuild vs web cg msbuild csproj retrying findpackagesbyidasync for source the http request to get has timed out after retrying findpackagesbyidasync for source the http request to get has timed out after users runner runners work s dotnet sdk preview nuget targets error failed to retrieve information about microsoft codeanalysis common from remote source users runner runners work s dotnet sdk preview nuget targets error the http request to get has timed out after dotnet sdk preview nuget targets error failed to retrieve information about microsoft codeanalysis common from remote source the http request to get has timed out after this has not yet hit our ci but is already impacting a number of prs core eng issue definition build kind job name build tvos debug macos product build tvos debug product build ios arm debug product build osx release product build osx release product build osx checked product build tvos release llvm product build osx debug product build osx debug product build ios arm release product build tvos release product build ios release product build ios debug build osx debug llvm product build osx release product build tvos debug product build ios release product build ios debug build tvos debug product build osx release build ios debug build tvos release build ios release build tvos debug build ios debug build tvos release build ios release build and test tvos debug build ios debug build tvos release build tvos debug test run release mono osx debug most failure pm | 0 |
224,551 | 24,781,646,492 | IssuesEvent | 2022-10-24 05:56:34 | sast-automation-dev/openmrs-core-43 | https://api.github.com/repos/sast-automation-dev/openmrs-core-43 | opened | standard-1.1.2.jar: 1 vulnerabilities (highest severity is: 7.3) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>standard-1.1.2.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /webapp/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/taglibs/standard/1.1.2/standard-1.1.2.jar,/itory/taglibs/standard/1.1.2/standard-1.1.2.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-43/commit/1cef8c0ef8d97d37160ad09ab71a696204c7c39d">1cef8c0ef8d97d37160ad09ab71a696204c7c39d</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (standard version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2015-0254](https://www.mend.io/vulnerability-database/CVE-2015-0254) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | standard-1.1.2.jar | Direct | org.apache.taglibs:taglibs-standard-impl:1.2.3 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-0254</summary>
### Vulnerable Library - <b>standard-1.1.2.jar</b></p>
<p></p>
<p>Path to dependency file: /webapp/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/taglibs/standard/1.1.2/standard-1.1.2.jar,/itory/taglibs/standard/1.1.2/standard-1.1.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **standard-1.1.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-43/commit/1cef8c0ef8d97d37160ad09ab71a696204c7c39d">1cef8c0ef8d97d37160ad09ab71a696204c7c39d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) <x:parse> or (2) <x:transform> JSTL XML tag.
<p>Publish Date: Mar 9, 2015 2:59:00 PM
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-0254>CVE-2015-0254</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tomcat.apache.org/taglibs/standard/">https://tomcat.apache.org/taglibs/standard/</a></p>
<p>Release Date: Mar 9, 2015 2:59:00 PM</p>
<p>Fix Resolution: org.apache.taglibs:taglibs-standard-impl:1.2.3</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | standard-1.1.2.jar: 1 vulnerabilities (highest severity is: 7.3) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>standard-1.1.2.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /webapp/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/taglibs/standard/1.1.2/standard-1.1.2.jar,/itory/taglibs/standard/1.1.2/standard-1.1.2.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-43/commit/1cef8c0ef8d97d37160ad09ab71a696204c7c39d">1cef8c0ef8d97d37160ad09ab71a696204c7c39d</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (standard version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2015-0254](https://www.mend.io/vulnerability-database/CVE-2015-0254) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | standard-1.1.2.jar | Direct | org.apache.taglibs:taglibs-standard-impl:1.2.3 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-0254</summary>
### Vulnerable Library - <b>standard-1.1.2.jar</b></p>
<p></p>
<p>Path to dependency file: /webapp/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/taglibs/standard/1.1.2/standard-1.1.2.jar,/itory/taglibs/standard/1.1.2/standard-1.1.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **standard-1.1.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/openmrs-core-43/commit/1cef8c0ef8d97d37160ad09ab71a696204c7c39d">1cef8c0ef8d97d37160ad09ab71a696204c7c39d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) <x:parse> or (2) <x:transform> JSTL XML tag.
<p>Publish Date: Mar 9, 2015 2:59:00 PM
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-0254>CVE-2015-0254</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tomcat.apache.org/taglibs/standard/">https://tomcat.apache.org/taglibs/standard/</a></p>
<p>Release Date: Mar 9, 2015 2:59:00 PM</p>
<p>Fix Resolution: org.apache.taglibs:taglibs-standard-impl:1.2.3</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_priority | standard jar vulnerabilities highest severity is vulnerable library standard jar path to dependency file webapp pom xml path to vulnerable library home wss scanner repository taglibs standard standard jar itory taglibs standard standard jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in standard version remediation available high standard jar direct org apache taglibs taglibs standard impl details cve vulnerable library standard jar path to dependency file webapp pom xml path to vulnerable library home wss scanner repository taglibs standard standard jar itory taglibs standard standard jar dependency hierarchy x standard jar vulnerable library found in head commit a href found in base branch master vulnerability details apache standard taglibs before allows remote attackers to execute arbitrary code or conduct external xml entity xxe attacks via a crafted xslt extension in a or jstl xml tag publish date mar pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date mar pm fix resolution org apache taglibs taglibs standard impl rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
822,212 | 30,858,147,197 | IssuesEvent | 2023-08-02 22:53:11 | Memmy-App/memmy | https://api.github.com/repos/Memmy-App/memmy | closed | [Bug]: comment jump button will always jump to first comment on first press, even after scrolling past it | bug high priority in progress | ### Check for open issues
- [X] I have verified that another issue for this is not open, or it has been closed and has not been fixed.
### Minimal reproducible example
Scroll past top comment by any length, press comment jump button
### Expected Behavior
Jump to the next comment down
### Version
0.6.0 (1)
### App Store or TestFlight?
TestFlight
### iOS Version
16.x
### Device
iPhone 13 | 1.0 | [Bug]: comment jump button will always jump to first comment on first press, even after scrolling past it - ### Check for open issues
- [X] I have verified that another issue for this is not open, or it has been closed and has not been fixed.
### Minimal reproducible example
Scroll past top comment by any length, press comment jump button
### Expected Behavior
Jump to the next comment down
### Version
0.6.0 (1)
### App Store or TestFlight?
TestFlight
### iOS Version
16.x
### Device
iPhone 13 | priority | comment jump button will always jump to first comment on first press even after scrolling past it check for open issues i have verified that another issue for this is not open or it has been closed and has not been fixed minimal reproducible example scroll past top comment by any length press comment jump button expected behavior jump to the next comment down version app store or testflight testflight ios version x device iphone | 1 |
226,669 | 17,358,651,095 | IssuesEvent | 2021-07-29 17:21:21 | rokmoln/support-firecloud | https://api.github.com/repos/rokmoln/support-firecloud | opened | bats FTW | documentation enhancement | some shell scripts could really use some testing, and BATS is the tool for the job
* https://bats-core.readthedocs.io/en/latest/index.html
* https://opensource.com/article/19/2/testing-bash-bats
* https://medium.com/@pimterry/testing-your-shell-scripts-with-bats-abfca9bdc5b9
* https://github.com/dmlond/how_to_bats/
* https://github.com/martin-schulze-vireso/bats-tutorial | 1.0 | bats FTW - some shell scripts could really use some testing, and BATS is the tool for the job
* https://bats-core.readthedocs.io/en/latest/index.html
* https://opensource.com/article/19/2/testing-bash-bats
* https://medium.com/@pimterry/testing-your-shell-scripts-with-bats-abfca9bdc5b9
* https://github.com/dmlond/how_to_bats/
* https://github.com/martin-schulze-vireso/bats-tutorial | non_priority | bats ftw some shell scripts could really use some testing and bats is the tool for the job | 0 |
58,102 | 24,330,395,640 | IssuesEvent | 2022-09-30 18:49:40 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | AWS VPC Traffic Mirroring - enable GWLB as a valid traffic mirror target | enhancement good first issue service/vpc | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
The resource aws_ec2_traffic_mirror_target currently supports network_load_balancer_arn and network_interface_id as its valid input parameters. AWS Introduced GWLB as a valid Traffic Mirror target in May 2022 and hence this new input must be enabled in this resource.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
- aws_ec2_traffic_mirror_target
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "aws_ec2_traffic_mirror_target" "gwlb" {
description = "GWLB target"
gateway_load_balancer_endpoint = aws_vpc_endpoint.example.id
}
```
### References
https://aws.amazon.com/blogs/networking-and-content-delivery/introduction-to-traffic-mirroring-to-gwlb-endpoints-as-target/
https://docs.aws.amazon.com/vpc/latest/mirroring/tm-example-glb-endpoints.html
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/
--->
* #0000
| 1.0 | AWS VPC Traffic Mirroring - enable GWLB as a valid traffic mirror target - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
The resource aws_ec2_traffic_mirror_target currently supports network_load_balancer_arn and network_interface_id as its valid input parameters. AWS Introduced GWLB as a valid Traffic Mirror target in May 2022 and hence this new input must be enabled in this resource.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
- aws_ec2_traffic_mirror_target
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "aws_ec2_traffic_mirror_target" "gwlb" {
description = "GWLB target"
gateway_load_balancer_endpoint = aws_vpc_endpoint.example.id
}
```
### References
https://aws.amazon.com/blogs/networking-and-content-delivery/introduction-to-traffic-mirroring-to-gwlb-endpoints-as-target/
https://docs.aws.amazon.com/vpc/latest/mirroring/tm-example-glb-endpoints.html
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/
--->
* #0000
| non_priority | aws vpc traffic mirroring enable gwlb as a valid traffic mirror target community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the resource aws traffic mirror target currently supports network load balancer arn and network interface id as its valid input parameters aws introduced gwlb as a valid traffic mirror target in may and hence this new input must be enabled in this resource new or affected resource s aws traffic mirror target potential terraform configuration hcl resource aws traffic mirror target gwlb description gwlb target gateway load balancer endpoint aws vpc endpoint example id references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example | 0 |
245,633 | 26,549,314,034 | IssuesEvent | 2023-01-20 05:31:06 | nidhi7598/linux-3.0.35_CVE-2022-45934 | https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2022-45934 | opened | CVE-2019-15922 (Medium) detected in linuxlinux-3.0.49 | security vulnerability | ## CVE-2019-15922 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.49</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2022-45934/commit/5e23b7f9d2dd0154edd54986754eecd5b5308571">5e23b7f9d2dd0154edd54986754eecd5b5308571</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/paride/pf.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/paride/pf.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/paride/pf.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.0.9. There is a NULL pointer dereference for a pf data structure if alloc_disk fails in drivers/block/paride/pf.c.
<p>Publish Date: 2019-09-04
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15922>CVE-2019-15922</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.0.9">https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.0.9</a></p>
<p>Release Date: 2019-10-04</p>
<p>Fix Resolution: v5.1-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-15922 (Medium) detected in linuxlinux-3.0.49 - ## CVE-2019-15922 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.49</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2022-45934/commit/5e23b7f9d2dd0154edd54986754eecd5b5308571">5e23b7f9d2dd0154edd54986754eecd5b5308571</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/paride/pf.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/paride/pf.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/paride/pf.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.0.9. There is a NULL pointer dereference for a pf data structure if alloc_disk fails in drivers/block/paride/pf.c.
<p>Publish Date: 2019-09-04
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15922>CVE-2019-15922</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.0.9">https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.0.9</a></p>
<p>Release Date: 2019-10-04</p>
<p>Fix Resolution: v5.1-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers block paride pf c drivers block paride pf c drivers block paride pf c vulnerability details an issue was discovered in the linux kernel before there is a null pointer dereference for a pf data structure if alloc disk fails in drivers block paride pf c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
808,519 | 30,085,857,026 | IssuesEvent | 2023-06-29 08:35:09 | red-hat-storage/ocs-ci | https://api.github.com/repos/red-hat-storage/ocs-ci | opened | Undo the early return in `conftest.py::nb_ensure_endpoint_count` once underlying issue is resolved | High Priority workaround MCG Squad/Red | Due to https://url.corp.redhat.com/2eba588, we had to give up on ensuring the minimum of two noobaa-endpoint pods in MCG tests via https://github.com/red-hat-storage/ocs-ci/pull/7922.
While this allows the acceptance tests to pass, we can expect this to fail more demanding tests in the CI, so we should undo the compromise as soon as the BZ is fixed. | 1.0 | Undo the early return in `conftest.py::nb_ensure_endpoint_count` once underlying issue is resolved - Due to https://url.corp.redhat.com/2eba588, we had to give up on ensuring the minimum of two noobaa-endpoint pods in MCG tests via https://github.com/red-hat-storage/ocs-ci/pull/7922.
While this allows the acceptance tests to pass, we can expect this to fail more demanding tests in the CI, so we should undo the compromise as soon as the BZ is fixed. | priority | undo the early return in conftest py nb ensure endpoint count once underlying issue is resolved due to we had to give up on ensuring the minimum of two noobaa endpoint pods in mcg tests via while this allows the acceptance tests to pass we can expect this to fail more demanding tests in the ci so we should undo the compromise as soon as the bz is fixed | 1 |
196,357 | 15,591,230,802 | IssuesEvent | 2021-03-18 10:12:53 | VoidStoneDev/GradleMinecraftPluginHelper | https://api.github.com/repos/VoidStoneDev/GradleMinecraftPluginHelper | closed | Changing Brand Name To GradleMinecraftPluginHelper | documentation | Changing Brand Name To 'GradleMinecraftPluginHelper' Because Of Support For Non-Bukkit API Plugins | 1.0 | Changing Brand Name To GradleMinecraftPluginHelper - Changing Brand Name To 'GradleMinecraftPluginHelper' Because Of Support For Non-Bukkit API Plugins | non_priority | changing brand name to gradleminecraftpluginhelper changing brand name to gradleminecraftpluginhelper because of support for non bukkit api plugins | 0 |
236,117 | 19,514,507,037 | IssuesEvent | 2021-12-29 07:48:17 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | opened | Fail to add an entity that contains a boolean/int32 property in one emulator table | 🧪 testing :gear: tables :beetle: regression | **Storage Explorer Version:** 1.23.0-dev
**Build Number:** 20211220.7
**Branch:** main
**Platform/OS:** Windows 10
**Architecture:** ia32
**How Found:** AD hoc testing
**Regression From:** Previous release (1.21.3)
## Steps to Reproduce ##
1. Install and start Azurite 3.14.0.
2. Launch Storage Explorer -> Expand 'Local & Attached -> Storage Accounts -> (Emulator - Default Ports)(Key)'.
3. Right click the 'Tables' node -> Create a table.
4. Click 'Add' -> Add one boolean property -> Click 'Insert'.
5. Check whether succeeds to add the entity.
## Expected Experience ##
Succeed to add the entity.
## Actual Experience ##
Fail to add the entity.

| 1.0 | Fail to add an entity that contains a boolean/int32 property in one emulator table - **Storage Explorer Version:** 1.23.0-dev
**Build Number:** 20211220.7
**Branch:** main
**Platform/OS:** Windows 10
**Architecture:** ia32
**How Found:** AD hoc testing
**Regression From:** Previous release (1.21.3)
## Steps to Reproduce ##
1. Install and start Azurite 3.14.0.
2. Launch Storage Explorer -> Expand 'Local & Attached -> Storage Accounts -> (Emulator - Default Ports)(Key)'.
3. Right click the 'Tables' node -> Create a table.
4. Click 'Add' -> Add one boolean property -> Click 'Insert'.
5. Check whether succeeds to add the entity.
## Expected Experience ##
Succeed to add the entity.
## Actual Experience ##
Fail to add the entity.

| non_priority | fail to add an entity that contains a boolean property in one emulator table storage explorer version dev build number branch main platform os windows architecture how found ad hoc testing regression from previous release steps to reproduce install and start azurite launch storage explorer expand local attached storage accounts emulator default ports key right click the tables node create a table click add add one boolean property click insert check whether succeeds to add the entity expected experience succeed to add the entity actual experience fail to add the entity | 0 |
428,157 | 12,403,911,563 | IssuesEvent | 2020-05-21 14:41:30 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | Put NSX info into NSX tab | Category: Sunstone Priority: Normal Status: Accepted Type: Feature | **Description**
Sunstone should show NSX info into NSX tab
**Use case**
To view more easily that information
**Interface Changes**
Sunstone
**Additional Context**
Add any other context or screenshots about the feature request here. Or any other alternative you have considered to addressed this new feature.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| 1.0 | Put NSX info into NSX tab - **Description**
Sunstone should show NSX info into NSX tab
**Use case**
To view more easily that information
**Interface Changes**
Sunstone
**Additional Context**
Add any other context or screenshots about the feature request here. Or any other alternative you have considered to addressed this new feature.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| priority | put nsx info into nsx tab description sunstone should show nsx info into nsx tab use case to view more easily that information interface changes sunstone additional context add any other context or screenshots about the feature request here or any other alternative you have considered to addressed this new feature progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches | 1 |
106,450 | 4,272,690,282 | IssuesEvent | 2016-07-13 15:11:17 | fgpv-vpgf/rcs | https://api.github.com/repos/fgpv-vpgf/rcs | opened | Permit recursive flag on Esri Map Services | improvements: enhancement needs: estimate priority: high | #45 was targeting only Group Layers, modify RCS to also handle Map Service listings of layers. This will involve some complexity as separate calls will be needed to pull out the nested layers beneath Group Layers.
Sample registration:
```
{
"service_url": "http://geoappext.nrcan.gc.ca/arcgis/rest/services/NorthAmerica/energy_infrastructure_of_north_america_en/MapServer",
"recursive": true,
"service_type": "esriMapServer",
"service_name": "Energy Infrastructure of North America"
},
"fr": {
"service_url": "http://geoappext.nrcan.gc.ca/arcgis/rest/services/NorthAmerica/energy_infrastructure_of_north_america_fr/MapServer",
"recursive": true,
"service_type": "esriMapServer",
"service_name": "Infrastructure énergétique d'Amérique du Nord"
},
"version": "2.0"
}
``` | 1.0 | Permit recursive flag on Esri Map Services - #45 was targeting only Group Layers, modify RCS to also handle Map Service listings of layers. This will involve some complexity as separate calls will be needed to pull out the nested layers beneath Group Layers.
Sample registration:
```
{
"service_url": "http://geoappext.nrcan.gc.ca/arcgis/rest/services/NorthAmerica/energy_infrastructure_of_north_america_en/MapServer",
"recursive": true,
"service_type": "esriMapServer",
"service_name": "Energy Infrastructure of North America"
},
"fr": {
"service_url": "http://geoappext.nrcan.gc.ca/arcgis/rest/services/NorthAmerica/energy_infrastructure_of_north_america_fr/MapServer",
"recursive": true,
"service_type": "esriMapServer",
"service_name": "Infrastructure énergétique d'Amérique du Nord"
},
"version": "2.0"
}
``` | priority | permit recursive flag on esri map services was targeting only group layers modify rcs to also handle map service listings of layers this will involve some complexity as separate calls will be needed to pull out the nested layers beneath group layers sample registration service url recursive true service type esrimapserver service name energy infrastructure of north america fr service url recursive true service type esrimapserver service name infrastructure énergétique d amérique du nord version | 1 |
32,349 | 7,529,463,182 | IssuesEvent | 2018-04-14 05:11:02 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | Request for standard form field type for all media in library | J3 Issue No Code Attached Yet | Please could we have a Standard Form Field Type which provides modal access to all the files in the media library.
This would be like the current "media" form field type, but wouldn't be restricted to just images.
The new field type should allow filtering, so that you could specify the types of files (eg audio / video) and file extensions (eg zip, gpx) which you would want presented in the modal window.
A personal suggestion would be to extend the current media file type to include 2 other parameters which would control the filtering of what to display:
* an array of something like mime types (eg "audio/*" for presenting audio files)
* an array of file extensions
If these parameters weren't set then the functionality could revert to what is presented at the moment, which is like a mime type of image/*
| 1.0 | Request for standard form field type for all media in library - Please could we have a Standard Form Field Type which provides modal access to all the files in the media library.
This would be like the current "media" form field type, but wouldn't be restricted to just images.
The new field type should allow filtering, so that you could specify the types of files (eg audio / video) and file extensions (eg zip, gpx) which you would want presented in the modal window.
A personal suggestion would be to extend the current media file type to include 2 other parameters which would control the filtering of what to display:
* an array of something like mime types (eg "audio/*" for presenting audio files)
* an array of file extensions
If these parameters weren't set then the functionality could revert to what is presented at the moment, which is like a mime type of image/*
| non_priority | request for standard form field type for all media in library please could we have a standard form field type which provides modal access to all the files in the media library this would be like the current media form field type but wouldn t be restricted to just images the new field type should allow filtering so that you could specify the types of files eg audio video and file extensions eg zip gpx which you would want presented in the modal window a personal suggestion would be to extend the current media file type to include other parameters which would control the filtering of what to display an array of something like mime types eg audio for presenting audio files an array of file extensions if these parameters weren t set then the functionality could revert to what is presented at the moment which is like a mime type of image | 0 |
302,340 | 26,139,861,719 | IssuesEvent | 2022-12-29 16:49:52 | apache/beam | https://api.github.com/repos/apache/beam | closed | Deploy PerfKit Explorer for Beam | tests P3 bug |
Imported from Jira [BEAM-1596](https://issues.apache.org/jira/browse/BEAM-1596). Original Jira may contain additional context.
Reported by: jaku. | 1.0 | Deploy PerfKit Explorer for Beam -
Imported from Jira [BEAM-1596](https://issues.apache.org/jira/browse/BEAM-1596). Original Jira may contain additional context.
Reported by: jaku. | non_priority | deploy perfkit explorer for beam imported from jira original jira may contain additional context reported by jaku | 0 |
328,605 | 28,127,430,457 | IssuesEvent | 2023-03-31 19:03:23 | flutter/flutter | https://api.github.com/repos/flutter/flutter | reopened | integration_test doesn't measure CPU, memory, or GPU utilization | integration_test P4 | This makes the numbers fairly limited compared to what driver can do. | 1.0 | integration_test doesn't measure CPU, memory, or GPU utilization - This makes the numbers fairly limited compared to what driver can do. | non_priority | integration test doesn t measure cpu memory or gpu utilization this makes the numbers fairly limited compared to what driver can do | 0 |
53,580 | 13,850,391,324 | IssuesEvent | 2020-10-15 01:05:16 | kenferrara/highcharts | https://api.github.com/repos/kenferrara/highcharts | opened | CVE-2020-11979 (High) detected in ant-1.8.2.jar | security vulnerability | ## CVE-2020-11979 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.8.2.jar</b></p></summary>
<p>master POM</p>
<p>Path to vulnerable library: highcharts/tools/apache-ant/lib/ant.jar</p>
<p>
Dependency Hierarchy:
- :x: **ant-1.8.2.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
As mitigation for CVE-2020-1945 Apache Ant 1.10.8 changed the permissions of temporary files it created so that only the current user was allowed to access them. Unfortunately the fixcrlf task deleted the temporary file and created a new one without said protection, effectively nullifying the effort. This would still allow an attacker to inject modified source files into the build process.
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11979>CVE-2020-11979</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://ant.apache.org/security.html">https://ant.apache.org/security.html</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.ant:ant:1.10.9 </p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.ant","packageName":"ant","packageVersion":"1.8.2","isTransitiveDependency":false,"dependencyTree":"org.apache.ant:ant:1.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.ant:ant:1.10.9 "}],"vulnerabilityIdentifier":"CVE-2020-11979","vulnerabilityDetails":"As mitigation for CVE-2020-1945 Apache Ant 1.10.8 changed the permissions of temporary files it created so that only the current user was allowed to access them. Unfortunately the fixcrlf task deleted the temporary file and created a new one without said protection, effectively nullifying the effort. This would still allow an attacker to inject modified source files into the build process.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11979","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-11979 (High) detected in ant-1.8.2.jar - ## CVE-2020-11979 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.8.2.jar</b></p></summary>
<p>master POM</p>
<p>Path to vulnerable library: highcharts/tools/apache-ant/lib/ant.jar</p>
<p>
Dependency Hierarchy:
- :x: **ant-1.8.2.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
As mitigation for CVE-2020-1945 Apache Ant 1.10.8 changed the permissions of temporary files it created so that only the current user was allowed to access them. Unfortunately the fixcrlf task deleted the temporary file and created a new one without said protection, effectively nullifying the effort. This would still allow an attacker to inject modified source files into the build process.
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11979>CVE-2020-11979</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://ant.apache.org/security.html">https://ant.apache.org/security.html</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.ant:ant:1.10.9 </p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.ant","packageName":"ant","packageVersion":"1.8.2","isTransitiveDependency":false,"dependencyTree":"org.apache.ant:ant:1.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.ant:ant:1.10.9 "}],"vulnerabilityIdentifier":"CVE-2020-11979","vulnerabilityDetails":"As mitigation for CVE-2020-1945 Apache Ant 1.10.8 changed the permissions of temporary files it created so that only the current user was allowed to access them. Unfortunately the fixcrlf task deleted the temporary file and created a new one without said protection, effectively nullifying the effort. This would still allow an attacker to inject modified source files into the build process.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11979","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in ant jar cve high severity vulnerability vulnerable library ant jar master pom path to vulnerable library highcharts tools apache ant lib ant jar dependency hierarchy x ant jar vulnerable library vulnerability details as mitigation for cve apache ant changed the permissions of temporary files it created so that only the current user was allowed to access them unfortunately the fixcrlf task deleted the temporary file and created a new one without said protection effectively nullifying the effort this would still allow an attacker to inject modified source files into the build process publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache ant ant rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails as mitigation for cve apache ant changed the permissions of temporary files it created so that only the current user was allowed to access them unfortunately the fixcrlf task deleted the temporary file and created a new one without said protection effectively nullifying the effort this would still allow an attacker to inject modified source files into the build process vulnerabilityurl | 0 |
431,685 | 12,484,735,596 | IssuesEvent | 2020-05-30 16:11:11 | spectrochempy/test_issues_migration_from_redmine | https://api.github.com/repos/spectrochempy/test_issues_migration_from_redmine | closed | Check for new version at the program start up | Category: Deployment Priority: Minor Project: SpectroChemPy Status: Closed Tracker: Feature | Author: Christian Fernandez (Christian Fernandez )
Redmine Issue: 29, https://redmine.spectrochempy.fr/issues/29
---
Make a function call when the library is loaded to indicate whether or not a new version is present o PyPi and Conda.
| 1.0 | Check for new version at the program start up - Author: Christian Fernandez (Christian Fernandez )
Redmine Issue: 29, https://redmine.spectrochempy.fr/issues/29
---
Make a function call when the library is loaded to indicate whether or not a new version is present o PyPi and Conda.
| priority | check for new version at the program start up author christian fernandez christian fernandez redmine issue make a function call when the library is loaded to indicate whether or not a new version is present o pypi and conda | 1 |
742,340 | 25,849,995,158 | IssuesEvent | 2022-12-13 09:44:08 | l7mp/stunner | https://api.github.com/repos/l7mp/stunner | closed | stunnerd stops responding to filesystem events | priority: low status: cannot reproduce type: bug | The `stunnerd` pod seems to hang, with the following logs:
```
14:09:00.748041 main.go:148: stunnerd WARNING: unhnadled notify op on config file "/etc/stunnerd/stunnerd.conf" (ignoring): CHMOD
14:09:00.748077 main.go:133: stunnerd WARNING: config file deleted "REMOVE", disabling watcher
14:09:00.748090 main.go:138: stunnerd WARNING: could not remove config file "/etc/stunnerd/stunnerd.conf" from watcher: can't remove non-existent inotify watch for: /etc/stunnerd/stunnerd.conf
14:09:01.400101 reconcile.go:145: stunner INFO: reconciliation ready: new objects: 0, changed objects: 3, deleted objects: 0
```
After this point, config file updates are not picked up by `stunnerd` any more | 1.0 | stunnerd stops responding to filesystem events - The `stunnerd` pod seems to hang, with the following logs:
```
14:09:00.748041 main.go:148: stunnerd WARNING: unhnadled notify op on config file "/etc/stunnerd/stunnerd.conf" (ignoring): CHMOD
14:09:00.748077 main.go:133: stunnerd WARNING: config file deleted "REMOVE", disabling watcher
14:09:00.748090 main.go:138: stunnerd WARNING: could not remove config file "/etc/stunnerd/stunnerd.conf" from watcher: can't remove non-existent inotify watch for: /etc/stunnerd/stunnerd.conf
14:09:01.400101 reconcile.go:145: stunner INFO: reconciliation ready: new objects: 0, changed objects: 3, deleted objects: 0
```
After this point, config file updates are not picked up by `stunnerd` any more | priority | stunnerd stops responding to filesystem events the stunnerd pod seems to hang with the following logs main go stunnerd warning unhnadled notify op on config file etc stunnerd stunnerd conf ignoring chmod main go stunnerd warning config file deleted remove disabling watcher main go stunnerd warning could not remove config file etc stunnerd stunnerd conf from watcher can t remove non existent inotify watch for etc stunnerd stunnerd conf reconcile go stunner info reconciliation ready new objects changed objects deleted objects after this point config file updates are not picked up by stunnerd any more | 1 |
337,905 | 30,271,068,967 | IssuesEvent | 2023-07-07 15:23:08 | BoBAdministration/QA-Bug-Reports | https://api.github.com/repos/BoBAdministration/QA-Bug-Reports | closed | World border kill zone jutting out into area within the border | Fixed-PendingTesting | **Describe the Bug**
In some spots, the kill zone that is supposed to be outside the world border is inside the world border, making it kill players who arent doing anything wrong. this seems to exclusively be around corners that jut into the playable area
**To Reproduce**
1. log into any titania server
2. teleport or otherwise travel to (627865 726500 -13130) or (150637 339014 -9469)
3. touch the area around the corner
**Expected behavior**
you get stopped by the border
**Actual behavior**
you get killed by the border
**Screenshots & Video**
https://www.youtube.com/watch?v=APMNkwagma8
**Branch Version**
Live
**Character Information**
Any dino
**Additional Information**
Titania
Map bug
| 1.0 | World border kill zone jutting out into area within the border - **Describe the Bug**
In some spots, the kill zone that is supposed to be outside the world border is inside the world border, making it kill players who arent doing anything wrong. this seems to exclusively be around corners that jut into the playable area
**To Reproduce**
1. log into any titania server
2. teleport or otherwise travel to (627865 726500 -13130) or (150637 339014 -9469)
3. touch the area around the corner
**Expected behavior**
you get stopped by the border
**Actual behavior**
you get killed by the border
**Screenshots & Video**
https://www.youtube.com/watch?v=APMNkwagma8
**Branch Version**
Live
**Character Information**
Any dino
**Additional Information**
Titania
Map bug
| non_priority | world border kill zone jutting out into area within the border describe the bug in some spots the kill zone that is supposed to be outside the world border is inside the world border making it kill players who arent doing anything wrong this seems to exclusively be around corners that jut into the playable area to reproduce log into any titania server teleport or otherwise travel to or touch the area around the corner expected behavior you get stopped by the border actual behavior you get killed by the border screenshots video branch version live character information any dino additional information titania map bug | 0 |
2,142 | 4,272,560,830 | IssuesEvent | 2016-07-13 14:50:46 | IBM-Bluemix/logistics-wizard | https://api.github.com/repos/IBM-Bluemix/logistics-wizard | opened | Use transaction in ERP simulator to avoid corruption of data in concurrent API calls | erp-service story | Aim of this story is to review the ERP simulator and use Loopback transactions whenever needed.
Typically in demo.js, shipment.js, some remote methods alter multiple tables and objects at the same time. These methods should be protected against concurrent API calls that could potentially corrupt the data - as example executing two different shipment approvals at the same time could corrupt the inventory data today. | 1.0 | Use transaction in ERP simulator to avoid corruption of data in concurrent API calls - Aim of this story is to review the ERP simulator and use Loopback transactions whenever needed.
Typically in demo.js, shipment.js, some remote methods alter multiple tables and objects at the same time. These methods should be protected against concurrent API calls that could potentially corrupt the data - as example executing two different shipment approvals at the same time could corrupt the inventory data today. | non_priority | use transaction in erp simulator to avoid corruption of data in concurrent api calls aim of this story is to review the erp simulator and use loopback transactions whenever needed typically in demo js shipment js some remote methods alter multiple tables and objects at the same time these methods should be protected against concurrent api calls that could potentially corrupt the data as example executing two different shipment approvals at the same time could corrupt the inventory data today | 0 |
43,735 | 2,891,935,225 | IssuesEvent | 2015-06-15 09:37:44 | geometalab/osmaxx | https://api.github.com/repos/geometalab/osmaxx | opened | Production and development suitable docker-compose.yml | priority:high | As a developer, I want to deploy the docker containers as simple or simpler than on the local machine.
Write a simple docker-compose file, that can be adapted/copied if used in production or in development, while still having both environments stick very close to each other. | 1.0 | Production and development suitable docker-compose.yml - As a developer, I want to deploy the docker containers as simple or simpler than on the local machine.
Write a simple docker-compose file, that can be adapted/copied if used in production or in development, while still having both environments stick very close to each other. | priority | production and development suitable docker compose yml as a developer i want to deploy the docker containers as simple or simpler than on the local machine write a simple docker compose file that can be adapted copied if used in production or in development while still having both environments stick very close to each other | 1 |
61,972 | 25,813,186,113 | IssuesEvent | 2022-12-12 01:28:59 | fga-eps-mds/2022-2-MeasureSoftGram-Doc | https://api.github.com/repos/fga-eps-mds/2022-2-MeasureSoftGram-Doc | opened | US10: Visualizar `heatmap` de correlação entre as releases | Frontend Service US | ## Descrição da Issue
<!-- Descreva de forma sucinta a issue e caso necessite, as informações adicionais necessárias para sua realização. -->
Eu como **Pesquisador de Software/QA/Tech lead** gostaria de observar um **heatmap** com a correlação entre as releases.
## Tarefas:
<!-- - [ ] Tarefa 1
- [ ] Tarefa 2
- [ ] Tarefa 3 -->
## Critérios de Aceitação:
<!-- - [ ] Critério 1
- [ ] Critério 2
- [ ] Critério 3 --> | 1.0 | US10: Visualizar `heatmap` de correlação entre as releases - ## Descrição da Issue
<!-- Descreva de forma sucinta a issue e caso necessite, as informações adicionais necessárias para sua realização. -->
Eu como **Pesquisador de Software/QA/Tech lead** gostaria de observar um **heatmap** com a correlação entre as releases.
## Tarefas:
<!-- - [ ] Tarefa 1
- [ ] Tarefa 2
- [ ] Tarefa 3 -->
## Critérios de Aceitação:
<!-- - [ ] Critério 1
- [ ] Critério 2
- [ ] Critério 3 --> | non_priority | visualizar heatmap de correlação entre as releases descrição da issue eu como pesquisador de software qa tech lead gostaria de observar um heatmap com a correlação entre as releases tarefas tarefa tarefa tarefa critérios de aceitação critério critério critério | 0 |
80,118 | 3,550,920,478 | IssuesEvent | 2016-01-21 00:11:20 | washingtontrails/vms | https://api.github.com/repos/washingtontrails/vms | closed | Request higher Salesforce API call quota | High Priority Salesforce VMS BUDGET | @WaBirder57, we had discussed requesting a temporary increase in Salesforce's limit on API calls for the first few weeks that the live copy of VMS is running, until we have a better idea of what typical usage looks like and can figure out if we need to pay for more calls permanently or optimize the system.
Can you make this request to your account rep?
Your current limit is 36,000 per 24 hours (1000 per user license). The live instance of VMS is currently using about 12,000 calls per day for indexing but user interaction will use more and we don't know how heavily the system will be used. I'd suggest requesting an increase to 100k/day through the date that signup for volunteer vacations begins.
/cc @mscholtz | 1.0 | Request higher Salesforce API call quota - @WaBirder57, we had discussed requesting a temporary increase in Salesforce's limit on API calls for the first few weeks that the live copy of VMS is running, until we have a better idea of what typical usage looks like and can figure out if we need to pay for more calls permanently or optimize the system.
Can you make this request to your account rep?
Your current limit is 36,000 per 24 hours (1000 per user license). The live instance of VMS is currently using about 12,000 calls per day for indexing but user interaction will use more and we don't know how heavily the system will be used. I'd suggest requesting an increase to 100k/day through the date that signup for volunteer vacations begins.
/cc @mscholtz | priority | request higher salesforce api call quota we had discussed requesting a temporary increase in salesforce s limit on api calls for the first few weeks that the live copy of vms is running until we have a better idea of what typical usage looks like and can figure out if we need to pay for more calls permanently or optimize the system can you make this request to your account rep your current limit is per hours per user license the live instance of vms is currently using about calls per day for indexing but user interaction will use more and we don t know how heavily the system will be used i d suggest requesting an increase to day through the date that signup for volunteer vacations begins cc mscholtz | 1 |
589,511 | 17,703,531,333 | IssuesEvent | 2021-08-25 03:13:19 | woowa-techcamp-2021/store-6 | https://api.github.com/repos/woowa-techcamp-2021/store-6 | closed | [FE] useDebounce 개선 | refactor high priority | ## :hammer: 기능 설명
### useDebounce 개선
## 📑 완료 조건
- [x] 인자에 값이 아닌 콜백을 받도록 수정한다.
## :thought_balloon: 관련 Backlog
> [대분류] - [중분류] - [Backlog 이름]
리팩토링 - useDebounce | 1.0 | [FE] useDebounce 개선 - ## :hammer: 기능 설명
### useDebounce 개선
## 📑 완료 조건
- [x] 인자에 값이 아닌 콜백을 받도록 수정한다.
## :thought_balloon: 관련 Backlog
> [대분류] - [중분류] - [Backlog 이름]
리팩토링 - useDebounce | priority | usedebounce 개선 hammer 기능 설명 usedebounce 개선 📑 완료 조건 인자에 값이 아닌 콜백을 받도록 수정한다 thought balloon 관련 backlog 리팩토링 usedebounce | 1 |
44,003 | 9,530,755,758 | IssuesEvent | 2019-04-29 14:33:26 | jwilm/alacritty | https://api.github.com/repos/jwilm/alacritty | closed | Chinese character selection width | S - selection S - unicode help wanted | A single Chinese character requires two default widths to be fully selected. Otherwise only half can be selected. Just like this:

Running in Archlinux, Xorg environment. Alacritty version is `0.2.9`
| 1.0 | Chinese character selection width - A single Chinese character requires two default widths to be fully selected. Otherwise only half can be selected. Just like this:

Running in Archlinux, Xorg environment. Alacritty version is `0.2.9`
| non_priority | chinese character selection width a single chinese character requires two default widths to be fully selected otherwise only half can be selected just like this running in archlinux xorg environment alacritty version is | 0 |
55,407 | 14,008,903,983 | IssuesEvent | 2020-10-29 00:55:36 | mwilliams7197/ksa | https://api.github.com/repos/mwilliams7197/ksa | closed | WS-2018-0021 (Medium) detected in bootstrap-2.1.0.js - autoclosed | security vulnerability | ## WS-2018-0021 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-2.1.0.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js</a></p>
<p>Path to vulnerable library: /ksa/ksa-web-root/ksa-web/target/ROOT/rs/bootstrap/js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-2.1.0.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/ksa/commits/e0a0b31a5caa386ebc07377c9a7403d049ee4e94">e0a0b31a5caa386ebc07377c9a7403d049ee4e94</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XSS in data-target in bootstrap (3.3.7 and before)
<p>Publish Date: 2017-06-27
<p>URL: <a href=https://github.com/twbs/bootstrap/issues/20184>WS-2018-0021</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/commit/d9be1da55bf0f94a81e8a2c9acf5574fb801306e">https://github.com/twbs/bootstrap/commit/d9be1da55bf0f94a81e8a2c9acf5574fb801306e</a></p>
<p>Release Date: 2017-08-25</p>
<p>Fix Resolution: Replace or update the following files: alert.js, carousel.js, collapse.js, dropdown.js, modal.js</p>
</p>
</details>
<p></p>
| True | WS-2018-0021 (Medium) detected in bootstrap-2.1.0.js - autoclosed - ## WS-2018-0021 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-2.1.0.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js</a></p>
<p>Path to vulnerable library: /ksa/ksa-web-root/ksa-web/target/ROOT/rs/bootstrap/js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-2.1.0.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/ksa/commits/e0a0b31a5caa386ebc07377c9a7403d049ee4e94">e0a0b31a5caa386ebc07377c9a7403d049ee4e94</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XSS in data-target in bootstrap (3.3.7 and before)
<p>Publish Date: 2017-06-27
<p>URL: <a href=https://github.com/twbs/bootstrap/issues/20184>WS-2018-0021</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/commit/d9be1da55bf0f94a81e8a2c9acf5574fb801306e">https://github.com/twbs/bootstrap/commit/d9be1da55bf0f94a81e8a2c9acf5574fb801306e</a></p>
<p>Release Date: 2017-08-25</p>
<p>Fix Resolution: Replace or update the following files: alert.js, carousel.js, collapse.js, dropdown.js, modal.js</p>
</p>
</details>
<p></p>
| non_priority | ws medium detected in bootstrap js autoclosed ws medium severity vulnerability vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library ksa ksa web root ksa web target root rs bootstrap js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href vulnerability details xss in data target in bootstrap and before publish date url a href cvss score details base score metrics not available suggested fix type change files origin a href release date fix resolution replace or update the following files alert js carousel js collapse js dropdown js modal js | 0 |
55,406 | 14,008,903,714 | IssuesEvent | 2020-10-29 00:55:33 | mwilliams7197/ksa | https://api.github.com/repos/mwilliams7197/ksa | closed | WS-2017-0330 (Medium) detected in mime-1.2.4.tgz - autoclosed | security vulnerability | ## WS-2017-0330 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.2.4.tgz</b></p></summary>
<p>A comprehensive library for mime-type mapping</p>
<p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.2.4.tgz">https://registry.npmjs.org/mime/-/mime-1.2.4.tgz</a></p>
<p>Path to dependency file: /ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json</p>
<p>Path to vulnerable library: /tmp/git/ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/node_modules/mime/package.json</p>
<p>
Dependency Hierarchy:
- connect-2.1.3.tgz (Root Library)
- :x: **mime-1.2.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/ksa/commits/e0a0b31a5caa386ebc07377c9a7403d049ee4e94">e0a0b31a5caa386ebc07377c9a7403d049ee4e94</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Affected version of mime (1.0.0 throw 1.4.0 and 2.0.0 throw 2.0.2), are vulnerable to regular expression denial of service.
<p>Publish Date: 2017-09-27
<p>URL: <a href=https://github.com/broofa/node-mime/commit/1df903fdeb9ae7eaa048795b8d580ce2c98f40b0>WS-2017-0330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/broofa/node-mime/commit/1df903fdeb9ae7eaa048795b8d580ce2c98f40b0">https://github.com/broofa/node-mime/commit/1df903fdeb9ae7eaa048795b8d580ce2c98f40b0</a></p>
<p>Release Date: 2019-04-03</p>
<p>Fix Resolution: 1.4.1,2.0.3</p>
</p>
</details>
<p></p>
| True | WS-2017-0330 (Medium) detected in mime-1.2.4.tgz - autoclosed - ## WS-2017-0330 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.2.4.tgz</b></p></summary>
<p>A comprehensive library for mime-type mapping</p>
<p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.2.4.tgz">https://registry.npmjs.org/mime/-/mime-1.2.4.tgz</a></p>
<p>Path to dependency file: /ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json</p>
<p>Path to vulnerable library: /tmp/git/ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/node_modules/mime/package.json</p>
<p>
Dependency Hierarchy:
- connect-2.1.3.tgz (Root Library)
- :x: **mime-1.2.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/ksa/commits/e0a0b31a5caa386ebc07377c9a7403d049ee4e94">e0a0b31a5caa386ebc07377c9a7403d049ee4e94</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Affected version of mime (1.0.0 throw 1.4.0 and 2.0.0 throw 2.0.2), are vulnerable to regular expression denial of service.
<p>Publish Date: 2017-09-27
<p>URL: <a href=https://github.com/broofa/node-mime/commit/1df903fdeb9ae7eaa048795b8d580ce2c98f40b0>WS-2017-0330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/broofa/node-mime/commit/1df903fdeb9ae7eaa048795b8d580ce2c98f40b0">https://github.com/broofa/node-mime/commit/1df903fdeb9ae7eaa048795b8d580ce2c98f40b0</a></p>
<p>Release Date: 2019-04-03</p>
<p>Fix Resolution: 1.4.1,2.0.3</p>
</p>
</details>
<p></p>
| non_priority | ws medium detected in mime tgz autoclosed ws medium severity vulnerability vulnerable library mime tgz a comprehensive library for mime type mapping library home page a href path to dependency file ksa ksa web root ksa web src main webapp rs bootstrap package json path to vulnerable library tmp git ksa ksa web root ksa web src main webapp rs bootstrap node modules mime package json dependency hierarchy connect tgz root library x mime tgz vulnerable library found in head commit a href vulnerability details affected version of mime throw and throw are vulnerable to regular expression denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution | 0 |
520,722 | 15,091,782,005 | IssuesEvent | 2021-02-06 16:53:42 | ridhambhat/SlateVim | https://api.github.com/repos/ridhambhat/SlateVim | opened | Add React Router routing for group editing | priority.Medium type.Enhancement | Use `React Router` and have a route to something like `groups/:id` to allow for different editing "rooms" | 1.0 | Add React Router routing for group editing - Use `React Router` and have a route to something like `groups/:id` to allow for different editing "rooms" | priority | add react router routing for group editing use react router and have a route to something like groups id to allow for different editing rooms | 1 |
162,317 | 6,150,799,408 | IssuesEvent | 2017-06-27 23:50:02 | chaos/pdsh | https://api.github.com/repos/chaos/pdsh | closed | Support user defined variables in pdsh | auto-migrated Priority-Low Type-Enhancement | ```
Hi!
It would be nice having user defined variables per host in pdsh.
This could be the first "execute different command by host (I see a wishlist
bug here talking about)" implementation:
Like %h is hostname.
We could define:
$ cat ~/.pdsh/vars/host01:
ports="8080 8081"
$ cat ~/.pdsh/vars/host02:
ports="8080"
Then:
pdsh -w ^filewithhosts "
for port in %ports
do
wget -O - --user user --password pass localhost:\$port/manager/list
done
"
Do you think it could be possible in the future ?
Thank you developping this awesome tool !
```
Original issue reported on code.google.com by `javibarr...@gmail.com` on 9 Sep 2011 at 6:32
| 1.0 | Support user defined variables in pdsh - ```
Hi!
It would be nice having user defined variables per host in pdsh.
This could be the first "execute different command by host (I see a wishlist
bug here talking about)" implementation:
Like %h is hostname.
We could define:
$ cat ~/.pdsh/vars/host01:
ports="8080 8081"
$ cat ~/.pdsh/vars/host02:
ports="8080"
Then:
pdsh -w ^filewithhosts "
for port in %ports
do
wget -O - --user user --password pass localhost:\$port/manager/list
done
"
Do you think it could be possible in the future ?
Thank you developping this awesome tool !
```
Original issue reported on code.google.com by `javibarr...@gmail.com` on 9 Sep 2011 at 6:32
| priority | support user defined variables in pdsh hi it would be nice having user defined variables per host in pdsh this could be the first execute different command by host i see a wishlist bug here talking about implementation like h is hostname we could define cat pdsh vars ports cat pdsh vars ports then pdsh w filewithhosts for port in ports do wget o user user password pass localhost port manager list done do you think it could be possible in the future thank you developping this awesome tool original issue reported on code google com by javibarr gmail com on sep at | 1 |
186,843 | 6,743,025,513 | IssuesEvent | 2017-10-20 10:11:20 | CS2103AUG2017-W09-B3/main | https://api.github.com/repos/CS2103AUG2017-W09-B3/main | closed | As a forgetful user, I want to find specific contacts by attributes other than name | priority.high status.completed type.enhancement | ...so that I can find out the contact’s name and other details.
- Command class implementation
- Update Parser
- JUnit test | 1.0 | As a forgetful user, I want to find specific contacts by attributes other than name - ...so that I can find out the contact’s name and other details.
- Command class implementation
- Update Parser
- JUnit test | priority | as a forgetful user i want to find specific contacts by attributes other than name so that i can find out the contact’s name and other details command class implementation update parser junit test | 1 |
15,539 | 8,954,696,862 | IssuesEvent | 2019-01-26 00:07:42 | vim/vim | https://api.github.com/repos/vim/vim | closed | `rop=type:directx` in combination with `encoding=utf-8` slows down global command | patch performance platform-windows | Os: Windows 10 x64 (v1709)
Vim version:
```
VIM - Vi IMproved 8.0 (2016 Sep 12, compiled Mar 11 2018 23:08:09)
MS-Windows 32-bit GUI version with OLE support
Included patches: 1-1599
Compiled by appveyor@APPVYR-WIN
```
Steps to reproduce:
`gvim --clean +"set rop=type:directx encoding=utf-8" +"norm Itest" +"norm yy10000p" +"g/./norm Itest"`
7.5 s
`gvim --clean +"set encoding=utf-8" +"norm Itest" +"norm yy10000p" +"g/./norm Itest"`
0.7 s
:global command is ten times slower when directx is enabled.
| True | `rop=type:directx` in combination with `encoding=utf-8` slows down global command - Os: Windows 10 x64 (v1709)
Vim version:
```
VIM - Vi IMproved 8.0 (2016 Sep 12, compiled Mar 11 2018 23:08:09)
MS-Windows 32-bit GUI version with OLE support
Included patches: 1-1599
Compiled by appveyor@APPVYR-WIN
```
Steps to reproduce:
`gvim --clean +"set rop=type:directx encoding=utf-8" +"norm Itest" +"norm yy10000p" +"g/./norm Itest"`
7.5 s
`gvim --clean +"set encoding=utf-8" +"norm Itest" +"norm yy10000p" +"g/./norm Itest"`
0.7 s
:global command is ten times slower when directx is enabled.
| non_priority | rop type directx in combination with encoding utf slows down global command os windows vim version vim vi improved sep compiled mar ms windows bit gui version with ole support included patches compiled by appveyor appvyr win steps to reproduce gvim clean set rop type directx encoding utf norm itest norm g norm itest s gvim clean set encoding utf norm itest norm g norm itest s global command is ten times slower when directx is enabled | 0 |
207,603 | 7,131,595,064 | IssuesEvent | 2018-01-22 11:35:44 | GreenDelta/Sophena | https://api.github.com/repos/GreenDelta/Sophena | opened | Einführung einer Brennstoff-Gruppe und Zuordnung der Wärmeerzeuger dazu | enhancement medium priority | Bestehende Wärmeerzeuger können erfasste Brennstoffe nicht verwenden. | 1.0 | Einführung einer Brennstoff-Gruppe und Zuordnung der Wärmeerzeuger dazu - Bestehende Wärmeerzeuger können erfasste Brennstoffe nicht verwenden. | priority | einführung einer brennstoff gruppe und zuordnung der wärmeerzeuger dazu bestehende wärmeerzeuger können erfasste brennstoffe nicht verwenden | 1 |
173,272 | 6,522,819,979 | IssuesEvent | 2017-08-29 05:28:19 | Templarian/MaterialDesign | https://api.github.com/repos/Templarian/MaterialDesign | closed | Floorplan / blueprint icon | Contribution High Priority Home Assistant Icon Request | As a home assistant home automation user I found the need for a floorplan icon, but there is nothing suitable in MDI.
Here's one I rolled. It would be nice if it can be added to MDI.
[floorplan.zip](https://github.com/Templarian/MaterialDesign/files/1213052/floorplan.zip)
| 1.0 | Floorplan / blueprint icon - As a home assistant home automation user I found the need for a floorplan icon, but there is nothing suitable in MDI.
Here's one I rolled. It would be nice if it can be added to MDI.
[floorplan.zip](https://github.com/Templarian/MaterialDesign/files/1213052/floorplan.zip)
| priority | floorplan blueprint icon as a home assistant home automation user i found the need for a floorplan icon but there is nothing suitable in mdi here s one i rolled it would be nice if it can be added to mdi | 1 |
161,536 | 25,358,047,356 | IssuesEvent | 2022-11-20 15:21:10 | dotnet/fsharp | https://api.github.com/repos/dotnet/fsharp | closed | FSharpChecker.CompileToDynamicAssembly | Resolution-By Design Needs-Triage | i used this function to make in-memory dynamic assembiles from fsx scripts in 41.0.7
Does this function just no longer exist?
Is there a replacement?
I cant find a single thing about it in the notes | 1.0 | FSharpChecker.CompileToDynamicAssembly - i used this function to make in-memory dynamic assembiles from fsx scripts in 41.0.7
Does this function just no longer exist?
Is there a replacement?
I cant find a single thing about it in the notes | non_priority | fsharpchecker compiletodynamicassembly i used this function to make in memory dynamic assembiles from fsx scripts in does this function just no longer exist is there a replacement i cant find a single thing about it in the notes | 0 |
375,553 | 26,167,915,526 | IssuesEvent | 2023-01-01 14:28:06 | appium/appium | https://api.github.com/repos/appium/appium | closed | 'css selector' is not supported for this session | Documentation Help Needed | I am following the code from http://appium.io/docs/en/commands/element/actions/click/
`$('#SomeId').click();`
"webdriverio": "^5.18.5"
| 1.0 | 'css selector' is not supported for this session - I am following the code from http://appium.io/docs/en/commands/element/actions/click/
`$('#SomeId').click();`
"webdriverio": "^5.18.5"
| non_priority | css selector is not supported for this session i am following the code from someid click webdriverio | 0 |
224,622 | 17,762,237,766 | IssuesEvent | 2021-08-29 22:43:35 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | test/test_op_aliases.py generates illegal test method names | module: tests triaged | Note the double period:
```
TestOpNormalizationCPU.test_jit_op_alias_normalization_linalg.det_cpu
```
cc @mruberry @VitalyFedyunin @walterddr
| 1.0 | test/test_op_aliases.py generates illegal test method names - Note the double period:
```
TestOpNormalizationCPU.test_jit_op_alias_normalization_linalg.det_cpu
```
cc @mruberry @VitalyFedyunin @walterddr
| non_priority | test test op aliases py generates illegal test method names note the double period testopnormalizationcpu test jit op alias normalization linalg det cpu cc mruberry vitalyfedyunin walterddr | 0 |
115 | 3,298,962,295 | IssuesEvent | 2015-11-02 16:38:24 | magnumripper/JohnTheRipper | https://api.github.com/repos/magnumripper/JohnTheRipper | closed | clang build for Raspberry Pi 2 fails | portability | A regular autoconf build works, with and without openmp. But using clang instead of gcc doesn't.
```
$ clang --version
Raspbian clang version 3.5.0-10+rpi1 (tags/RELEASE_350/final) (based on LLVM 3.5.0)
Target: arm-unknown-linux-gnueabihf
Thread model: posix
$ clang -dumpversion
4.2.1
$ uname -a
Linux raspberrypi 4.1.7-v7+ #817 SMP PREEMPT Sat Sep 19 15:32:00 BST 2015 armv7l GNU/Linux
```
```
clang -DAC_BUILT -march=native -c -g -O2 -I/usr/local/include -DARCH_LITTLE_ENDIAN=1 -Wall -Wdeclaration-after-statement -fomit-frame-pointer -Wno-deprecated-declarations -Wno-format-extra-args -Qunused-arguments -std=gnu89 -Wdate-time -D_POSIX_SOURCE -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE64_SOURCE -pthread -I/usr/local/include -funroll-loops MD5_fmt.c -o MD5_fmt.o
MD5_fmt.c:389:10: error: expected '}'
"MD5 " MD5_ALGORITHM_NAME,
^
./simd-intrinsics.h:79:29: note: expanded from macro 'MD5_ALGORITHM_NAME'
#define MD5_ALGORITHM_NAME BITS " " SIMD_TYPE " " MD5_N_STR
^
MD5_fmt.c:386:2: note: to match this '{'
{
^
1 error generated.
Makefile:1534: recipe for target 'MD5_fmt.o' failed
make[1]: *** [MD5_fmt.o] Error 1
make[1]: Leaving directory '/usb/home/fd/git/JtR/src'
Makefile:187: recipe for target 'default' failed
make: *** [default] Error 2
```
If I try to compile that that single file with gcc, I get the same error:
```
$ gcc -DAC_BUILT -march=native -c -g -O2 -I/usr/local/include -DARCH_LITTLE_ENDIAN=1 -Wall -Wdeclaration-after-statement -fomit-frame-pointer -Wno-deprecated-declarations -Wno-format-extra-args -std=gnu89 -Wdate-time -D_POSIX_SOURCE -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE64_SOURCE -pthread -I/usr/local/include -funroll-loops MD5_fmt.c -o MD5_fmt.o
In file included from MD5_fmt.c:17:0:
simd-intrinsics.h:79:29: error: expected ‘}’ before ‘BITS’
#define MD5_ALGORITHM_NAME BITS " " SIMD_TYPE " " MD5_N_STR
^
MD5_fmt.c:389:10: note: in expansion of macro ‘MD5_ALGORITHM_NAME’
"MD5 " MD5_ALGORITHM_NAME,
^
MD5_fmt.c:52:25: warning: ‘tests’ defined but not used [-Wunused-variable]
static struct fmt_tests tests[] = {
^
```
To see what else fails, I just touched MD5_fmt.o and continued.
```
$ touch MD5_fmt.o; make -s
dynamic_fmt.c:4837:8: error: use of undeclared identifier 'input_buf'
p = input_buf[i+j].w;
^
dynamic_fmt.c:4838:20: error: use of undeclared identifier 'SIMD_COEF_32'
for (k = 0; k < SIMD_COEF_32; k++)
^
dynamic_fmt.c:4839:10: error: use of undeclared identifier 'SIMD_COEF_32'
p[14*SIMD_COEF_32+k] = total_len[i+j][k] << 3;
^
dynamic_fmt.c:4839:28: error: use of undeclared identifier 'total_len'
p[14*SIMD_COEF_32+k] = total_len[i+j][k] << 3;
^
dynamic_fmt.c:4846:8: error: use of undeclared identifier 'input_buf2'
p = input_buf2[i+j].w;
^
dynamic_fmt.c:4847:20: error: use of undeclared identifier 'SIMD_COEF_32'
for (k = 0; k < SIMD_COEF_32; k++)
^
dynamic_fmt.c:4848:10: error: use of undeclared identifier 'SIMD_COEF_32'
p[14*SIMD_COEF_32+k] = total_len2[i+j][k] << 3;
^
dynamic_fmt.c:4848:28: error: use of undeclared identifier 'total_len2'
p[14*SIMD_COEF_32+k] = total_len2[i+j][k] << 3;
^
dynamic_fmt.c:4862:8: error: use of undeclared identifier 'input_buf'
p = input_buf[i+j].w;
^
dynamic_fmt.c:4863:20: error: use of undeclared identifier 'SIMD_COEF_32'
for (k = 0; k < SIMD_COEF_32; k++)
^
dynamic_fmt.c:4864:10: error: use of undeclared identifier 'SIMD_COEF_32'
p[14*SIMD_COEF_32+k] = total_len[i+j][k] << 3;
^
dynamic_fmt.c:4864:28: error: use of undeclared identifier 'total_len'
p[14*SIMD_COEF_32+k] = total_len[i+j][k] << 3;
^
dynamic_fmt.c:4871:8: error: use of undeclared identifier 'input_buf2'
p = input_buf2[i+j].w;
^
dynamic_fmt.c:4872:20: error: use of undeclared identifier 'SIMD_COEF_32'
for (k = 0; k < SIMD_COEF_32; k++)
^
dynamic_fmt.c:4873:10: error: use of undeclared identifier 'SIMD_COEF_32'
p[14*SIMD_COEF_32+k] = total_len2[i+j][k] << 3;
^
dynamic_fmt.c:4873:28: error: use of undeclared identifier 'total_len2'
p[14*SIMD_COEF_32+k] = total_len2[i+j][k] << 3;
^
16 errors generated.
Makefile:1534: recipe for target 'dynamic_fmt.o' failed
make[1]: *** [dynamic_fmt.o] Error 1
Makefile:187: recipe for target 'default' failed
make: *** [default] Error 2
```
Next error:
```
$ touch dynamic_fmt.o; make -s
dynamic_big_crypt.c:521:33: error: use of undeclared identifier 'SIMD_COEF_32'
static const uint32_t MD5_inc = MD5_LOOPS;
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:523:54: error: use of undeclared identifier 'SIMD_COEF_32'
static void DoMD5_crypt_f_sse(void *in, uint32_t len[MD5_LOOPS], void *out) {
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:524:47: error: use of undeclared identifier 'SIMD_COEF_32'
JTR_ALIGN(MEM_ALIGN_SIMD) ARCH_WORD_32 a[(16*MD5_LOOPS)/sizeof(ARCH_WORD_32)];
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:525:23: error: use of undeclared identifier 'SIMD_COEF_32'
uint32_t i, j, loops[MD5_LOOPS], bMore, cnt;
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:528:18: error: use of undeclared identifier 'SIMD_COEF_32'
for (i = 0; i < MD5_LOOPS; ++i) {
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:529:3: error: use of undeclared identifier 'loops'
loops[i] = Do_FixBufferLen32(cp, len[i], 0);
^
dynamic_big_crypt.c:535:2: error: use of undeclared identifier 'bMore'
bMore = 1;
^
dynamic_big_crypt.c:536:2: error: use of undeclared identifier 'cnt'
cnt = 1;
^
dynamic_big_crypt.c:537:9: error: use of undeclared identifier 'bMore'
while (bMore) {
^
dynamic_big_crypt.c:538:66: error: use of undeclared identifier 'cnt'; did you mean 'int'?
SIMDmd5body(cp, a, a, SSEi_FLAT_IN |SSEi_4BUF_INPUT_FIRST_BLK|(cnt==1?0:SSEi_RELOAD));
^
dynamic_big_crypt.c:539:3: error: use of undeclared identifier 'bMore'
bMore = 0;
^
dynamic_big_crypt.c:540:19: error: use of undeclared identifier 'SIMD_COEF_32'
for (i = 0; i < MD5_LOOPS; ++i) {
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:541:8: error: use of undeclared identifier 'cnt'
if (cnt == loops[i]) {
^
dynamic_big_crypt.c:541:15: error: use of undeclared identifier 'loops'
if (cnt == loops[i]) {
^
dynamic_big_crypt.c:542:25: error: use of undeclared identifier 'SIMD_COEF_32'
uint32_t offx = ((i/SIMD_COEF_32)*(16/sizeof(ARCH_WORD_32))*SIMD_COEF_32)+(i&(SIMD_COEF_32-1));
^
dynamic_big_crypt.c:542:65: error: use of undeclared identifier 'SIMD_COEF_32'
uint32_t offx = ((i/SIMD_COEF_32)*(16/sizeof(ARCH_WORD_32))*SIMD_COEF_32)+(i&(SIMD_COEF_32-1));
^
dynamic_big_crypt.c:542:83: error: use of undeclared identifier 'SIMD_COEF_32'
uint32_t offx = ((i/SIMD_COEF_32)*(16/sizeof(ARCH_WORD_32))*SIMD_COEF_32)+(i&(SIMD_COEF_32-1));
^
dynamic_big_crypt.c:544:44: error: use of undeclared identifier 'SIMD_COEF_32'
((ARCH_WORD_32*)out)[(i*4)+j] = a[(j*SIMD_COEF_32)+offx];
^
dynamic_big_crypt.c:546:15: error: use of undeclared identifier 'cnt'
} else if (cnt < loops[i]) bMore = 1;
^
fatal error: too many errors emitted, stopping now [-ferror-limit=]
20 errors generated.
Makefile:1534: recipe for target 'dynamic_big_crypt.o' failed
make[1]: *** [dynamic_big_crypt.o] Error 1
Makefile:187: recipe for target 'default' failed
make: *** [default] Error 2
```
Next error:
```
$ touch dynamic_big_crypt.o; make -s
FGT_fmt_plug.c:245:3: error: expected '}'
ALGORITHM_NAME,
^
FGT_fmt_plug.c:54:33: note: expanded from macro 'ALGORITHM_NAME'
#define ALGORITHM_NAME "SHA1 " SHA1_ALGORITHM_NAME
^
./simd-intrinsics.h:98:30: note: expanded from macro 'SHA1_ALGORITHM_NAME'
#define SHA1_ALGORITHM_NAME BITS " " SIMD_TYPE " " SHA1_N_STR
^
FGT_fmt_plug.c:242:5: note: to match this '{'
{
^
1 error generated.
Makefile:1534: recipe for target 'FGT_fmt_plug.o' failed
make[1]: *** [FGT_fmt_plug.o] Error 1
Makefile:187: recipe for target 'default' failed
make: *** [default] Error 2
```
and so on... | True | clang build for Raspberry Pi 2 fails - A regular autoconf build works, with and without openmp. But using clang instead of gcc doesn't.
```
$ clang --version
Raspbian clang version 3.5.0-10+rpi1 (tags/RELEASE_350/final) (based on LLVM 3.5.0)
Target: arm-unknown-linux-gnueabihf
Thread model: posix
$ clang -dumpversion
4.2.1
$ uname -a
Linux raspberrypi 4.1.7-v7+ #817 SMP PREEMPT Sat Sep 19 15:32:00 BST 2015 armv7l GNU/Linux
```
```
clang -DAC_BUILT -march=native -c -g -O2 -I/usr/local/include -DARCH_LITTLE_ENDIAN=1 -Wall -Wdeclaration-after-statement -fomit-frame-pointer -Wno-deprecated-declarations -Wno-format-extra-args -Qunused-arguments -std=gnu89 -Wdate-time -D_POSIX_SOURCE -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE64_SOURCE -pthread -I/usr/local/include -funroll-loops MD5_fmt.c -o MD5_fmt.o
MD5_fmt.c:389:10: error: expected '}'
"MD5 " MD5_ALGORITHM_NAME,
^
./simd-intrinsics.h:79:29: note: expanded from macro 'MD5_ALGORITHM_NAME'
#define MD5_ALGORITHM_NAME BITS " " SIMD_TYPE " " MD5_N_STR
^
MD5_fmt.c:386:2: note: to match this '{'
{
^
1 error generated.
Makefile:1534: recipe for target 'MD5_fmt.o' failed
make[1]: *** [MD5_fmt.o] Error 1
make[1]: Leaving directory '/usb/home/fd/git/JtR/src'
Makefile:187: recipe for target 'default' failed
make: *** [default] Error 2
```
If I try to compile that that single file with gcc, I get the same error:
```
$ gcc -DAC_BUILT -march=native -c -g -O2 -I/usr/local/include -DARCH_LITTLE_ENDIAN=1 -Wall -Wdeclaration-after-statement -fomit-frame-pointer -Wno-deprecated-declarations -Wno-format-extra-args -std=gnu89 -Wdate-time -D_POSIX_SOURCE -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE64_SOURCE -pthread -I/usr/local/include -funroll-loops MD5_fmt.c -o MD5_fmt.o
In file included from MD5_fmt.c:17:0:
simd-intrinsics.h:79:29: error: expected ‘}’ before ‘BITS’
#define MD5_ALGORITHM_NAME BITS " " SIMD_TYPE " " MD5_N_STR
^
MD5_fmt.c:389:10: note: in expansion of macro ‘MD5_ALGORITHM_NAME’
"MD5 " MD5_ALGORITHM_NAME,
^
MD5_fmt.c:52:25: warning: ‘tests’ defined but not used [-Wunused-variable]
static struct fmt_tests tests[] = {
^
```
To see what else fails, I just touched MD5_fmt.o and continued.
```
$ touch MD5_fmt.o; make -s
dynamic_fmt.c:4837:8: error: use of undeclared identifier 'input_buf'
p = input_buf[i+j].w;
^
dynamic_fmt.c:4838:20: error: use of undeclared identifier 'SIMD_COEF_32'
for (k = 0; k < SIMD_COEF_32; k++)
^
dynamic_fmt.c:4839:10: error: use of undeclared identifier 'SIMD_COEF_32'
p[14*SIMD_COEF_32+k] = total_len[i+j][k] << 3;
^
dynamic_fmt.c:4839:28: error: use of undeclared identifier 'total_len'
p[14*SIMD_COEF_32+k] = total_len[i+j][k] << 3;
^
dynamic_fmt.c:4846:8: error: use of undeclared identifier 'input_buf2'
p = input_buf2[i+j].w;
^
dynamic_fmt.c:4847:20: error: use of undeclared identifier 'SIMD_COEF_32'
for (k = 0; k < SIMD_COEF_32; k++)
^
dynamic_fmt.c:4848:10: error: use of undeclared identifier 'SIMD_COEF_32'
p[14*SIMD_COEF_32+k] = total_len2[i+j][k] << 3;
^
dynamic_fmt.c:4848:28: error: use of undeclared identifier 'total_len2'
p[14*SIMD_COEF_32+k] = total_len2[i+j][k] << 3;
^
dynamic_fmt.c:4862:8: error: use of undeclared identifier 'input_buf'
p = input_buf[i+j].w;
^
dynamic_fmt.c:4863:20: error: use of undeclared identifier 'SIMD_COEF_32'
for (k = 0; k < SIMD_COEF_32; k++)
^
dynamic_fmt.c:4864:10: error: use of undeclared identifier 'SIMD_COEF_32'
p[14*SIMD_COEF_32+k] = total_len[i+j][k] << 3;
^
dynamic_fmt.c:4864:28: error: use of undeclared identifier 'total_len'
p[14*SIMD_COEF_32+k] = total_len[i+j][k] << 3;
^
dynamic_fmt.c:4871:8: error: use of undeclared identifier 'input_buf2'
p = input_buf2[i+j].w;
^
dynamic_fmt.c:4872:20: error: use of undeclared identifier 'SIMD_COEF_32'
for (k = 0; k < SIMD_COEF_32; k++)
^
dynamic_fmt.c:4873:10: error: use of undeclared identifier 'SIMD_COEF_32'
p[14*SIMD_COEF_32+k] = total_len2[i+j][k] << 3;
^
dynamic_fmt.c:4873:28: error: use of undeclared identifier 'total_len2'
p[14*SIMD_COEF_32+k] = total_len2[i+j][k] << 3;
^
16 errors generated.
Makefile:1534: recipe for target 'dynamic_fmt.o' failed
make[1]: *** [dynamic_fmt.o] Error 1
Makefile:187: recipe for target 'default' failed
make: *** [default] Error 2
```
Next error:
```
$ touch dynamic_fmt.o; make -s
dynamic_big_crypt.c:521:33: error: use of undeclared identifier 'SIMD_COEF_32'
static const uint32_t MD5_inc = MD5_LOOPS;
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:523:54: error: use of undeclared identifier 'SIMD_COEF_32'
static void DoMD5_crypt_f_sse(void *in, uint32_t len[MD5_LOOPS], void *out) {
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:524:47: error: use of undeclared identifier 'SIMD_COEF_32'
JTR_ALIGN(MEM_ALIGN_SIMD) ARCH_WORD_32 a[(16*MD5_LOOPS)/sizeof(ARCH_WORD_32)];
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:525:23: error: use of undeclared identifier 'SIMD_COEF_32'
uint32_t i, j, loops[MD5_LOOPS], bMore, cnt;
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:528:18: error: use of undeclared identifier 'SIMD_COEF_32'
for (i = 0; i < MD5_LOOPS; ++i) {
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:529:3: error: use of undeclared identifier 'loops'
loops[i] = Do_FixBufferLen32(cp, len[i], 0);
^
dynamic_big_crypt.c:535:2: error: use of undeclared identifier 'bMore'
bMore = 1;
^
dynamic_big_crypt.c:536:2: error: use of undeclared identifier 'cnt'
cnt = 1;
^
dynamic_big_crypt.c:537:9: error: use of undeclared identifier 'bMore'
while (bMore) {
^
dynamic_big_crypt.c:538:66: error: use of undeclared identifier 'cnt'; did you mean 'int'?
SIMDmd5body(cp, a, a, SSEi_FLAT_IN |SSEi_4BUF_INPUT_FIRST_BLK|(cnt==1?0:SSEi_RELOAD));
^
dynamic_big_crypt.c:539:3: error: use of undeclared identifier 'bMore'
bMore = 0;
^
dynamic_big_crypt.c:540:19: error: use of undeclared identifier 'SIMD_COEF_32'
for (i = 0; i < MD5_LOOPS; ++i) {
^
dynamic_big_crypt.c:520:20: note: expanded from macro 'MD5_LOOPS'
#define MD5_LOOPS (SIMD_COEF_32*SIMD_PARA_MD5)
^
dynamic_big_crypt.c:541:8: error: use of undeclared identifier 'cnt'
if (cnt == loops[i]) {
^
dynamic_big_crypt.c:541:15: error: use of undeclared identifier 'loops'
if (cnt == loops[i]) {
^
dynamic_big_crypt.c:542:25: error: use of undeclared identifier 'SIMD_COEF_32'
uint32_t offx = ((i/SIMD_COEF_32)*(16/sizeof(ARCH_WORD_32))*SIMD_COEF_32)+(i&(SIMD_COEF_32-1));
^
dynamic_big_crypt.c:542:65: error: use of undeclared identifier 'SIMD_COEF_32'
uint32_t offx = ((i/SIMD_COEF_32)*(16/sizeof(ARCH_WORD_32))*SIMD_COEF_32)+(i&(SIMD_COEF_32-1));
^
dynamic_big_crypt.c:542:83: error: use of undeclared identifier 'SIMD_COEF_32'
uint32_t offx = ((i/SIMD_COEF_32)*(16/sizeof(ARCH_WORD_32))*SIMD_COEF_32)+(i&(SIMD_COEF_32-1));
^
dynamic_big_crypt.c:544:44: error: use of undeclared identifier 'SIMD_COEF_32'
((ARCH_WORD_32*)out)[(i*4)+j] = a[(j*SIMD_COEF_32)+offx];
^
dynamic_big_crypt.c:546:15: error: use of undeclared identifier 'cnt'
} else if (cnt < loops[i]) bMore = 1;
^
fatal error: too many errors emitted, stopping now [-ferror-limit=]
20 errors generated.
Makefile:1534: recipe for target 'dynamic_big_crypt.o' failed
make[1]: *** [dynamic_big_crypt.o] Error 1
Makefile:187: recipe for target 'default' failed
make: *** [default] Error 2
```
Next error:
```
$ touch dynamic_big_crypt.o; make -s
FGT_fmt_plug.c:245:3: error: expected '}'
ALGORITHM_NAME,
^
FGT_fmt_plug.c:54:33: note: expanded from macro 'ALGORITHM_NAME'
#define ALGORITHM_NAME "SHA1 " SHA1_ALGORITHM_NAME
^
./simd-intrinsics.h:98:30: note: expanded from macro 'SHA1_ALGORITHM_NAME'
#define SHA1_ALGORITHM_NAME BITS " " SIMD_TYPE " " SHA1_N_STR
^
FGT_fmt_plug.c:242:5: note: to match this '{'
{
^
1 error generated.
Makefile:1534: recipe for target 'FGT_fmt_plug.o' failed
make[1]: *** [FGT_fmt_plug.o] Error 1
Makefile:187: recipe for target 'default' failed
make: *** [default] Error 2
```
and so on... | non_priority | clang build for raspberry pi fails a regular autoconf build works with and without openmp but using clang instead of gcc doesn t clang version raspbian clang version tags release final based on llvm target arm unknown linux gnueabihf thread model posix clang dumpversion uname a linux raspberrypi smp preempt sat sep bst gnu linux clang dac built march native c g i usr local include darch little endian wall wdeclaration after statement fomit frame pointer wno deprecated declarations wno format extra args qunused arguments std wdate time d posix source d gnu source d xopen source d file offset bits d source pthread i usr local include funroll loops fmt c o fmt o fmt c error expected algorithm name simd intrinsics h note expanded from macro algorithm name define algorithm name bits simd type n str fmt c note to match this error generated makefile recipe for target fmt o failed make error make leaving directory usb home fd git jtr src makefile recipe for target default failed make error if i try to compile that that single file with gcc i get the same error gcc dac built march native c g i usr local include darch little endian wall wdeclaration after statement fomit frame pointer wno deprecated declarations wno format extra args std wdate time d posix source d gnu source d xopen source d file offset bits d source pthread i usr local include funroll loops fmt c o fmt o in file included from fmt c simd intrinsics h error expected ‘ ’ before ‘bits’ define algorithm name bits simd type n str fmt c note in expansion of macro ‘ algorithm name’ algorithm name fmt c warning ‘tests’ defined but not used static struct fmt tests tests to see what else fails i just touched fmt o and continued touch fmt o make s dynamic fmt c error use of undeclared identifier input buf p input buf w dynamic fmt c error use of undeclared identifier simd coef for k k simd coef k dynamic fmt c error use of undeclared identifier simd coef p total len dynamic fmt c error use of undeclared identifier total len p total len dynamic fmt c error use of undeclared identifier input p input w dynamic fmt c error use of undeclared identifier simd coef for k k simd coef k dynamic fmt c error use of undeclared identifier simd coef p total dynamic fmt c error use of undeclared identifier total p total dynamic fmt c error use of undeclared identifier input buf p input buf w dynamic fmt c error use of undeclared identifier simd coef for k k simd coef k dynamic fmt c error use of undeclared identifier simd coef p total len dynamic fmt c error use of undeclared identifier total len p total len dynamic fmt c error use of undeclared identifier input p input w dynamic fmt c error use of undeclared identifier simd coef for k k simd coef k dynamic fmt c error use of undeclared identifier simd coef p total dynamic fmt c error use of undeclared identifier total p total errors generated makefile recipe for target dynamic fmt o failed make error makefile recipe for target default failed make error next error touch dynamic fmt o make s dynamic big crypt c error use of undeclared identifier simd coef static const t inc loops dynamic big crypt c note expanded from macro loops define loops simd coef simd para dynamic big crypt c error use of undeclared identifier simd coef static void crypt f sse void in t len void out dynamic big crypt c note expanded from macro loops define loops simd coef simd para dynamic big crypt c error use of undeclared identifier simd coef jtr align mem align simd arch word a dynamic big crypt c note expanded from macro loops define loops simd coef simd para dynamic big crypt c error use of undeclared identifier simd coef t i j loops bmore cnt dynamic big crypt c note expanded from macro loops define loops simd coef simd para dynamic big crypt c error use of undeclared identifier simd coef for i i loops i dynamic big crypt c note expanded from macro loops define loops simd coef simd para dynamic big crypt c error use of undeclared identifier loops loops do cp len dynamic big crypt c error use of undeclared identifier bmore bmore dynamic big crypt c error use of undeclared identifier cnt cnt dynamic big crypt c error use of undeclared identifier bmore while bmore dynamic big crypt c error use of undeclared identifier cnt did you mean int cp a a ssei flat in ssei input first blk cnt ssei reload dynamic big crypt c error use of undeclared identifier bmore bmore dynamic big crypt c error use of undeclared identifier simd coef for i i loops i dynamic big crypt c note expanded from macro loops define loops simd coef simd para dynamic big crypt c error use of undeclared identifier cnt if cnt loops dynamic big crypt c error use of undeclared identifier loops if cnt loops dynamic big crypt c error use of undeclared identifier simd coef t offx i simd coef sizeof arch word simd coef i simd coef dynamic big crypt c error use of undeclared identifier simd coef t offx i simd coef sizeof arch word simd coef i simd coef dynamic big crypt c error use of undeclared identifier simd coef t offx i simd coef sizeof arch word simd coef i simd coef dynamic big crypt c error use of undeclared identifier simd coef arch word out a dynamic big crypt c error use of undeclared identifier cnt else if cnt loops bmore fatal error too many errors emitted stopping now errors generated makefile recipe for target dynamic big crypt o failed make error makefile recipe for target default failed make error next error touch dynamic big crypt o make s fgt fmt plug c error expected algorithm name fgt fmt plug c note expanded from macro algorithm name define algorithm name algorithm name simd intrinsics h note expanded from macro algorithm name define algorithm name bits simd type n str fgt fmt plug c note to match this error generated makefile recipe for target fgt fmt plug o failed make error makefile recipe for target default failed make error and so on | 0 |
35,761 | 14,876,816,111 | IssuesEvent | 2021-01-20 01:42:47 | Azure/azure-sdk-for-net | https://api.github.com/repos/Azure/azure-sdk-for-net | closed | [FEATURE REQ] Way to log the client information for messages being read | Client Service Attention Service Bus customer-reported feature-request needs-team-attention | **Library or service name.**
Microsoft.Azure.ServiceBus
**Is your feature request related to a problem? Please describe.**
We are using the Azure service bus for message processing. We host one or more App Service instances(subscribers) that would be connected to the same Topic and subscription.
After a message is published from the publisher, the message becomes available in subscription and then it gets consumed by one of the App Service instances.
Without actually digging into logs for each App Service instance, it's not possible to know which instance has actually processed the message. Is there a way to get that information from the Service Bus side, if yes, please provide some details. If not, can this be considered as a feature request? I am sure it would add a lot of value to consumers.
| 2.0 | [FEATURE REQ] Way to log the client information for messages being read - **Library or service name.**
Microsoft.Azure.ServiceBus
**Is your feature request related to a problem? Please describe.**
We are using the Azure service bus for message processing. We host one or more App Service instances(subscribers) that would be connected to the same Topic and subscription.
After a message is published from the publisher, the message becomes available in subscription and then it gets consumed by one of the App Service instances.
Without actually digging into logs for each App Service instance, it's not possible to know which instance has actually processed the message. Is there a way to get that information from the Service Bus side, if yes, please provide some details. If not, can this be considered as a feature request? I am sure it would add a lot of value to consumers.
| non_priority | way to log the client information for messages being read library or service name microsoft azure servicebus is your feature request related to a problem please describe we are using the azure service bus for message processing we host one or more app service instances subscribers that would be connected to the same topic and subscription after a message is published from the publisher the message becomes available in subscription and then it gets consumed by one of the app service instances without actually digging into logs for each app service instance it s not possible to know which instance has actually processed the message is there a way to get that information from the service bus side if yes please provide some details if not can this be considered as a feature request i am sure it would add a lot of value to consumers | 0 |
66,137 | 12,728,322,340 | IssuesEvent | 2020-06-25 02:12:01 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | [4.0] Filter options in private messaging layout | No Code Attached Yet | ### Steps to reproduce the issue
Click filter options in private messaging in admin
### Actual result
Note the uber wide dropdown for a tiny field content
<img width="1650" alt="Screenshot 2020-06-09 at 21 44 37" src="https://user-images.githubusercontent.com/400092/84197856-87e77580-aa9a-11ea-8634-a73b56205da5.png">
| 1.0 | [4.0] Filter options in private messaging layout - ### Steps to reproduce the issue
Click filter options in private messaging in admin
### Actual result
Note the uber wide dropdown for a tiny field content
<img width="1650" alt="Screenshot 2020-06-09 at 21 44 37" src="https://user-images.githubusercontent.com/400092/84197856-87e77580-aa9a-11ea-8634-a73b56205da5.png">
| non_priority | filter options in private messaging layout steps to reproduce the issue click filter options in private messaging in admin actual result note the uber wide dropdown for a tiny field content img width alt screenshot at src | 0 |
99,724 | 8,709,493,936 | IssuesEvent | 2018-12-06 14:07:42 | edenlabllc/ehealth.api | https://api.github.com/repos/edenlabllc/ehealth.api | closed | Validate managing_organization==token.client_id on Update episode | [zube]: In Test kind/task priority/high project/medical_events status/test | On *Update Episode* validate ME.patient.episode.managing_organization==client_id
throw error *403* "User is not allowed to perform actions with an episode that belongs to another legal entity"
edenlabllc/ehealth.api#3241
| 2.0 | Validate managing_organization==token.client_id on Update episode - On *Update Episode* validate ME.patient.episode.managing_organization==client_id
throw error *403* "User is not allowed to perform actions with an episode that belongs to another legal entity"
edenlabllc/ehealth.api#3241
| non_priority | validate managing organization token client id on update episode on update episode validate me patient episode managing organization client id throw error user is not allowed to perform actions with an episode that belongs to another legal entity edenlabllc ehealth api | 0 |
6,019 | 8,678,681,370 | IssuesEvent | 2018-11-30 20:46:06 | isawnyu/isaw.web | https://api.github.com/repos/isawnyu/isaw.web | closed | make "filter results" functionality on search accessible: 3 | deploy requirement | > The “filter the results” option should not automatically update the results on the selection of the
checkbox to exclude the item type. This option should require a submit option or should have
the same submit button as the search text field and be placed visually before the search submit.

Traceability: Microassist "Complex Interaction Plan" | 1.0 | make "filter results" functionality on search accessible: 3 - > The “filter the results” option should not automatically update the results on the selection of the
checkbox to exclude the item type. This option should require a submit option or should have
the same submit button as the search text field and be placed visually before the search submit.

Traceability: Microassist "Complex Interaction Plan" | non_priority | make filter results functionality on search accessible the “filter the results” option should not automatically update the results on the selection of the checkbox to exclude the item type this option should require a submit option or should have the same submit button as the search text field and be placed visually before the search submit traceability microassist complex interaction plan | 0 |
198,947 | 22,674,199,161 | IssuesEvent | 2022-07-04 01:26:52 | Techini/WebGoat | https://api.github.com/repos/Techini/WebGoat | closed | CVE-2021-24122 (Medium) detected in tomcat-embed-core-9.0.27.jar - autoclosed | security vulnerability | ## CVE-2021-24122 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.27.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: /webgoat-integration-tests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.27/tomcat-embed-core-9.0.27.jar</p>
<p>
Dependency Hierarchy:
- webwolf-v8.0.0-SNAPSHOT.jar (Root Library)
- spring-boot-starter-web-2.2.0.RELEASE.jar
- spring-boot-starter-tomcat-2.2.0.RELEASE.jar
- :x: **tomcat-embed-core-9.0.27.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Techini/WebGoat/commit/d33cc0e32a0d1b949ff1b85af16890cd452276f8">d33cc0e32a0d1b949ff1b85af16890cd452276f8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When serving resources from a network location using the NTFS file system, Apache Tomcat versions 10.0.0-M1 to 10.0.0-M9, 9.0.0.M1 to 9.0.39, 8.5.0 to 8.5.59 and 7.0.0 to 7.0.106 were susceptible to JSP source code disclosure in some configurations. The root cause was the unexpected behaviour of the JRE API File.getCanonicalPath() which in turn was caused by the inconsistent behaviour of the Windows API (FindFirstFileW) in some circumstances.
<p>Publish Date: 2021-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24122>CVE-2021-24122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122</a></p>
<p>Release Date: 2021-01-14</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.107,8.5.60,9.0.40,10.0.0-M10;org.apache.tomcat:tomcat-catalina:7.0.107,8.5.60,9.0.40,10.0.0-M10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-24122 (Medium) detected in tomcat-embed-core-9.0.27.jar - autoclosed - ## CVE-2021-24122 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.27.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: /webgoat-integration-tests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.27/tomcat-embed-core-9.0.27.jar</p>
<p>
Dependency Hierarchy:
- webwolf-v8.0.0-SNAPSHOT.jar (Root Library)
- spring-boot-starter-web-2.2.0.RELEASE.jar
- spring-boot-starter-tomcat-2.2.0.RELEASE.jar
- :x: **tomcat-embed-core-9.0.27.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Techini/WebGoat/commit/d33cc0e32a0d1b949ff1b85af16890cd452276f8">d33cc0e32a0d1b949ff1b85af16890cd452276f8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When serving resources from a network location using the NTFS file system, Apache Tomcat versions 10.0.0-M1 to 10.0.0-M9, 9.0.0.M1 to 9.0.39, 8.5.0 to 8.5.59 and 7.0.0 to 7.0.106 were susceptible to JSP source code disclosure in some configurations. The root cause was the unexpected behaviour of the JRE API File.getCanonicalPath() which in turn was caused by the inconsistent behaviour of the Windows API (FindFirstFileW) in some circumstances.
<p>Publish Date: 2021-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24122>CVE-2021-24122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122</a></p>
<p>Release Date: 2021-01-14</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.107,8.5.60,9.0.40,10.0.0-M10;org.apache.tomcat:tomcat-catalina:7.0.107,8.5.60,9.0.40,10.0.0-M10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in tomcat embed core jar autoclosed cve medium severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file webgoat integration tests pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy webwolf snapshot jar root library spring boot starter web release jar spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in head commit a href vulnerability details when serving resources from a network location using the ntfs file system apache tomcat versions to to to and to were susceptible to jsp source code disclosure in some configurations the root cause was the unexpected behaviour of the jre api file getcanonicalpath which in turn was caused by the inconsistent behaviour of the windows api findfirstfilew in some circumstances publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource | 0 |
718,547 | 24,722,100,130 | IssuesEvent | 2022-10-20 11:31:00 | googleapis/release-please | https://api.github.com/repos/googleapis/release-please | opened | Support interactive editing PR | type: feature request priority: p3 | **Is your feature request related to a problem? Please describe.**
Listen for PR events to react to any changes.
**Describe the solution you'd like**
1. React when the version in the PR title is edited, which means that the version to be released will change.
2. React when the PR body is edited, which means that the changelog has been manually edited and should be pushed new content to `CHANGELOG.md`.
3. Or even provide options in the PR body (the UI can refer to **renovate**) to customize more! (like `prerelease`).
---
So we can quickly edit it when we are dissatisfied with the changelog of any specific release. | 1.0 | Support interactive editing PR - **Is your feature request related to a problem? Please describe.**
Listen for PR events to react to any changes.
**Describe the solution you'd like**
1. React when the version in the PR title is edited, which means that the version to be released will change.
2. React when the PR body is edited, which means that the changelog has been manually edited and should be pushed new content to `CHANGELOG.md`.
3. Or even provide options in the PR body (the UI can refer to **renovate**) to customize more! (like `prerelease`).
---
So we can quickly edit it when we are dissatisfied with the changelog of any specific release. | priority | support interactive editing pr is your feature request related to a problem please describe listen for pr events to react to any changes describe the solution you d like react when the version in the pr title is edited which means that the version to be released will change react when the pr body is edited which means that the changelog has been manually edited and should be pushed new content to changelog md or even provide options in the pr body the ui can refer to renovate to customize more like prerelease so we can quickly edit it when we are dissatisfied with the changelog of any specific release | 1 |
89,384 | 10,597,249,683 | IssuesEvent | 2019-10-09 23:52:19 | AllenInstitute/AllenSDK | https://api.github.com/repos/AllenInstitute/AllenSDK | opened | Create a style guide | documentation | - [ ] Create a style guide for technology informatics projects
- [ ] Update AllenSDK contributing docs to also reference style guide | 1.0 | Create a style guide - - [ ] Create a style guide for technology informatics projects
- [ ] Update AllenSDK contributing docs to also reference style guide | non_priority | create a style guide create a style guide for technology informatics projects update allensdk contributing docs to also reference style guide | 0 |
306,534 | 26,478,262,553 | IssuesEvent | 2023-01-17 12:53:13 | redhat-beyond/Smarticle | https://api.github.com/repos/redhat-beyond/Smarticle | closed | Test returned values in my_articles | priority: high type: tests | After authentication is added, fix the missing test of returned values in `test_client.py`, `my_articles`. | 1.0 | Test returned values in my_articles - After authentication is added, fix the missing test of returned values in `test_client.py`, `my_articles`. | non_priority | test returned values in my articles after authentication is added fix the missing test of returned values in test client py my articles | 0 |
129,755 | 18,109,735,216 | IssuesEvent | 2021-09-23 01:01:29 | maorkuriel/ksa | https://api.github.com/repos/maorkuriel/ksa | opened | CVE-2017-16138 (High) detected in mime-1.2.4.tgz | security vulnerability | ## CVE-2017-16138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.2.4.tgz</b></p></summary>
<p>A comprehensive library for mime-type mapping</p>
<p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.2.4.tgz">https://registry.npmjs.org/mime/-/mime-1.2.4.tgz</a></p>
<p>Path to dependency file: ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json</p>
<p>Path to vulnerable library: ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/node_modules/mime/package.json</p>
<p>
Dependency Hierarchy:
- connect-2.1.3.tgz (Root Library)
- :x: **mime-1.2.4.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 1.4.1,2.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.2.4","packageFilePaths":["/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json"],"isTransitiveDependency":true,"dependencyTree":"connect:2.1.3;mime:1.2.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-16138 (High) detected in mime-1.2.4.tgz - ## CVE-2017-16138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.2.4.tgz</b></p></summary>
<p>A comprehensive library for mime-type mapping</p>
<p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.2.4.tgz">https://registry.npmjs.org/mime/-/mime-1.2.4.tgz</a></p>
<p>Path to dependency file: ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json</p>
<p>Path to vulnerable library: ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/node_modules/mime/package.json</p>
<p>
Dependency Hierarchy:
- connect-2.1.3.tgz (Root Library)
- :x: **mime-1.2.4.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 1.4.1,2.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.2.4","packageFilePaths":["/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json"],"isTransitiveDependency":true,"dependencyTree":"connect:2.1.3;mime:1.2.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in mime tgz cve high severity vulnerability vulnerable library mime tgz a comprehensive library for mime type mapping library home page a href path to dependency file ksa ksa web root ksa web src main webapp rs bootstrap package json path to vulnerable library ksa ksa web root ksa web src main webapp rs bootstrap node modules mime package json dependency hierarchy connect tgz root library x mime tgz vulnerable library found in base branch master vulnerability details the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree connect mime isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input vulnerabilityurl | 0 |
192,153 | 6,847,231,469 | IssuesEvent | 2017-11-13 14:50:18 | fxi/map-x-mgl | https://api.github.com/repos/fxi/map-x-mgl | closed | Edit view: disable the possibility of saving edits when calculating the summary | Priority 3 | ### Expected Behavior
1) the summary calculation should be quick
2) users should not be allowed to save edits of a view while the summary is calculated
### Current Behavior
At the moment, it is possible to save edits of a view when the summary of the displayed variable is calculated (this process can be a bit long -> tens of seconds). The consequence is that the geometry of the view is no more displayed in the map.
### Possible solution
The save button is not clickable while the summary of the variable is calculated. | 1.0 | Edit view: disable the possibility of saving edits when calculating the summary - ### Expected Behavior
1) the summary calculation should be quick
2) users should not be allowed to save edits of a view while the summary is calculated
### Current Behavior
At the moment, it is possible to save edits of a view when the summary of the displayed variable is calculated (this process can be a bit long -> tens of seconds). The consequence is that the geometry of the view is no more displayed in the map.
### Possible solution
The save button is not clickable while the summary of the variable is calculated. | priority | edit view disable the possibility of saving edits when calculating the summary expected behavior the summary calculation should be quick users should not be allowed to save edits of a view while the summary is calculated current behavior at the moment it is possible to save edits of a view when the summary of the displayed variable is calculated this process can be a bit long tens of seconds the consequence is that the geometry of the view is no more displayed in the map possible solution the save button is not clickable while the summary of the variable is calculated | 1 |
657,720 | 21,802,451,785 | IssuesEvent | 2022-05-16 07:11:36 | oceanprotocol/contracts | https://api.github.com/repos/oceanprotocol/contracts | opened | Dispenser: Unprotected ownerWithdraw() | Type: Bug Priority: Mid | The function ownerWithdraw() is not protected with proper access controls and leaves anyone to call the function and sweeps the entire balance of data token that which Dispenser contract holds.
https://github.com/oceanprotocol/contracts/blob/b937a12b50dc4bdb7a6901c33e5c8fa136697df7/contracts/pools/dispenser/Dispenser.sol#L253 | 1.0 | Dispenser: Unprotected ownerWithdraw() - The function ownerWithdraw() is not protected with proper access controls and leaves anyone to call the function and sweeps the entire balance of data token that which Dispenser contract holds.
https://github.com/oceanprotocol/contracts/blob/b937a12b50dc4bdb7a6901c33e5c8fa136697df7/contracts/pools/dispenser/Dispenser.sol#L253 | priority | dispenser unprotected ownerwithdraw the function ownerwithdraw is not protected with proper access controls and leaves anyone to call the function and sweeps the entire balance of data token that which dispenser contract holds | 1 |
538,681 | 15,775,501,608 | IssuesEvent | 2021-04-01 02:58:35 | muccg/rdrf | https://api.github.com/repos/muccg/rdrf | closed | Secure LDAP modifications | area/cic area/ldap client/hss priority/p0 | For HSS , we need to use secure LDAP.
https://django-auth-ldap.readthedocs.io/en/latest/authentication.html
As per
AUTH_LDAP_CONNECTION_OPTIONS needs to be set with a dictionary of python ldap settings:
e.g
```
# Some optional TLS/SSL options
# AUTH_LDAP_GLOBAL_OPTIONS = {
ldap.OPT_X_TLS_CACERTFILE: "cacertfile",
ldap.OPT_X_TLS_CACERTDIR: "/certs",
ldap.OPT_X_TLS_CERTFILE: 'certfile',
ldap.OPT_X_TLS_KEYFILE: '/certs
ldap.OPT_X_TLS_REQUIRE_CERT: '
AUTH_LDAP_CONNECTION_OPTIONS = {
ldap.OPT_DEBUG_LEVEL: 1, # 0 to 255
ldap.OPT_REFERRALS: 0, # For Active Directory
}
```
including
The folllowing python ldap settings cover tls:
OPT_X_TLS
OPT_X_TLS', optional=True),
OPT_X_TLS_CTX'),
OPT_X_TLS_CACERTFILE'),
OPT_X_TLS_CACERTDIR'),
OPT_X_TLS_CERTFILE'),
OPT_X_TLS_KEYFILE'),
OPT_X_TLS_REQUIRE_CERT'),
OPT_X_TLS_CIPHER_SUITE'),
OPT_X_TLS_RANDOM_FILE'),
OPT_X_TLS_DHFILE'),
OPT_X_TLS_NEVER'),
OPT_X_TLS_HARD'),
OPT_X_TLS_DEMAND'),
OPT_X_TLS_ALLOW'),
OPT_X_TLS_TRY'),
OPT_X_TLS_PEERCERT', optional=True),
OPT_X_TLS_VERSION', optional=True),
OPT_X_TLS_CIPHER', optional=True),
OPT_X_TLS_PEERCERT', optional=True),
OPT_X_TLS_CRLCHECK', optional=True),
OPT_X_TLS_CRLFILE', optional=True),
OPT_X_TLS_CRL_NONE'),
OPT_X_TLS_CRL_PEER'),
OPT_X_TLS_CRL_ALL'),
OPT_X_TLS_NEWCTX', optional=True),
OPT_X_TLS_PROTOCOL_MIN', optional=True),
OPT_X_TLS_PACKAGE', optional=True),
OPT_X_TLS', optional=True),
OPT_X_TLS_CTX'),
OPT_X_TLS_CACERTFILE'),
OPT_X_TLS_CACERTDIR'),
OPT_X_TLS_CERTFILE'),
OPT_X_TLS_KEYFILE'),
OPT_X_TLS_REQUIRE_CERT'),
OPT_X_TLS_CIPHER_SUITE'),
OPT_X_TLS_RANDOM_FILE'),
OPT_X_TLS_DHFILE'),
OPT_X_TLS_NEVER'),
OPT_X_TLS_HARD'),
OPT_X_TLS_DEMAND'),
OPT_X_TLS_ALLOW'),
OPT_X_TLS_TRY'),
OPT_X_TLS_VERSION', optional=True),
OPT_X_TLS_CIPHER', optional=True),
OPT_X_TLS_PEERCERT', optional=True),
OPT_X_TLS_CRLCHECK', optional=True),
OPT_X_TLS_CRLFILE', optional=True),
OPT_X_TLS_CRL_NONE'),
OPT_X_TLS_CRL_PEER'),
OPT_X_TLS_CRL_ALL'),
OPT_X_TLS_NEWCTX', optional=True),
OPT_X_TLS_PROTOCOL_MIN', optional=True),
OPT_X_TLS_PACKAGE', optional=True),
| 1.0 | Secure LDAP modifications - For HSS , we need to use secure LDAP.
https://django-auth-ldap.readthedocs.io/en/latest/authentication.html
As per
AUTH_LDAP_CONNECTION_OPTIONS needs to be set with a dictionary of python ldap settings:
e.g
```
# Some optional TLS/SSL options
# AUTH_LDAP_GLOBAL_OPTIONS = {
ldap.OPT_X_TLS_CACERTFILE: "cacertfile",
ldap.OPT_X_TLS_CACERTDIR: "/certs",
ldap.OPT_X_TLS_CERTFILE: 'certfile',
ldap.OPT_X_TLS_KEYFILE: '/certs
ldap.OPT_X_TLS_REQUIRE_CERT: '
AUTH_LDAP_CONNECTION_OPTIONS = {
ldap.OPT_DEBUG_LEVEL: 1, # 0 to 255
ldap.OPT_REFERRALS: 0, # For Active Directory
}
```
including
The folllowing python ldap settings cover tls:
OPT_X_TLS
OPT_X_TLS', optional=True),
OPT_X_TLS_CTX'),
OPT_X_TLS_CACERTFILE'),
OPT_X_TLS_CACERTDIR'),
OPT_X_TLS_CERTFILE'),
OPT_X_TLS_KEYFILE'),
OPT_X_TLS_REQUIRE_CERT'),
OPT_X_TLS_CIPHER_SUITE'),
OPT_X_TLS_RANDOM_FILE'),
OPT_X_TLS_DHFILE'),
OPT_X_TLS_NEVER'),
OPT_X_TLS_HARD'),
OPT_X_TLS_DEMAND'),
OPT_X_TLS_ALLOW'),
OPT_X_TLS_TRY'),
OPT_X_TLS_PEERCERT', optional=True),
OPT_X_TLS_VERSION', optional=True),
OPT_X_TLS_CIPHER', optional=True),
OPT_X_TLS_PEERCERT', optional=True),
OPT_X_TLS_CRLCHECK', optional=True),
OPT_X_TLS_CRLFILE', optional=True),
OPT_X_TLS_CRL_NONE'),
OPT_X_TLS_CRL_PEER'),
OPT_X_TLS_CRL_ALL'),
OPT_X_TLS_NEWCTX', optional=True),
OPT_X_TLS_PROTOCOL_MIN', optional=True),
OPT_X_TLS_PACKAGE', optional=True),
OPT_X_TLS', optional=True),
OPT_X_TLS_CTX'),
OPT_X_TLS_CACERTFILE'),
OPT_X_TLS_CACERTDIR'),
OPT_X_TLS_CERTFILE'),
OPT_X_TLS_KEYFILE'),
OPT_X_TLS_REQUIRE_CERT'),
OPT_X_TLS_CIPHER_SUITE'),
OPT_X_TLS_RANDOM_FILE'),
OPT_X_TLS_DHFILE'),
OPT_X_TLS_NEVER'),
OPT_X_TLS_HARD'),
OPT_X_TLS_DEMAND'),
OPT_X_TLS_ALLOW'),
OPT_X_TLS_TRY'),
OPT_X_TLS_VERSION', optional=True),
OPT_X_TLS_CIPHER', optional=True),
OPT_X_TLS_PEERCERT', optional=True),
OPT_X_TLS_CRLCHECK', optional=True),
OPT_X_TLS_CRLFILE', optional=True),
OPT_X_TLS_CRL_NONE'),
OPT_X_TLS_CRL_PEER'),
OPT_X_TLS_CRL_ALL'),
OPT_X_TLS_NEWCTX', optional=True),
OPT_X_TLS_PROTOCOL_MIN', optional=True),
OPT_X_TLS_PACKAGE', optional=True),
| priority | secure ldap modifications for hss we need to use secure ldap as per auth ldap connection options needs to be set with a dictionary of python ldap settings e g some optional tls ssl options auth ldap global options ldap opt x tls cacertfile cacertfile ldap opt x tls cacertdir certs ldap opt x tls certfile certfile ldap opt x tls keyfile certs ldap opt x tls require cert auth ldap connection options ldap opt debug level to ldap opt referrals for active directory including the folllowing python ldap settings cover tls opt x tls opt x tls optional true opt x tls ctx opt x tls cacertfile opt x tls cacertdir opt x tls certfile opt x tls keyfile opt x tls require cert opt x tls cipher suite opt x tls random file opt x tls dhfile opt x tls never opt x tls hard opt x tls demand opt x tls allow opt x tls try opt x tls peercert optional true opt x tls version optional true opt x tls cipher optional true opt x tls peercert optional true opt x tls crlcheck optional true opt x tls crlfile optional true opt x tls crl none opt x tls crl peer opt x tls crl all opt x tls newctx optional true opt x tls protocol min optional true opt x tls package optional true opt x tls optional true opt x tls ctx opt x tls cacertfile opt x tls cacertdir opt x tls certfile opt x tls keyfile opt x tls require cert opt x tls cipher suite opt x tls random file opt x tls dhfile opt x tls never opt x tls hard opt x tls demand opt x tls allow opt x tls try opt x tls version optional true opt x tls cipher optional true opt x tls peercert optional true opt x tls crlcheck optional true opt x tls crlfile optional true opt x tls crl none opt x tls crl peer opt x tls crl all opt x tls newctx optional true opt x tls protocol min optional true opt x tls package optional true | 1 |
300,653 | 22,690,917,946 | IssuesEvent | 2022-07-04 20:08:43 | uber/piranha | https://api.github.com/repos/uber/piranha | closed | [Piranha Generic] [Documentation] Broken links in the documentation | piranha polyglot documentation | In [Getting started with Piranha](https://github.com/uber/piranha/tree/master/generic#getting-started-with-piranha) below links are broken:
* [edges.toml](https://github.com/uber/piranha/blob/master/demo/java/configurations/edges.toml)
* [java-ff_system1](https://github.com/uber/piranha/blob/master/piranha/test-resources/java/feature_flag_system_1/control/configurations/rules.toml)
* [java-ff_system2](https://github.com/uber/piranha/blob/master/piranha/test-resources/java/feature_flag_system_2/control/configurations/rules.toml)
* [Adding support for a new feature flag system](https://github.com/uber/piranha/blob/master/generic/adding-support-for-a-new-feature-flag-system) - I think we now have subsections in the README.md itself for this now.
* [Adding Support for a new language](https://github.com/uber/piranha/blob/master/generic/adding-support-for-a-new-language) - I think we now have subsections in the README.md itself for this now.
| 1.0 | [Piranha Generic] [Documentation] Broken links in the documentation - In [Getting started with Piranha](https://github.com/uber/piranha/tree/master/generic#getting-started-with-piranha) below links are broken:
* [edges.toml](https://github.com/uber/piranha/blob/master/demo/java/configurations/edges.toml)
* [java-ff_system1](https://github.com/uber/piranha/blob/master/piranha/test-resources/java/feature_flag_system_1/control/configurations/rules.toml)
* [java-ff_system2](https://github.com/uber/piranha/blob/master/piranha/test-resources/java/feature_flag_system_2/control/configurations/rules.toml)
* [Adding support for a new feature flag system](https://github.com/uber/piranha/blob/master/generic/adding-support-for-a-new-feature-flag-system) - I think we now have subsections in the README.md itself for this now.
* [Adding Support for a new language](https://github.com/uber/piranha/blob/master/generic/adding-support-for-a-new-language) - I think we now have subsections in the README.md itself for this now.
| non_priority | broken links in the documentation in below links are broken i think we now have subsections in the readme md itself for this now i think we now have subsections in the readme md itself for this now | 0 |
322,782 | 23,922,623,818 | IssuesEvent | 2022-09-09 18:32:12 | JohanFalt/Foundry_WoD20 | https://api.github.com/repos/JohanFalt/Foundry_WoD20 | closed | [C20] Changeling the Dreaming Character Sheet | Documentation/discussion Done CtD | ### Top
Name
Player
Chronicle
Court
Legacies
House
Seeming
Kith
Motley
### Attributes
(as usual)
### Talents
Alertness
Athletics
Brawl
Empathy
Expression
Intimidation
**Kenning**
Leadership
Streetwise
Subterfuge
### Skills
Animal Ken
Crafts
Drive
Etiquette
Firearms
Larceny
Melee
Performance
Stealth
Survival
### Knowledges
Academics
Computer
Enigmas
**Gremayre**
Investigation
Law
Medicine
Politics
Science
Technology
### Advantages
Backgrounds
Arts
### Realms
Actor
Fae
Nature
Prop
Scene
Time
Birthright
Frailty
Antithesis
Ravaging Threshold
Musing Threshold
Glamour
Willpower
Nightmare
Banality
[C20_2-Page_Official_Interactive.pdf](https://github.com/JohanFalt/Foundry_WoD20/files/8933600/C20_2-Page_Official_Interactive.pdf) | 1.0 | [C20] Changeling the Dreaming Character Sheet - ### Top
Name
Player
Chronicle
Court
Legacies
House
Seeming
Kith
Motley
### Attributes
(as usual)
### Talents
Alertness
Athletics
Brawl
Empathy
Expression
Intimidation
**Kenning**
Leadership
Streetwise
Subterfuge
### Skills
Animal Ken
Crafts
Drive
Etiquette
Firearms
Larceny
Melee
Performance
Stealth
Survival
### Knowledges
Academics
Computer
Enigmas
**Gremayre**
Investigation
Law
Medicine
Politics
Science
Technology
### Advantages
Backgrounds
Arts
### Realms
Actor
Fae
Nature
Prop
Scene
Time
Birthright
Frailty
Antithesis
Ravaging Threshold
Musing Threshold
Glamour
Willpower
Nightmare
Banality
[C20_2-Page_Official_Interactive.pdf](https://github.com/JohanFalt/Foundry_WoD20/files/8933600/C20_2-Page_Official_Interactive.pdf) | non_priority | changeling the dreaming character sheet top name player chronicle court legacies house seeming kith motley attributes as usual talents alertness athletics brawl empathy expression intimidation kenning leadership streetwise subterfuge skills animal ken crafts drive etiquette firearms larceny melee performance stealth survival knowledges academics computer enigmas gremayre investigation law medicine politics science technology advantages backgrounds arts realms actor fae nature prop scene time birthright frailty antithesis ravaging threshold musing threshold glamour willpower nightmare banality | 0 |
2,396 | 2,925,076,076 | IssuesEvent | 2015-06-26 01:18:11 | Homebrew/linuxbrew | https://api.github.com/repos/Homebrew/linuxbrew | closed | Error: binutils cannot be built with any available compilers. | build-error | Following the standalone instructions homebrew can't install anything because it wants a gcc that only it can provide.
```
$ brew install glibc
==> Installing dependencies for glibc: binutils, linux-headers
==> Installing glibc dependency: binutils
Error: binutils cannot be built with any available compilers.
To install this formula, you may need to:
brew install gcc
```
I had installed linuxbrew on this sytem before but today I wanted to just start over.
Previously, I went through this process and it worked great so I guess this is a new bug.
| 1.0 | Error: binutils cannot be built with any available compilers. - Following the standalone instructions homebrew can't install anything because it wants a gcc that only it can provide.
```
$ brew install glibc
==> Installing dependencies for glibc: binutils, linux-headers
==> Installing glibc dependency: binutils
Error: binutils cannot be built with any available compilers.
To install this formula, you may need to:
brew install gcc
```
I had installed linuxbrew on this sytem before but today I wanted to just start over.
Previously, I went through this process and it worked great so I guess this is a new bug.
| non_priority | error binutils cannot be built with any available compilers following the standalone instructions homebrew can t install anything because it wants a gcc that only it can provide brew install glibc installing dependencies for glibc binutils linux headers installing glibc dependency binutils error binutils cannot be built with any available compilers to install this formula you may need to brew install gcc i had installed linuxbrew on this sytem before but today i wanted to just start over previously i went through this process and it worked great so i guess this is a new bug | 0 |
720,609 | 24,798,665,324 | IssuesEvent | 2022-10-24 19:36:34 | fredo-ai/Fredo-Public | https://api.github.com/repos/fredo-ai/Fredo-Public | opened | Allow to initiate reminder with the words 'remind' or 'reminder' (without the '/') | priority-1 | As a result the following option will also work as new reminders:
1. remind 10:00 buy milk
2. reminder 10:00 buy milk | 1.0 | Allow to initiate reminder with the words 'remind' or 'reminder' (without the '/') - As a result the following option will also work as new reminders:
1. remind 10:00 buy milk
2. reminder 10:00 buy milk | priority | allow to initiate reminder with the words remind or reminder without the as a result the following option will also work as new reminders remind buy milk reminder buy milk | 1 |
130,215 | 18,155,242,367 | IssuesEvent | 2021-09-26 23:45:23 | ghc-dev/Jamie-Larson | https://api.github.com/repos/ghc-dev/Jamie-Larson | opened | CVE-2019-5413 (High) detected in morgan-1.6.1.tgz | security vulnerability | ## CVE-2019-5413 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.6.1.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz">https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz</a></p>
<p>Path to dependency file: Jamie-Larson/package.json</p>
<p>Path to vulnerable library: /node_modules/morgan/package.json</p>
<p>
Dependency Hierarchy:
- :x: **morgan-1.6.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Jamie-Larson/commit/8edca08892d40b47aab5a4e27ff990e11f84258f">8edca08892d40b47aab5a4e27ff990e11f84258f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An attacker can use the format parameter to inject arbitrary commands in the npm package morgan < 1.9.1.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-5413>CVE-2019-5413</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/390881">https://hackerone.com/reports/390881</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: 1.9.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.6.1","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"morgan:1.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-5413","vulnerabilityDetails":"An attacker can use the format parameter to inject arbitrary commands in the npm package morgan \u003c 1.9.1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-5413","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-5413 (High) detected in morgan-1.6.1.tgz - ## CVE-2019-5413 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.6.1.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz">https://registry.npmjs.org/morgan/-/morgan-1.6.1.tgz</a></p>
<p>Path to dependency file: Jamie-Larson/package.json</p>
<p>Path to vulnerable library: /node_modules/morgan/package.json</p>
<p>
Dependency Hierarchy:
- :x: **morgan-1.6.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Jamie-Larson/commit/8edca08892d40b47aab5a4e27ff990e11f84258f">8edca08892d40b47aab5a4e27ff990e11f84258f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An attacker can use the format parameter to inject arbitrary commands in the npm package morgan < 1.9.1.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-5413>CVE-2019-5413</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/390881">https://hackerone.com/reports/390881</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: 1.9.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.6.1","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"morgan:1.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-5413","vulnerabilityDetails":"An attacker can use the format parameter to inject arbitrary commands in the npm package morgan \u003c 1.9.1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-5413","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in morgan tgz cve high severity vulnerability vulnerable library morgan tgz http request logger middleware for node js library home page a href path to dependency file jamie larson package json path to vulnerable library node modules morgan package json dependency hierarchy x morgan tgz vulnerable library found in head commit a href found in base branch master vulnerability details an attacker can use the format parameter to inject arbitrary commands in the npm package morgan publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree morgan isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails an attacker can use the format parameter to inject arbitrary commands in the npm package morgan vulnerabilityurl | 0 |
143,570 | 11,569,724,782 | IssuesEvent | 2020-02-20 18:06:53 | godaddy-wordpress/coblocks | https://api.github.com/repos/godaddy-wordpress/coblocks | closed | ISBAT ensure the integrity of the Click to Tweet block through automated tests | [Type] Tests | Automated tests need to be added to the `click-to-tweet` block which follow the groundwork from #835.
#### Tests Required:
- [x] The save function
- [x] Block transforms #1184
- [x] Block deprecation #956
#### AC
- Any attribute that modifies the serialized block output needs to have tests.
- Any existing transforms to or from a block needs to have tests (if applicable).
- Any existing block deprecations need to have tests (if applicable). | 1.0 | ISBAT ensure the integrity of the Click to Tweet block through automated tests - Automated tests need to be added to the `click-to-tweet` block which follow the groundwork from #835.
#### Tests Required:
- [x] The save function
- [x] Block transforms #1184
- [x] Block deprecation #956
#### AC
- Any attribute that modifies the serialized block output needs to have tests.
- Any existing transforms to or from a block needs to have tests (if applicable).
- Any existing block deprecations need to have tests (if applicable). | non_priority | isbat ensure the integrity of the click to tweet block through automated tests automated tests need to be added to the click to tweet block which follow the groundwork from tests required the save function block transforms block deprecation ac any attribute that modifies the serialized block output needs to have tests any existing transforms to or from a block needs to have tests if applicable any existing block deprecations need to have tests if applicable | 0 |
348,726 | 31,712,334,153 | IssuesEvent | 2023-09-09 12:35:19 | IntellectualSites/fastasyncvoxelsniper | https://api.github.com/repos/IntellectualSites/fastasyncvoxelsniper | opened | `//goto` without arguments throws `java.lang.ArrayIndexOutOfBoundsException` | Requires Testing | ### Server Implementation
Paper
### Server Version
1.20
### Describe the bug
When running the `//goto` command without providing any arguments, a `java.lang.ArrayIndexOutOfBoundsException` is thrown
### To Reproduce
1. Perform`//goto` command without any additional arguments
### Expected behaviour
The correct error message is to be sent to the player
### Screenshots / Videos
_No response_
### Error log (if applicable)
https://hastebin.com/share/dalibapoku.ruby
### Favs Debugpaste
https://athion.net/ISPaster/paste/view/08afc60f5cca48f9b064356393c0d155
### Favs Version
FastAsyncVoxelSniper version 2.9.3
### Checklist
- [X] I have included a Fawe debugpaste.
- [X] I am using the newest build from https://dev.bukkit.org/projects/favs and the issue still persists.
### Anything else?
In the [GotoExecutor](https://github.com/IntellectualSites/fastasyncvoxelsniper/blob/69370d854576a73cb4c5f2618ebb5e26dc7a8e45/src/main/java/com/thevoxelbox/voxelsniper/command/executor/GotoExecutor.java#L36) the `java.lang.ArrayIndexOutOfBoundsException` is caught.
However using
```java
x = Integer.parseInt(arguments[0]);
z = Integer.parseInt(arguments[1]);
```
for getting the x- and z-coordinate input throws the new exception | 1.0 | `//goto` without arguments throws `java.lang.ArrayIndexOutOfBoundsException` - ### Server Implementation
Paper
### Server Version
1.20
### Describe the bug
When running the `//goto` command without providing any arguments, a `java.lang.ArrayIndexOutOfBoundsException` is thrown
### To Reproduce
1. Perform`//goto` command without any additional arguments
### Expected behaviour
The correct error message is to be sent to the player
### Screenshots / Videos
_No response_
### Error log (if applicable)
https://hastebin.com/share/dalibapoku.ruby
### Favs Debugpaste
https://athion.net/ISPaster/paste/view/08afc60f5cca48f9b064356393c0d155
### Favs Version
FastAsyncVoxelSniper version 2.9.3
### Checklist
- [X] I have included a Fawe debugpaste.
- [X] I am using the newest build from https://dev.bukkit.org/projects/favs and the issue still persists.
### Anything else?
In the [GotoExecutor](https://github.com/IntellectualSites/fastasyncvoxelsniper/blob/69370d854576a73cb4c5f2618ebb5e26dc7a8e45/src/main/java/com/thevoxelbox/voxelsniper/command/executor/GotoExecutor.java#L36) the `java.lang.ArrayIndexOutOfBoundsException` is caught.
However using
```java
x = Integer.parseInt(arguments[0]);
z = Integer.parseInt(arguments[1]);
```
for getting the x- and z-coordinate input throws the new exception | non_priority | goto without arguments throws java lang arrayindexoutofboundsexception server implementation paper server version describe the bug when running the goto command without providing any arguments a java lang arrayindexoutofboundsexception is thrown to reproduce perform goto command without any additional arguments expected behaviour the correct error message is to be sent to the player screenshots videos no response error log if applicable favs debugpaste favs version fastasyncvoxelsniper version checklist i have included a fawe debugpaste i am using the newest build from and the issue still persists anything else in the the java lang arrayindexoutofboundsexception is caught however using java x integer parseint arguments z integer parseint arguments for getting the x and z coordinate input throws the new exception | 0 |
52,117 | 21,994,682,892 | IssuesEvent | 2022-05-26 04:17:44 | googleapis/python-service-directory | https://api.github.com/repos/googleapis/python-service-directory | closed | tests.unit.gapic.servicedirectory_v1beta1.test_registration_service: test_list_namespaces_pager failed | type: bug priority: p1 api: servicedirectory flakybot: issue | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 3e89efea0d6c8f9940594ae7cd8249688c126369
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/a14573f0-be5f-4a69-81d7-fb46bc1a0dc3), [Sponge](http://sponge2/a14573f0-be5f-4a69-81d7-fb46bc1a0dc3)
status: failed
<details><summary>Test output</summary><br><pre>transport_name = 'grpc'
def test_list_namespaces_pager(transport_name: str = "grpc"):
client = RegistrationServiceClient(
credentials=ga_credentials.AnonymousCredentials,
transport=transport_name,
)
# Mock the actual call within the gRPC stub, and fake the request.
with mock.patch.object(type(client.transport.list_namespaces), "__call__") as call:
# Set the response to a series of pages.
call.side_effect = (
registration_service.ListNamespacesResponse(
namespaces=[
namespace.Namespace(),
namespace.Namespace(),
namespace.Namespace(),
],
next_page_token="abc",
),
registration_service.ListNamespacesResponse(
namespaces=[],
next_page_token="def",
),
registration_service.ListNamespacesResponse(
namespaces=[
namespace.Namespace(),
],
next_page_token="ghi",
),
registration_service.ListNamespacesResponse(
namespaces=[
namespace.Namespace(),
namespace.Namespace(),
],
),
RuntimeError,
)
metadata = ()
metadata = tuple(metadata) + (
gapic_v1.routing_header.to_grpc_metadata((("parent", ""),)),
)
pager = client.list_namespaces(request={})
assert pager._metadata == metadata
results = list(pager)
assert len(results) == 6
> assert all(isinstance(i, namespace.Namespace) for i in results)
E assert False
E + where False = all(<generator object test_list_namespaces_pager.<locals>.<genexpr> at 0x7fa977464580>)
tests/unit/gapic/servicedirectory_v1beta1/test_registration_service.py:1214: AssertionError</pre></details> | 1.0 | tests.unit.gapic.servicedirectory_v1beta1.test_registration_service: test_list_namespaces_pager failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 3e89efea0d6c8f9940594ae7cd8249688c126369
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/a14573f0-be5f-4a69-81d7-fb46bc1a0dc3), [Sponge](http://sponge2/a14573f0-be5f-4a69-81d7-fb46bc1a0dc3)
status: failed
<details><summary>Test output</summary><br><pre>transport_name = 'grpc'
def test_list_namespaces_pager(transport_name: str = "grpc"):
client = RegistrationServiceClient(
credentials=ga_credentials.AnonymousCredentials,
transport=transport_name,
)
# Mock the actual call within the gRPC stub, and fake the request.
with mock.patch.object(type(client.transport.list_namespaces), "__call__") as call:
# Set the response to a series of pages.
call.side_effect = (
registration_service.ListNamespacesResponse(
namespaces=[
namespace.Namespace(),
namespace.Namespace(),
namespace.Namespace(),
],
next_page_token="abc",
),
registration_service.ListNamespacesResponse(
namespaces=[],
next_page_token="def",
),
registration_service.ListNamespacesResponse(
namespaces=[
namespace.Namespace(),
],
next_page_token="ghi",
),
registration_service.ListNamespacesResponse(
namespaces=[
namespace.Namespace(),
namespace.Namespace(),
],
),
RuntimeError,
)
metadata = ()
metadata = tuple(metadata) + (
gapic_v1.routing_header.to_grpc_metadata((("parent", ""),)),
)
pager = client.list_namespaces(request={})
assert pager._metadata == metadata
results = list(pager)
assert len(results) == 6
> assert all(isinstance(i, namespace.Namespace) for i in results)
E assert False
E + where False = all(<generator object test_list_namespaces_pager.<locals>.<genexpr> at 0x7fa977464580>)
tests/unit/gapic/servicedirectory_v1beta1/test_registration_service.py:1214: AssertionError</pre></details> | non_priority | tests unit gapic servicedirectory test registration service test list namespaces pager failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output transport name grpc def test list namespaces pager transport name str grpc client registrationserviceclient credentials ga credentials anonymouscredentials transport transport name mock the actual call within the grpc stub and fake the request with mock patch object type client transport list namespaces call as call set the response to a series of pages call side effect registration service listnamespacesresponse namespaces namespace namespace namespace namespace namespace namespace next page token abc registration service listnamespacesresponse namespaces next page token def registration service listnamespacesresponse namespaces namespace namespace next page token ghi registration service listnamespacesresponse namespaces namespace namespace namespace namespace runtimeerror metadata metadata tuple metadata gapic routing header to grpc metadata parent pager client list namespaces request assert pager metadata metadata results list pager assert len results assert all isinstance i namespace namespace for i in results e assert false e where false all at tests unit gapic servicedirectory test registration service py assertionerror | 0 |
404,005 | 27,446,179,199 | IssuesEvent | 2023-03-02 14:26:17 | NetAppDocs/ontap-select | https://api.github.com/repos/NetAppDocs/ontap-select | closed | ONTAP Select versions need to be updated to included latest version / current verison | documentation | Page: [Storage efficiency support](https://docs.netapp.com/us-en/ontap-select/concept_stor_eff_support.html)
Versioning needs to be updated in multiple places, for example "9.6 All Licenses" should be something like "9.6 and newer - All licenses" | 1.0 | ONTAP Select versions need to be updated to included latest version / current verison - Page: [Storage efficiency support](https://docs.netapp.com/us-en/ontap-select/concept_stor_eff_support.html)
Versioning needs to be updated in multiple places, for example "9.6 All Licenses" should be something like "9.6 and newer - All licenses" | non_priority | ontap select versions need to be updated to included latest version current verison page versioning needs to be updated in multiple places for example all licenses should be something like and newer all licenses | 0 |
256,978 | 22,139,286,714 | IssuesEvent | 2022-06-03 04:13:48 | microsoft/playwright | https://api.github.com/repos/microsoft/playwright | closed | [Playwright BUG]: Worker process exited unexpectedly | feature-test-runner v1.23 | **Context:**
Playwright Version: [1.22]
Operating System: [Windows]
Node.js version: [v12.18.3]
Browser: [ Chrome v101.0.4951.67]
**Code Snippet**
```js
for (let executeNo = 0; executeNo < 480; executeNo++) {
let describeName = `iterate on loopEach block_${executeNo}`
test.describe(describeName, () => {
test.beforeEach(async ({ page }) => {
await page.goto(`http://${argv.testServer}:5555`, { timeout: 60000 });
});
test.afterEach(async ({ page }) => {
await page.close();
});
test('iterate on loopEach block', async ({ page }) => {
//Action Performed ""
});
}
```
**Conf File Details**
```
use: {
browserName: "chromium",
headless: false,
viewport: { width: 1920, height: 1080 },
ignoreHTTPSErrors: true,
trace: 'retain-on-failure',
screenshot: 'only-on-failure',
channel: "chrome",
video: 'retain-on-failure',
},
```
**Describe the bug**
On running one Testcase in Loop (i.e 480)as described above ,it breaks in between i.e till 371 times loops the testcase works fine
but rest all Testcase get skipped with below error
**Worker process exited unexpectedly**
Please help here in understanding why this issue happening
| 1.0 | [Playwright BUG]: Worker process exited unexpectedly - **Context:**
Playwright Version: [1.22]
Operating System: [Windows]
Node.js version: [v12.18.3]
Browser: [ Chrome v101.0.4951.67]
**Code Snippet**
```js
for (let executeNo = 0; executeNo < 480; executeNo++) {
let describeName = `iterate on loopEach block_${executeNo}`
test.describe(describeName, () => {
test.beforeEach(async ({ page }) => {
await page.goto(`http://${argv.testServer}:5555`, { timeout: 60000 });
});
test.afterEach(async ({ page }) => {
await page.close();
});
test('iterate on loopEach block', async ({ page }) => {
//Action Performed ""
});
}
```
**Conf File Details**
```
use: {
browserName: "chromium",
headless: false,
viewport: { width: 1920, height: 1080 },
ignoreHTTPSErrors: true,
trace: 'retain-on-failure',
screenshot: 'only-on-failure',
channel: "chrome",
video: 'retain-on-failure',
},
```
**Describe the bug**
On running one Testcase in Loop (i.e 480)as described above ,it breaks in between i.e till 371 times loops the testcase works fine
but rest all Testcase get skipped with below error
**Worker process exited unexpectedly**
Please help here in understanding why this issue happening
| non_priority | worker process exited unexpectedly context playwright version operating system node js version browser code snippet js for let executeno executeno executeno let describename iterate on loopeach block executeno test describe describename test beforeeach async page await page goto timeout test aftereach async page await page close test iterate on loopeach block async page action performed conf file details use browsername chromium headless false viewport width height ignorehttpserrors true trace retain on failure screenshot only on failure channel chrome video retain on failure describe the bug on running one testcase in loop i e as described above it breaks in between i e till times loops the testcase works fine but rest all testcase get skipped with below error worker process exited unexpectedly please help here in understanding why this issue happening | 0 |
158,698 | 24,879,532,126 | IssuesEvent | 2022-10-27 22:46:38 | angelolab/ark-analysis | https://api.github.com/repos/angelolab/ark-analysis | closed | Integrating example dataset with pixel, cell, and post clustering | design_doc | *This is for internal use only; if you'd like to open an issue or request a new feature, please open a bug or enhancement issue*
**Section 1: Design details**
**Relevant background**
After the example dataset is finalized for the segmentation notebook, it needs to integrate seamlessly into pixel and cell clustering.
**Design overview**
This project will require us to refactor the directory structure of `example_dataset`. We should have no less than all the required input files for each script, and no extraneous files either.
**Design list/flowchart**
Our sample dataset folder should look like:
```sh
data/
└── example_dataset/
├── input_data/
│ ├── fov0/
│ │ ├── ...
│ ├── ...
├── segmentation/
│ ├── deepcell_input/
│ ├── deepcell_output/
│ │ ├── fov0_feature_0.tif
│ │ ├── ...
│ ├── deepcell_visualization/
│ ├── pixel_mat_data/
│ │ ├── fov0.feather
│ │ ├── ...
│ ├── {pixel_cluster_prefix}_pixel_output_dir/
│ │ ├── cell_clustering_params.json
│ │ ├── pixel_channel_avg_som_cluster.csv
│ │ ├── pixel_channel_avg_meta_cluster.csv
│ └── cell_table/
│ │ ├── example_cell_table_size_normalized.csv
│ │ ├── example_cell_table_arcsinh_transformed.csv
```
`cell_clustering_params.json` should be set up as follows:
```
{
'fovs': ['fov0', 'fov1', ...],
'channels': [//list of channels used],
'segmentation_dir': 'path_to_deepcell_output',
'seg_suffix': '_feature_0.tif',
'pixel_data_dir': 'path_to_pixel_mat_data',
'pc_chan_avg_som_cluster_name': 'path_to_pixel_channel_avg_som_cluster.csv',
'pc_chan_avg_meta_cluster_name': 'path_to_pixel_channel_avg_meta_cluster.csv'
}
```
A few notes:
- `pixel_mat_data` normally gets placed inside the `pixel_output_dir` folder. We need to separate this out in the example dataset so pixel clustering doesn't skip training/assignment/consensus clustering and cell clustering can use example pixel clustered data. `cell_clustering_params.json` needs to be set with the appropriate paths.
- The data in `pixel_mat_data` must include `pixel_som_cluster`, `pixel_meta_cluster`, and `pixel_meta_cluster_rename` columns.
- The cell tables need to include `pixel_meta_cluster` columns in order for notebook 4 to be demoed at any time. Notebook 4 should reference these example cell tables, and not the ones created by notebook 1 (because the cell tables generated by the latter don't contain a `pixel_meta_cluster` column).
- The data other than the `pixel_meta_cluster` column will otherwise be exactly identical.
For Hugging Face, we will need to create the following datasets:
```sh
segmentation/
├── fov0_feature_0.tif
├── ...
└── fov10_feature_0.tif
```
```sh
cell_tables/
├── cell_table_arcsinh_transformed.csv
└── cell_table_size_normalized.csv
```
```sh
{pixel_cluster_prefix}_pixel_output_dir/
├── pixel_channel_avg_meta_cluster.csv
├── pixel_channel_avg_som_cluster.csv
└── cell_clustering_params.json
```
```sh
pixel_mat_data/
├── fov0.feather
├── ...
└── fov10.feather
```
In addition the loading script needs to be adjusted (by adding builder configs) for each of these. In addition the loading function in Ark, will need to be adjusted to account for them as well.
**Required inputs**
This is detailed in the above section.
**Output files**
The outputs each notebook produces will be the same as running the example dataset through the entire process as it stands.
**Section 2: Implementation details**
Once you have completed section 1, please tag the relevant parties and iterate on the initial design details until everyone is satisfied. Then,
proceed to section 2
**Control flow**
Provide additional, more granular details (if necessary) about how the proposed coding logic will be put together
**Milestones and timeline**
List each of the major components of the project, and provide an estimated completion date for each one.
Timeline Give a rough estimate for how long you think the project will take. In general, it's better to be too conservative rather than too optimistic.
**Multiple weeks.** For large projects, make sure to agree on a plan that isn't just a single monster PR at the end.
Create 1 PR for each notebook:
1. Notebook 2: Cluster Pixels -- #736
1. [x] Update the dataset on HuggingFace
2. [x] Edit the notebook paths
3. [x] Edit the test functions
4. [x] Edit the notebook tests
2. Notebook 3: Cluster Cells -- #760
1. [x] Update the dataset on HuggingFace
2. [x] Edit the notebook paths
3. [x] Edit the test functions
4. [x] Edit the notebook tests
3. Notebook 4: Post Clustering -- #763
1. [x] Update the dataset on HuggingFace
2. [x] Edit the notebook paths
3. [x] Edit the test functions
4. [x] Edit the notebook tests
Estimated date when a fully implemented version will be ready for review: 9/30
Estimated date when the finalized project will be merged in: 10/7 | 1.0 | Integrating example dataset with pixel, cell, and post clustering - *This is for internal use only; if you'd like to open an issue or request a new feature, please open a bug or enhancement issue*
**Section 1: Design details**
**Relevant background**
After the example dataset is finalized for the segmentation notebook, it needs to integrate seamlessly into pixel and cell clustering.
**Design overview**
This project will require us to refactor the directory structure of `example_dataset`. We should have no less than all the required input files for each script, and no extraneous files either.
**Design list/flowchart**
Our sample dataset folder should look like:
```sh
data/
└── example_dataset/
├── input_data/
│ ├── fov0/
│ │ ├── ...
│ ├── ...
├── segmentation/
│ ├── deepcell_input/
│ ├── deepcell_output/
│ │ ├── fov0_feature_0.tif
│ │ ├── ...
│ ├── deepcell_visualization/
│ ├── pixel_mat_data/
│ │ ├── fov0.feather
│ │ ├── ...
│ ├── {pixel_cluster_prefix}_pixel_output_dir/
│ │ ├── cell_clustering_params.json
│ │ ├── pixel_channel_avg_som_cluster.csv
│ │ ├── pixel_channel_avg_meta_cluster.csv
│ └── cell_table/
│ │ ├── example_cell_table_size_normalized.csv
│ │ ├── example_cell_table_arcsinh_transformed.csv
```
`cell_clustering_params.json` should be set up as follows:
```
{
'fovs': ['fov0', 'fov1', ...],
'channels': [//list of channels used],
'segmentation_dir': 'path_to_deepcell_output',
'seg_suffix': '_feature_0.tif',
'pixel_data_dir': 'path_to_pixel_mat_data',
'pc_chan_avg_som_cluster_name': 'path_to_pixel_channel_avg_som_cluster.csv',
'pc_chan_avg_meta_cluster_name': 'path_to_pixel_channel_avg_meta_cluster.csv'
}
```
A few notes:
- `pixel_mat_data` normally gets placed inside the `pixel_output_dir` folder. We need to separate this out in the example dataset so pixel clustering doesn't skip training/assignment/consensus clustering and cell clustering can use example pixel clustered data. `cell_clustering_params.json` needs to be set with the appropriate paths.
- The data in `pixel_mat_data` must include `pixel_som_cluster`, `pixel_meta_cluster`, and `pixel_meta_cluster_rename` columns.
- The cell tables need to include `pixel_meta_cluster` columns in order for notebook 4 to be demoed at any time. Notebook 4 should reference these example cell tables, and not the ones created by notebook 1 (because the cell tables generated by the latter don't contain a `pixel_meta_cluster` column).
- The data other than the `pixel_meta_cluster` column will otherwise be exactly identical.
For Hugging Face, we will need to create the following datasets:
```sh
segmentation/
├── fov0_feature_0.tif
├── ...
└── fov10_feature_0.tif
```
```sh
cell_tables/
├── cell_table_arcsinh_transformed.csv
└── cell_table_size_normalized.csv
```
```sh
{pixel_cluster_prefix}_pixel_output_dir/
├── pixel_channel_avg_meta_cluster.csv
├── pixel_channel_avg_som_cluster.csv
└── cell_clustering_params.json
```
```sh
pixel_mat_data/
├── fov0.feather
├── ...
└── fov10.feather
```
In addition the loading script needs to be adjusted (by adding builder configs) for each of these. In addition the loading function in Ark, will need to be adjusted to account for them as well.
**Required inputs**
This is detailed in the above section.
**Output files**
The outputs each notebook produces will be the same as running the example dataset through the entire process as it stands.
**Section 2: Implementation details**
Once you have completed section 1, please tag the relevant parties and iterate on the initial design details until everyone is satisfied. Then,
proceed to section 2
**Control flow**
Provide additional, more granular details (if necessary) about how the proposed coding logic will be put together
**Milestones and timeline**
List each of the major components of the project, and provide an estimated completion date for each one.
Timeline Give a rough estimate for how long you think the project will take. In general, it's better to be too conservative rather than too optimistic.
**Multiple weeks.** For large projects, make sure to agree on a plan that isn't just a single monster PR at the end.
Create 1 PR for each notebook:
1. Notebook 2: Cluster Pixels -- #736
1. [x] Update the dataset on HuggingFace
2. [x] Edit the notebook paths
3. [x] Edit the test functions
4. [x] Edit the notebook tests
2. Notebook 3: Cluster Cells -- #760
1. [x] Update the dataset on HuggingFace
2. [x] Edit the notebook paths
3. [x] Edit the test functions
4. [x] Edit the notebook tests
3. Notebook 4: Post Clustering -- #763
1. [x] Update the dataset on HuggingFace
2. [x] Edit the notebook paths
3. [x] Edit the test functions
4. [x] Edit the notebook tests
Estimated date when a fully implemented version will be ready for review: 9/30
Estimated date when the finalized project will be merged in: 10/7 | non_priority | integrating example dataset with pixel cell and post clustering this is for internal use only if you d like to open an issue or request a new feature please open a bug or enhancement issue section design details relevant background after the example dataset is finalized for the segmentation notebook it needs to integrate seamlessly into pixel and cell clustering design overview this project will require us to refactor the directory structure of example dataset we should have no less than all the required input files for each script and no extraneous files either design list flowchart our sample dataset folder should look like sh data └── example dataset ├── input data │ ├── │ │ ├── │ ├── ├── segmentation │ ├── deepcell input │ ├── deepcell output │ │ ├── feature tif │ │ ├── │ ├── deepcell visualization │ ├── pixel mat data │ │ ├── feather │ │ ├── │ ├── pixel cluster prefix pixel output dir │ │ ├── cell clustering params json │ │ ├── pixel channel avg som cluster csv │ │ ├── pixel channel avg meta cluster csv │ └── cell table │ │ ├── example cell table size normalized csv │ │ ├── example cell table arcsinh transformed csv cell clustering params json should be set up as follows fovs channels segmentation dir path to deepcell output seg suffix feature tif pixel data dir path to pixel mat data pc chan avg som cluster name path to pixel channel avg som cluster csv pc chan avg meta cluster name path to pixel channel avg meta cluster csv a few notes pixel mat data normally gets placed inside the pixel output dir folder we need to separate this out in the example dataset so pixel clustering doesn t skip training assignment consensus clustering and cell clustering can use example pixel clustered data cell clustering params json needs to be set with the appropriate paths the data in pixel mat data must include pixel som cluster pixel meta cluster and pixel meta cluster rename columns the cell tables need to include pixel meta cluster columns in order for notebook to be demoed at any time notebook should reference these example cell tables and not the ones created by notebook because the cell tables generated by the latter don t contain a pixel meta cluster column the data other than the pixel meta cluster column will otherwise be exactly identical for hugging face we will need to create the following datasets sh segmentation ├── feature tif ├── └── feature tif sh cell tables ├── cell table arcsinh transformed csv └── cell table size normalized csv sh pixel cluster prefix pixel output dir ├── pixel channel avg meta cluster csv ├── pixel channel avg som cluster csv └── cell clustering params json sh pixel mat data ├── feather ├── └── feather in addition the loading script needs to be adjusted by adding builder configs for each of these in addition the loading function in ark will need to be adjusted to account for them as well required inputs this is detailed in the above section output files the outputs each notebook produces will be the same as running the example dataset through the entire process as it stands section implementation details once you have completed section please tag the relevant parties and iterate on the initial design details until everyone is satisfied then proceed to section control flow provide additional more granular details if necessary about how the proposed coding logic will be put together milestones and timeline list each of the major components of the project and provide an estimated completion date for each one timeline give a rough estimate for how long you think the project will take in general it s better to be too conservative rather than too optimistic multiple weeks for large projects make sure to agree on a plan that isn t just a single monster pr at the end create pr for each notebook notebook cluster pixels update the dataset on huggingface edit the notebook paths edit the test functions edit the notebook tests notebook cluster cells update the dataset on huggingface edit the notebook paths edit the test functions edit the notebook tests notebook post clustering update the dataset on huggingface edit the notebook paths edit the test functions edit the notebook tests estimated date when a fully implemented version will be ready for review estimated date when the finalized project will be merged in | 0 |
111,958 | 4,499,720,586 | IssuesEvent | 2016-09-01 00:12:19 | socrata/opendatanetwork.com | https://api.github.com/repos/socrata/opendatanetwork.com | closed | page toggles not generating all constraint permutations, so no map appearing | high priority | go to: http://opendatanetwork-staging.herokuapp.com/region/310M200US42660/Seattle_Metro_Area_(WA)/cost_of_living/economy.cost_of_living.index/2013?
note: no map
note: map error: https://odn-backend.herokuapp.com/data/v1/map/new?variable=economy.cost_of_living.index&entity_id=310M200US42660&app_token=cQovpGcdUT1CSzgYk0KPYdAI0&year=2013
{"error":{"message":"must specify values for constraints: component"},"statusCode":422,"url":"/data/v1/map/new?variable=economy.cost_of_living.index&entity_id=310M200US42660&app_token=cQovpGcdUT1CSzgYk0KPYdAI0&year=2013"}
what we want is a drop down for every constraint in the /availability call. note, in this case there are 2, year and component. The drop down values should contain all of the permutations for that constraint using this call:
http://api.opendatanetwork.com/data/v1/constraint/demographics.population.count?entity_id=310M200US42660&constraint=year&app_token=cQovpGcdUT1CSzgYk0KPYdAI0
where the UI picks the order and default. | 1.0 | page toggles not generating all constraint permutations, so no map appearing - go to: http://opendatanetwork-staging.herokuapp.com/region/310M200US42660/Seattle_Metro_Area_(WA)/cost_of_living/economy.cost_of_living.index/2013?
note: no map
note: map error: https://odn-backend.herokuapp.com/data/v1/map/new?variable=economy.cost_of_living.index&entity_id=310M200US42660&app_token=cQovpGcdUT1CSzgYk0KPYdAI0&year=2013
{"error":{"message":"must specify values for constraints: component"},"statusCode":422,"url":"/data/v1/map/new?variable=economy.cost_of_living.index&entity_id=310M200US42660&app_token=cQovpGcdUT1CSzgYk0KPYdAI0&year=2013"}
what we want is a drop down for every constraint in the /availability call. note, in this case there are 2, year and component. The drop down values should contain all of the permutations for that constraint using this call:
http://api.opendatanetwork.com/data/v1/constraint/demographics.population.count?entity_id=310M200US42660&constraint=year&app_token=cQovpGcdUT1CSzgYk0KPYdAI0
where the UI picks the order and default. | priority | page toggles not generating all constraint permutations so no map appearing go to note no map note map error error message must specify values for constraints component statuscode url data map new variable economy cost of living index entity id app token year what we want is a drop down for every constraint in the availability call note in this case there are year and component the drop down values should contain all of the permutations for that constraint using this call where the ui picks the order and default | 1 |
608,506 | 18,840,730,429 | IssuesEvent | 2021-11-11 09:15:19 | harvester/harvester | https://api.github.com/repos/harvester/harvester | closed | [BUG] VM's resource restriction is not match to hosts state | bug area/ui priority/2 | ### Describe the bug
A host shows 0.62 of 6 core is used, but we can't allocate more than 2 cores for VM.
** VM's test cases No.71-76
### To Reproduce
Steps to reproduce the behavior:
1. install Harvester on hosts with different resources
2. create a VM which CPU is more than half of maximum
3. VM goes `starting` state and shows error `Insufficient cpu`
### Expected behavior
User can't know how many CPUs is dedicated for Node and the dashboard shows there still many resource is free to allocated for VMs.
## Environment:
- Harvester ISO version: **master-a4528869-head**
- Underlying Infrastructure (e.g. Baremetal with Dell PowerEdge R630): **QEMU/KVM** 3nodes
### Additional context
[hosts state]

[VM's status]

[allocate over half of maximum RAM]

[hosts state after the VM is running]

| 1.0 | [BUG] VM's resource restriction is not match to hosts state - ### Describe the bug
A host shows 0.62 of 6 core is used, but we can't allocate more than 2 cores for VM.
** VM's test cases No.71-76
### To Reproduce
Steps to reproduce the behavior:
1. install Harvester on hosts with different resources
2. create a VM which CPU is more than half of maximum
3. VM goes `starting` state and shows error `Insufficient cpu`
### Expected behavior
User can't know how many CPUs is dedicated for Node and the dashboard shows there still many resource is free to allocated for VMs.
## Environment:
- Harvester ISO version: **master-a4528869-head**
- Underlying Infrastructure (e.g. Baremetal with Dell PowerEdge R630): **QEMU/KVM** 3nodes
### Additional context
[hosts state]

[VM's status]

[allocate over half of maximum RAM]

[hosts state after the VM is running]

| priority | vm s resource restriction is not match to hosts state describe the bug a host shows of core is used but we can t allocate more than cores for vm vm s test cases no to reproduce steps to reproduce the behavior install harvester on hosts with different resources create a vm which cpu is more than half of maximum vm goes starting state and shows error insufficient cpu expected behavior user can t know how many cpus is dedicated for node and the dashboard shows there still many resource is free to allocated for vms environment harvester iso version master head underlying infrastructure e g baremetal with dell poweredge qemu kvm additional context | 1 |
804,033 | 29,326,866,319 | IssuesEvent | 2023-05-26 00:32:43 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Rename pushdown related internal names | kind/bug area/ysql priority/medium | Jira Link: [DB-6627](https://yugabyte.atlassian.net/browse/DB-6627)
### Description
Some pushdown vars/funcs could be better named:
- [x] some "params" to "colrefs"
- [x] some "qual" to "quals"
- [ ] some "remote" to "pushdown"
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-6627]: https://yugabyte.atlassian.net/browse/DB-6627?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [YSQL] Rename pushdown related internal names - Jira Link: [DB-6627](https://yugabyte.atlassian.net/browse/DB-6627)
### Description
Some pushdown vars/funcs could be better named:
- [x] some "params" to "colrefs"
- [x] some "qual" to "quals"
- [ ] some "remote" to "pushdown"
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-6627]: https://yugabyte.atlassian.net/browse/DB-6627?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | rename pushdown related internal names jira link description some pushdown vars funcs could be better named some params to colrefs some qual to quals some remote to pushdown warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information | 1 |
526 | 2,497,833,277 | IssuesEvent | 2015-01-07 11:25:42 | HubTurbo/HubTurbo | https://api.github.com/repos/HubTurbo/HubTurbo | opened | Make '<' the default 'updated' qualifier | aspect-ui feature-filters priority.low status.accepted type.enhancement | I guess this will be the most used one, so it helps if it can be the default | 1.0 | Make '<' the default 'updated' qualifier - I guess this will be the most used one, so it helps if it can be the default | priority | make the default updated qualifier i guess this will be the most used one so it helps if it can be the default | 1 |
170,948 | 20,888,435,902 | IssuesEvent | 2022-03-23 08:31:42 | kedacore/keda-docs | https://api.github.com/repos/kedacore/keda-docs | opened | CVE-2021-44906 (Medium) detected in minimist-1.2.5.tgz | security vulnerability | ## CVE-2021-44906 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/netlify-cli/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- netlify-cli-9.13.3.tgz (Root Library)
- :x: **minimist-1.2.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kedacore/keda-docs/commit/689fc2f263626cb2198644b68f537463821c0e81">689fc2f263626cb2198644b68f537463821c0e81</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-44906">https://nvd.nist.gov/vuln/detail/CVE-2021-44906</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: BumperLane.Public.Service.Contracts - 0.23.35.214-prerelease;cloudscribe.templates - 5.2.0;Virteom.Tenant.Mobile.Bluetooth - 0.21.29.159-prerelease;ShowingVault.DotNet.Sdk - 0.13.41.190-prerelease;Envisia.DotNet.Templates - 3.0.1;Yarnpkg.Yarn - 0.26.1;Virteom.Tenant.Mobile.Framework.UWP - 0.20.41.103-prerelease;Virteom.Tenant.Mobile.Framework.iOS - 0.20.41.103-prerelease;BumperLane.Public.Api.V2.ClientModule - 0.23.35.214-prerelease;VueJS.NetCore - 1.1.1;Dianoga - 4.0.0,3.0.0-RC02;Virteom.Tenant.Mobile.Bluetooth.iOS - 0.20.41.103-prerelease;Virteom.Public.Utilities - 0.23.37.212-prerelease;Indianadavy.VueJsWebAPITemplate.CSharp - 1.0.1;NorDroN.AngularTemplate - 0.1.6;Virteom.Tenant.Mobile.Framework - 0.21.29.159-prerelease;Virteom.Tenant.Mobile.Bluetooth.Android - 0.20.41.103-prerelease;z4a-dotnet-scaffold - 1.0.0.2;Raml.Parser - 1.0.7;CoreVueWebTest - 3.0.101;dotnetng.template - 1.0.0.4;SitecoreMaster.TrueDynamicPlaceholders - 1.0.3;Virteom.Tenant.Mobile.Framework.Android - 0.20.41.103-prerelease;Fable.Template.Elmish.React - 0.1.6;BlazorPolyfill.Build - 6.0.100.2;Fable.Snowpack.Template - 2.1.0;BumperLane.Public.Api.Client - 0.23.35.214-prerelease;Yarn.MSBuild - 0.22.0,0.24.6;Blazor.TailwindCSS.BUnit - 1.0.2;Bridge.AWS - 0.3.30.36;tslint - 5.6.0;SAFE.Template - 3.0.1;GR.PageRender.Razor - 1.8.0;MIDIator.WebClient - 1.0.105</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-44906 (Medium) detected in minimist-1.2.5.tgz - ## CVE-2021-44906 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/netlify-cli/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- netlify-cli-9.13.3.tgz (Root Library)
- :x: **minimist-1.2.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kedacore/keda-docs/commit/689fc2f263626cb2198644b68f537463821c0e81">689fc2f263626cb2198644b68f537463821c0e81</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-44906">https://nvd.nist.gov/vuln/detail/CVE-2021-44906</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: BumperLane.Public.Service.Contracts - 0.23.35.214-prerelease;cloudscribe.templates - 5.2.0;Virteom.Tenant.Mobile.Bluetooth - 0.21.29.159-prerelease;ShowingVault.DotNet.Sdk - 0.13.41.190-prerelease;Envisia.DotNet.Templates - 3.0.1;Yarnpkg.Yarn - 0.26.1;Virteom.Tenant.Mobile.Framework.UWP - 0.20.41.103-prerelease;Virteom.Tenant.Mobile.Framework.iOS - 0.20.41.103-prerelease;BumperLane.Public.Api.V2.ClientModule - 0.23.35.214-prerelease;VueJS.NetCore - 1.1.1;Dianoga - 4.0.0,3.0.0-RC02;Virteom.Tenant.Mobile.Bluetooth.iOS - 0.20.41.103-prerelease;Virteom.Public.Utilities - 0.23.37.212-prerelease;Indianadavy.VueJsWebAPITemplate.CSharp - 1.0.1;NorDroN.AngularTemplate - 0.1.6;Virteom.Tenant.Mobile.Framework - 0.21.29.159-prerelease;Virteom.Tenant.Mobile.Bluetooth.Android - 0.20.41.103-prerelease;z4a-dotnet-scaffold - 1.0.0.2;Raml.Parser - 1.0.7;CoreVueWebTest - 3.0.101;dotnetng.template - 1.0.0.4;SitecoreMaster.TrueDynamicPlaceholders - 1.0.3;Virteom.Tenant.Mobile.Framework.Android - 0.20.41.103-prerelease;Fable.Template.Elmish.React - 0.1.6;BlazorPolyfill.Build - 6.0.100.2;Fable.Snowpack.Template - 2.1.0;BumperLane.Public.Api.Client - 0.23.35.214-prerelease;Yarn.MSBuild - 0.22.0,0.24.6;Blazor.TailwindCSS.BUnit - 1.0.2;Bridge.AWS - 0.3.30.36;tslint - 5.6.0;SAFE.Template - 3.0.1;GR.PageRender.Razor - 1.8.0;MIDIator.WebClient - 1.0.105</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in minimist tgz cve medium severity vulnerability vulnerable library minimist tgz parse argument options library home page a href path to dependency file package json path to vulnerable library node modules netlify cli node modules minimist package json dependency hierarchy netlify cli tgz root library x minimist tgz vulnerable library found in head commit a href found in base branch main vulnerability details minimist is vulnerable to prototype pollution via file index js function setkey lines publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bumperlane public service contracts prerelease cloudscribe templates virteom tenant mobile bluetooth prerelease showingvault dotnet sdk prerelease envisia dotnet templates yarnpkg yarn virteom tenant mobile framework uwp prerelease virteom tenant mobile framework ios prerelease bumperlane public api clientmodule prerelease vuejs netcore dianoga virteom tenant mobile bluetooth ios prerelease virteom public utilities prerelease indianadavy vuejswebapitemplate csharp nordron angulartemplate virteom tenant mobile framework prerelease virteom tenant mobile bluetooth android prerelease dotnet scaffold raml parser corevuewebtest dotnetng template sitecoremaster truedynamicplaceholders virteom tenant mobile framework android prerelease fable template elmish react blazorpolyfill build fable snowpack template bumperlane public api client prerelease yarn msbuild blazor tailwindcss bunit bridge aws tslint safe template gr pagerender razor midiator webclient step up your open source security game with whitesource | 0 |
366,868 | 25,705,851,506 | IssuesEvent | 2022-12-07 00:35:08 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | SDN Oncall Documentation - General On-call Documentation | documentation team/DXC ops and shared services NSXT/SDN | **Describe the issue**
Similar to `OCP4/On-Call.md` for regular Openshift 4 clusters, we need a matching document created for Openshift clusters built on NSX-T. The premise will be that the new document will link back to the OCP4 document since most if not all info will still apply, and then add new items specific to NSX-T in the new document.
**Additional context**
Link to GitHub PR: <https://github.com/bcgov-c/advsol-docs/pull/246>
**How does this benefit the users of our platform?**
Proper documentation ensures all Platform Operations on-call staff have access to essential information to best help them quickly assess issues on NSX-backed Openshift clusters.
**Definition of Done**
- [x] create initial PR with starting content.
- [x] promote PR for internal review/approval.
- [x] update PR with requested changes as appropriate.
- [x] merge PR when approved.
| 1.0 | SDN Oncall Documentation - General On-call Documentation - **Describe the issue**
Similar to `OCP4/On-Call.md` for regular Openshift 4 clusters, we need a matching document created for Openshift clusters built on NSX-T. The premise will be that the new document will link back to the OCP4 document since most if not all info will still apply, and then add new items specific to NSX-T in the new document.
**Additional context**
Link to GitHub PR: <https://github.com/bcgov-c/advsol-docs/pull/246>
**How does this benefit the users of our platform?**
Proper documentation ensures all Platform Operations on-call staff have access to essential information to best help them quickly assess issues on NSX-backed Openshift clusters.
**Definition of Done**
- [x] create initial PR with starting content.
- [x] promote PR for internal review/approval.
- [x] update PR with requested changes as appropriate.
- [x] merge PR when approved.
| non_priority | sdn oncall documentation general on call documentation describe the issue similar to on call md for regular openshift clusters we need a matching document created for openshift clusters built on nsx t the premise will be that the new document will link back to the document since most if not all info will still apply and then add new items specific to nsx t in the new document additional context link to github pr how does this benefit the users of our platform proper documentation ensures all platform operations on call staff have access to essential information to best help them quickly assess issues on nsx backed openshift clusters definition of done create initial pr with starting content promote pr for internal review approval update pr with requested changes as appropriate merge pr when approved | 0 |
310,382 | 9,489,390,893 | IssuesEvent | 2019-04-22 22:23:47 | jbadlato/Markov-Rankings | https://api.github.com/repos/jbadlato/Markov-Rankings | closed | Show team's schedule when you click on team name | Priority: High Type: Feature | - [x] Create API call for a team's schedule & scores
- [x] Create new page in React router for team schedule (can this be a pop-up/new window? Overlay?)
- [x] Create React elements for team, score, game, schedule
- [x] Display home team on bottom
- [x] Don't show scores for games scheduled in the future
- [x] Mark winning team of each game | 1.0 | Show team's schedule when you click on team name - - [x] Create API call for a team's schedule & scores
- [x] Create new page in React router for team schedule (can this be a pop-up/new window? Overlay?)
- [x] Create React elements for team, score, game, schedule
- [x] Display home team on bottom
- [x] Don't show scores for games scheduled in the future
- [x] Mark winning team of each game | priority | show team s schedule when you click on team name create api call for a team s schedule scores create new page in react router for team schedule can this be a pop up new window overlay create react elements for team score game schedule display home team on bottom don t show scores for games scheduled in the future mark winning team of each game | 1 |
193,778 | 22,216,336,739 | IssuesEvent | 2022-06-08 02:19:45 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | reopened | CVE-2017-17856 (High) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2017-17856 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
kernel/bpf/verifier.c in the Linux kernel through 4.14.8 allows local users to cause a denial of service (memory corruption) or possibly have unspecified other impact by leveraging the lack of stack-pointer alignment enforcement.
<p>Publish Date: 2017-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-17856>CVE-2017-17856</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2017-17856">https://www.linuxkernelcves.com/cves/CVE-2017-17856</a></p>
<p>Release Date: 2017-12-27</p>
<p>Fix Resolution: v4.15-rc5,v4.14.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-17856 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-17856 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
kernel/bpf/verifier.c in the Linux kernel through 4.14.8 allows local users to cause a denial of service (memory corruption) or possibly have unspecified other impact by leveraging the lack of stack-pointer alignment enforcement.
<p>Publish Date: 2017-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-17856>CVE-2017-17856</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2017-17856">https://www.linuxkernelcves.com/cves/CVE-2017-17856</a></p>
<p>Release Date: 2017-12-27</p>
<p>Fix Resolution: v4.15-rc5,v4.14.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files kernel bpf verifier c kernel bpf verifier c vulnerability details kernel bpf verifier c in the linux kernel through allows local users to cause a denial of service memory corruption or possibly have unspecified other impact by leveraging the lack of stack pointer alignment enforcement publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.