Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
310,955 | 9,526,388,025 | IssuesEvent | 2019-04-28 19:36:48 | zanderbowen/AAE560-SP2019 | https://api.github.com/repos/zanderbowen/AAE560-SP2019 | closed | General Work Order Class Items | high priority stochastic | - [x] Add a column to obj.routing.Edges for TotalActual in the WorkOrder constructor method
- [x] Write a method to sum the actual delivery time and actual hours worked time. This will need to run after each cycle to update accordingly. | 1.0 | General Work Order Class Items - - [x] Add a column to obj.routing.Edges for TotalActual in the WorkOrder constructor method
- [x] Write a method to sum the actual delivery time and actual hours worked time. This will need to run after each cycle to update accordingly. | priority | general work order class items add a column to obj routing edges for totalactual in the workorder constructor method write a method to sum the actual delivery time and actual hours worked time this will need to run after each cycle to update accordingly | 1 |
606,587 | 18,765,545,118 | IssuesEvent | 2021-11-05 23:10:16 | gpbl/react-day-picker | https://api.github.com/repos/gpbl/react-day-picker | closed | v8. Markup is broken with "showOutsideDays" and "fixedWeeks" options in December 2022 | Type: Bug Priority: High | **Describe the bug**
Layout is broken with `"react-day-picker": "8.0.0-beta.29"` under certain conditions.
**To Reproduce**
reproduced here:
https://codesandbox.io/s/daypicker-basicscustomization-weeknumber-example-forked-7ej5b?file=/example.tsx
using next month button scroll to December 2022. While other months are rendered just fine, this one brakes layout
**Screenshots**

**Additional context**
used beta 29 (as 30 doesn't seemed to be published to npm at the moment of writing)
| 1.0 | v8. Markup is broken with "showOutsideDays" and "fixedWeeks" options in December 2022 - **Describe the bug**
Layout is broken with `"react-day-picker": "8.0.0-beta.29"` under certain conditions.
**To Reproduce**
reproduced here:
https://codesandbox.io/s/daypicker-basicscustomization-weeknumber-example-forked-7ej5b?file=/example.tsx
using next month button scroll to December 2022. While other months are rendered just fine, this one brakes layout
**Screenshots**

**Additional context**
used beta 29 (as 30 doesn't seemed to be published to npm at the moment of writing)
| priority | markup is broken with showoutsidedays and fixedweeks options in december describe the bug layout is broken with react day picker beta under certain conditions to reproduce reproduced here using next month button scroll to december while other months are rendered just fine this one brakes layout screenshots additional context used beta as doesn t seemed to be published to npm at the moment of writing | 1 |
512,646 | 14,906,265,190 | IssuesEvent | 2021-01-22 00:10:03 | netlify/next-on-netlify | https://api.github.com/repos/netlify/next-on-netlify | closed | _redirects sorted wrong when using catch all route and dynamic routes | priority: high type: bug | The change [#145](https://github.com/netlify/next-on-netlify/pull/145) "revert route/redirect sorting logic to static then dynamic" causes the `_redirects` file to be sorted wrong when using an optional catch all route and dynamic paths.
## How to reproduce
This problem happens when we have the following routes:
```
pages/_app.js
pages/[[...any]].js
pages/[bar]/test.js
pages/test.js
```
Clone https://github.com/amuttsch/non-routes
Execute the following command
```bash
yarn install
yarn netlify
```
### Problem
The resulting `_redirects` file looks like this on version `^2.8.3`
```
# Next-on-Netlify Redirects
/_next/data/tveFS5JTA64mmnqiHIWfA/test.json /.netlify/functions/next_test 200
/_next/data/tveFS5JTA64mmnqiHIWfA/index.json /.netlify/functions/next_any 200
/_next/data/tveFS5JTA64mmnqiHIWfA/* /.netlify/functions/next_any 200 <------ catches ":bar/test.json"
/api/hello /.netlify/functions/next_api_hello 200
/test /.netlify/functions/next_test 200
/_next/data/tveFS5JTA64mmnqiHIWfA/:bar/test.json /.netlify/functions/next_bar_test 200
/_next/image* url=:url w=:width q=:quality /.netlify/functions/next_image?url=:url&w=:width&q=:quality 200
/:bar/test /.netlify/functions/next_bar_test 200
/ /.netlify/functions/next_any 200
/_next/* /_next/:splat 200
/* /.netlify/functions/next_any 200
```
The data routes `/_next/data` are split into two groups. `/_next/data/tveFS5JTA64mmnqiHIWfA/* /.netlify/functions/next_any 200` catches any data route, but `/_next/data/tveFS5JTA64mmnqiHIWfA/:bar/test.json /.netlify/functions/next_bar_test 200` is below this redirect. This causes our application to fail to load data when the user does a client side transition.
### Expected
Previous to `2.8.3` the file looks like this:
```
# Next-on-Netlify Redirects
/_next/data/xcFf_YKEf4dQzabGKlRTV/test.json /.netlify/functions/next_test 200
/_next/data/xcFf_YKEf4dQzabGKlRTV/:bar/test.json /.netlify/functions/next_bar_test 200
/_next/data/xcFf_YKEf4dQzabGKlRTV/index.json /.netlify/functions/next_any 200
/_next/data/xcFf_YKEf4dQzabGKlRTV/* /.netlify/functions/next_any 200
/_next/image* url=:url w=:width q=:quality /.netlify/functions/next_image?url=:url&w=:width&q=:quality 200
/api/hello /.netlify/functions/next_api_hello 200
/test /.netlify/functions/next_test 200
/:bar/test /.netlify/functions/next_bar_test 200
/ /.netlify/functions/next_any 200
/_next/* /_next/:splat 200
/* /.netlify/functions/next_any 200
```
Workaround: Downgrade `next-on-netlify` to `2.8.2`. | 1.0 | _redirects sorted wrong when using catch all route and dynamic routes - The change [#145](https://github.com/netlify/next-on-netlify/pull/145) "revert route/redirect sorting logic to static then dynamic" causes the `_redirects` file to be sorted wrong when using an optional catch all route and dynamic paths.
## How to reproduce
This problem happens when we have the following routes:
```
pages/_app.js
pages/[[...any]].js
pages/[bar]/test.js
pages/test.js
```
Clone https://github.com/amuttsch/non-routes
Execute the following command
```bash
yarn install
yarn netlify
```
### Problem
The resulting `_redirects` file looks like this on version `^2.8.3`
```
# Next-on-Netlify Redirects
/_next/data/tveFS5JTA64mmnqiHIWfA/test.json /.netlify/functions/next_test 200
/_next/data/tveFS5JTA64mmnqiHIWfA/index.json /.netlify/functions/next_any 200
/_next/data/tveFS5JTA64mmnqiHIWfA/* /.netlify/functions/next_any 200 <------ catches ":bar/test.json"
/api/hello /.netlify/functions/next_api_hello 200
/test /.netlify/functions/next_test 200
/_next/data/tveFS5JTA64mmnqiHIWfA/:bar/test.json /.netlify/functions/next_bar_test 200
/_next/image* url=:url w=:width q=:quality /.netlify/functions/next_image?url=:url&w=:width&q=:quality 200
/:bar/test /.netlify/functions/next_bar_test 200
/ /.netlify/functions/next_any 200
/_next/* /_next/:splat 200
/* /.netlify/functions/next_any 200
```
The data routes `/_next/data` are split into two groups. `/_next/data/tveFS5JTA64mmnqiHIWfA/* /.netlify/functions/next_any 200` catches any data route, but `/_next/data/tveFS5JTA64mmnqiHIWfA/:bar/test.json /.netlify/functions/next_bar_test 200` is below this redirect. This causes our application to fail to load data when the user does a client side transition.
### Expected
Previous to `2.8.3` the file looks like this:
```
# Next-on-Netlify Redirects
/_next/data/xcFf_YKEf4dQzabGKlRTV/test.json /.netlify/functions/next_test 200
/_next/data/xcFf_YKEf4dQzabGKlRTV/:bar/test.json /.netlify/functions/next_bar_test 200
/_next/data/xcFf_YKEf4dQzabGKlRTV/index.json /.netlify/functions/next_any 200
/_next/data/xcFf_YKEf4dQzabGKlRTV/* /.netlify/functions/next_any 200
/_next/image* url=:url w=:width q=:quality /.netlify/functions/next_image?url=:url&w=:width&q=:quality 200
/api/hello /.netlify/functions/next_api_hello 200
/test /.netlify/functions/next_test 200
/:bar/test /.netlify/functions/next_bar_test 200
/ /.netlify/functions/next_any 200
/_next/* /_next/:splat 200
/* /.netlify/functions/next_any 200
```
Workaround: Downgrade `next-on-netlify` to `2.8.2`. | priority | redirects sorted wrong when using catch all route and dynamic routes the change revert route redirect sorting logic to static then dynamic causes the redirects file to be sorted wrong when using an optional catch all route and dynamic paths how to reproduce this problem happens when we have the following routes pages app js pages js pages test js pages test js clone execute the following command bash yarn install yarn netlify problem the resulting redirects file looks like this on version next on netlify redirects next data test json netlify functions next test next data index json netlify functions next any next data netlify functions next any catches bar test json api hello netlify functions next api hello test netlify functions next test next data bar test json netlify functions next bar test next image url url w width q quality netlify functions next image url url w width q quality bar test netlify functions next bar test netlify functions next any next next splat netlify functions next any the data routes next data are split into two groups next data netlify functions next any catches any data route but next data bar test json netlify functions next bar test is below this redirect this causes our application to fail to load data when the user does a client side transition expected previous to the file looks like this next on netlify redirects next data xcff test json netlify functions next test next data xcff bar test json netlify functions next bar test next data xcff index json netlify functions next any next data xcff netlify functions next any next image url url w width q quality netlify functions next image url url w width q quality api hello netlify functions next api hello test netlify functions next test bar test netlify functions next bar test netlify functions next any next next splat netlify functions next any workaround downgrade next on netlify to | 1 |
636,378 | 20,598,670,187 | IssuesEvent | 2022-03-05 23:02:01 | localstack/localstack | https://api.github.com/repos/localstack/localstack | closed | bug: SSL certificate issue since v0.12.18 | bug priority-high needs-triaging infra-startup networking | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
LocalStack is taking a long time to start up since updating to v0.12.18 due to timing out when attempting to pull down the local test SSL certificate. This works fine on v0.12.17.
```
localstack_1 | Starting edge router (https port 4566)...
localstack_1 | 2021-11-11T09:33:23:INFO:bootstrap.py: Execution of "load_plugin_from_path" took 2430.59ms
localstack_1 | 2021-11-11T09:33:23:INFO:bootstrap.py: Execution of "load_plugins" took 2431.15ms
localstack_1 | Waiting for all LocalStack services to be ready
localstack_1 | ...
localstack_1 | Waiting for all LocalStack services to be ready
localstack_1 | 2021-11-11T09:46:28:INFO:localstack_ext.bootstrap.install: Unable to download local test SSL certificate from https://cdn.jsdelivr.net/gh/localstack/localstack-artifacts@master/local-certs/server.key to /tmp/localstack/server.test.pem: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out')) Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 175, in _new_conn
localstack_1 | (self._dns_host, self.port), self.timeout, **extra_kw
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/connection.py", line 96, in create_connection
localstack_1 | raise err
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/connection.py", line 86, in create_connection
localstack_1 | sock.connect(sa)
localstack_1 | TimeoutError: [Errno 110] Operation timed out
localstack_1 |
localstack_1 | During handling of the above exception, another exception occurred:
localstack_1 |
localstack_1 | Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 706, in urlopen
localstack_1 | chunked=chunked,
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 382, in _make_request
localstack_1 | self._validate_conn(conn)
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 1010, in _validate_conn
localstack_1 | conn.connect()
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 358, in connect
localstack_1 | conn = self._new_conn()
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 187, in _new_conn
localstack_1 | self, "Failed to establish a new connection: %s" % e
localstack_1 | urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out
localstack_1 |
localstack_1 | During handling of the above exception, another exception occurred:
localstack_1 |
localstack_1 | Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/adapters.py", line 449, in send
localstack_1 | timeout=timeout
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 756, in urlopen
localstack_1 | method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/retry.py", line 574, in increment
localstack_1 | raise MaxRetryError(_pool, url, error or ResponseError(cause))
localstack_1 | urllib3.exceptions.MaxRetryError: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out'))
localstack_1 |
localstack_1 | During handling of the above exception, another exception occurred:
localstack_1 |
localstack_1 | Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/localstack_ext/bootstrap/install.py", line 87, in do_download
localstack_1 | download(url,target_file)
localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 1072, in download
localstack_1 | r = s.get(url, stream=True, verify=os.getenv("REQUESTS_CA_BUNDLE", verify_ssl))
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 555, in get
localstack_1 | return self.request('GET', url, **kwargs)
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 542, in request
localstack_1 | resp = self.send(prep, **send_kwargs)
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 655, in send
localstack_1 | r = adapter.send(request, **kwargs)
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/adapters.py", line 516, in send
localstack_1 | raise ConnectionError(e, request=request)
localstack_1 | requests.exceptions.ConnectionError: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out'))
```
I expect this has something to do with this line in the release notes (but not 100%)
> add startup logic to install prebuilt SSL cert if available
### Expected Behavior
LocalStack to start up in a normal amount of time
### How are you starting LocalStack?
With a docker-compose file
### Steps To Reproduce
#### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`)
Using a custom docker-compose that starts localstack up as part of a suite of services
#### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands)
N/A
### Environment
```markdown
- OS:
- LocalStack: v0.12.18 to latest
```
### Anything else?
The localstack image works fine on my local machine, but does not work when running on the CI server. The CI server is more locked down and lacks external connectivity without a proxy, so this may be causing the issue. | 1.0 | bug: SSL certificate issue since v0.12.18 - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
LocalStack is taking a long time to start up since updating to v0.12.18 due to timing out when attempting to pull down the local test SSL certificate. This works fine on v0.12.17.
```
localstack_1 | Starting edge router (https port 4566)...
localstack_1 | 2021-11-11T09:33:23:INFO:bootstrap.py: Execution of "load_plugin_from_path" took 2430.59ms
localstack_1 | 2021-11-11T09:33:23:INFO:bootstrap.py: Execution of "load_plugins" took 2431.15ms
localstack_1 | Waiting for all LocalStack services to be ready
localstack_1 | ...
localstack_1 | Waiting for all LocalStack services to be ready
localstack_1 | 2021-11-11T09:46:28:INFO:localstack_ext.bootstrap.install: Unable to download local test SSL certificate from https://cdn.jsdelivr.net/gh/localstack/localstack-artifacts@master/local-certs/server.key to /tmp/localstack/server.test.pem: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out')) Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 175, in _new_conn
localstack_1 | (self._dns_host, self.port), self.timeout, **extra_kw
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/connection.py", line 96, in create_connection
localstack_1 | raise err
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/connection.py", line 86, in create_connection
localstack_1 | sock.connect(sa)
localstack_1 | TimeoutError: [Errno 110] Operation timed out
localstack_1 |
localstack_1 | During handling of the above exception, another exception occurred:
localstack_1 |
localstack_1 | Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 706, in urlopen
localstack_1 | chunked=chunked,
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 382, in _make_request
localstack_1 | self._validate_conn(conn)
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 1010, in _validate_conn
localstack_1 | conn.connect()
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 358, in connect
localstack_1 | conn = self._new_conn()
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 187, in _new_conn
localstack_1 | self, "Failed to establish a new connection: %s" % e
localstack_1 | urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out
localstack_1 |
localstack_1 | During handling of the above exception, another exception occurred:
localstack_1 |
localstack_1 | Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/adapters.py", line 449, in send
localstack_1 | timeout=timeout
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 756, in urlopen
localstack_1 | method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/retry.py", line 574, in increment
localstack_1 | raise MaxRetryError(_pool, url, error or ResponseError(cause))
localstack_1 | urllib3.exceptions.MaxRetryError: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out'))
localstack_1 |
localstack_1 | During handling of the above exception, another exception occurred:
localstack_1 |
localstack_1 | Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/localstack_ext/bootstrap/install.py", line 87, in do_download
localstack_1 | download(url,target_file)
localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 1072, in download
localstack_1 | r = s.get(url, stream=True, verify=os.getenv("REQUESTS_CA_BUNDLE", verify_ssl))
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 555, in get
localstack_1 | return self.request('GET', url, **kwargs)
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 542, in request
localstack_1 | resp = self.send(prep, **send_kwargs)
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 655, in send
localstack_1 | r = adapter.send(request, **kwargs)
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/adapters.py", line 516, in send
localstack_1 | raise ConnectionError(e, request=request)
localstack_1 | requests.exceptions.ConnectionError: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out'))
```
I expect this has something to do with this line in the release notes (but not 100%)
> add startup logic to install prebuilt SSL cert if available
### Expected Behavior
LocalStack to start up in a normal amount of time
### How are you starting LocalStack?
With a docker-compose file
### Steps To Reproduce
#### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`)
Using a custom docker-compose that starts localstack up as part of a suite of services
#### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands)
N/A
### Environment
```markdown
- OS:
- LocalStack: v0.12.18 to latest
```
### Anything else?
The localstack image works fine on my local machine, but does not work when running on the CI server. The CI server is more locked down and lacks external connectivity without a proxy, so this may be causing the issue. | priority | bug ssl certificate issue since is there an existing issue for this i have searched the existing issues current behavior localstack is taking a long time to start up since updating to due to timing out when attempting to pull down the local test ssl certificate this works fine on localstack starting edge router https port localstack info bootstrap py execution of load plugin from path took localstack info bootstrap py execution of load plugins took localstack waiting for all localstack services to be ready localstack localstack waiting for all localstack services to be ready localstack info localstack ext bootstrap install unable to download local test ssl certificate from to tmp localstack server test pem myhttpsconnectionpool host cdn jsdelivr net port max retries exceeded with url gh localstack localstack artifacts master local certs server key caused by newconnectionerror failed to establish a new connection operation timed out traceback most recent call last localstack file opt code localstack venv lib site packages connection py line in new conn localstack self dns host self port self timeout extra kw localstack file opt code localstack venv lib site packages util connection py line in create connection localstack raise err localstack file opt code localstack venv lib site packages util connection py line in create connection localstack sock connect sa localstack timeouterror operation timed out localstack localstack during handling of the above exception another exception occurred localstack localstack traceback most recent call last localstack file opt code localstack venv lib site packages connectionpool py line in urlopen localstack chunked chunked localstack file opt code localstack venv lib site packages connectionpool py line in make request localstack self validate conn conn localstack file opt code localstack venv lib site packages connectionpool py line in validate conn localstack conn connect localstack file opt code localstack venv lib site packages connection py line in connect localstack conn self new conn localstack file opt code localstack venv lib site packages connection py line in new conn localstack self failed to establish a new connection s e localstack exceptions newconnectionerror failed to establish a new connection operation timed out localstack localstack during handling of the above exception another exception occurred localstack localstack traceback most recent call last localstack file opt code localstack venv lib site packages requests adapters py line in send localstack timeout timeout localstack file opt code localstack venv lib site packages connectionpool py line in urlopen localstack method url error e pool self stacktrace sys exc info localstack file opt code localstack venv lib site packages util retry py line in increment localstack raise maxretryerror pool url error or responseerror cause localstack exceptions maxretryerror myhttpsconnectionpool host cdn jsdelivr net port max retries exceeded with url gh localstack localstack artifacts master local certs server key caused by newconnectionerror failed to establish a new connection operation timed out localstack localstack during handling of the above exception another exception occurred localstack localstack traceback most recent call last localstack file opt code localstack venv lib site packages localstack ext bootstrap install py line in do download localstack download url target file localstack file opt code localstack localstack utils common py line in download localstack r s get url stream true verify os getenv requests ca bundle verify ssl localstack file opt code localstack venv lib site packages requests sessions py line in get localstack return self request get url kwargs localstack file opt code localstack venv lib site packages requests sessions py line in request localstack resp self send prep send kwargs localstack file opt code localstack venv lib site packages requests sessions py line in send localstack r adapter send request kwargs localstack file opt code localstack venv lib site packages requests adapters py line in send localstack raise connectionerror e request request localstack requests exceptions connectionerror myhttpsconnectionpool host cdn jsdelivr net port max retries exceeded with url gh localstack localstack artifacts master local certs server key caused by newconnectionerror failed to establish a new connection operation timed out i expect this has something to do with this line in the release notes but not add startup logic to install prebuilt ssl cert if available expected behavior localstack to start up in a normal amount of time how are you starting localstack with a docker compose file steps to reproduce how are you starting localstack e g bin localstack command arguments or docker compose yml using a custom docker compose that starts localstack up as part of a suite of services client commands e g aws sdk code snippet or sequence of awslocal commands n a environment markdown os localstack to latest anything else the localstack image works fine on my local machine but does not work when running on the ci server the ci server is more locked down and lacks external connectivity without a proxy so this may be causing the issue | 1 |
451,235 | 13,031,722,226 | IssuesEvent | 2020-07-28 02:06:56 | scprogramming/Open-Source-Scan | https://api.github.com/repos/scprogramming/Open-Source-Scan | opened | Add a way to clear scans and all non cpe/cve data | Analysis High Priority | I should have a way to delete scans, clear all scans, projects, etc to preserve cpe/cve data, but allow for clearning unneeded data. | 1.0 | Add a way to clear scans and all non cpe/cve data - I should have a way to delete scans, clear all scans, projects, etc to preserve cpe/cve data, but allow for clearning unneeded data. | priority | add a way to clear scans and all non cpe cve data i should have a way to delete scans clear all scans projects etc to preserve cpe cve data but allow for clearning unneeded data | 1 |
236,909 | 7,753,586,792 | IssuesEvent | 2018-05-31 01:33:49 | Gloirin/m2gTest | https://api.github.com/repos/Gloirin/m2gTest | closed | 0006634:
custom fields missing in XLS export | Addressbook bug high priority | **Reported by cweiss on 18 Jun 2012 07:49**
**Version:** Milan (2012-03-3)
custom fields missing in XLS export
| 1.0 | 0006634:
custom fields missing in XLS export - **Reported by cweiss on 18 Jun 2012 07:49**
**Version:** Milan (2012-03-3)
custom fields missing in XLS export
| priority | custom fields missing in xls export reported by cweiss on jun version milan custom fields missing in xls export | 1 |
718,537 | 24,721,619,448 | IssuesEvent | 2022-10-20 11:08:43 | harvester/harvester | https://api.github.com/repos/harvester/harvester | closed | [FEATURE]Dedicated storage network | kind/enhancement area/ui area/installer area/network priority/0 area/storage highlight blocker area/longhorn-related require-ui/small | We would like to have a dedicated storage network interface that can be specified (either as single nic or bonded NICs) to separate the storage traffic from the overlay traffic.
Depends on https://github.com/longhorn/longhorn/issues/2285 and #1048 | 1.0 | [FEATURE]Dedicated storage network - We would like to have a dedicated storage network interface that can be specified (either as single nic or bonded NICs) to separate the storage traffic from the overlay traffic.
Depends on https://github.com/longhorn/longhorn/issues/2285 and #1048 | priority | dedicated storage network we would like to have a dedicated storage network interface that can be specified either as single nic or bonded nics to separate the storage traffic from the overlay traffic depends on and | 1 |
212,122 | 7,228,799,288 | IssuesEvent | 2018-02-11 13:41:13 | allure-framework/allure2 | https://api.github.com/repos/allure-framework/allure2 | closed | Packages tab is not shown in demo report due to invalid initializer | priority:high theme:ui type:bug work:review | [//]: # (
. Note: for support questions, please use Stackoverflow or Gitter**.
. This repository's issues are reserved for feature requests and bug reports.
.
. In case of any problems with Allure Jenkins plugin** please use the following repository
. to create an issue: https://github.com/jenkinsci/allure-plugin/issues
.
. Make sure you have a clear name for your issue. The name should start with a capital
. letter and no dot is required in the end of the sentence. An example of good issue names:
.
. - The report is broken in IE11
. - Add an ability to disable default plugins
. - Support emoji in test descriptions
)
#### I'm submitting a ...
- [x] bug report
- [ ] feature request
- [ ] support request => Please do not submit support request here, see note at the top of this template.
#### What is the current behavior?

#### If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem
Screenshot is attached.
#### What is the expected behavior?
Correct identifier, Packages tab is shown
#### What is the motivation / use case for changing the behavior?
#### Please tell us about your environment:
| Allure version | 2.2.0 |
| --- | --- |
| Test framework | testng@6.8 |
| Allure adaptor | allure-testng@2.0-BETA11 |
| Generate report using | allure-maven@2.18 |
#### Other information
[//]: # (
. e.g. detailed explanation, stacktraces, related issues, suggestions
. how to fix, links for us to have more context, eg. Stackoverflow, Gitter etc
)
<!-- Love allure-report? Please consider supporting our collective:
👉 https://opencollective.com/allure-report/donate --> | 1.0 | Packages tab is not shown in demo report due to invalid initializer - [//]: # (
. Note: for support questions, please use Stackoverflow or Gitter**.
. This repository's issues are reserved for feature requests and bug reports.
.
. In case of any problems with Allure Jenkins plugin** please use the following repository
. to create an issue: https://github.com/jenkinsci/allure-plugin/issues
.
. Make sure you have a clear name for your issue. The name should start with a capital
. letter and no dot is required in the end of the sentence. An example of good issue names:
.
. - The report is broken in IE11
. - Add an ability to disable default plugins
. - Support emoji in test descriptions
)
#### I'm submitting a ...
- [x] bug report
- [ ] feature request
- [ ] support request => Please do not submit support request here, see note at the top of this template.
#### What is the current behavior?

#### If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem
Screenshot is attached.
#### What is the expected behavior?
Correct identifier, Packages tab is shown
#### What is the motivation / use case for changing the behavior?
#### Please tell us about your environment:
| Allure version | 2.2.0 |
| --- | --- |
| Test framework | testng@6.8 |
| Allure adaptor | allure-testng@2.0-BETA11 |
| Generate report using | allure-maven@2.18 |
#### Other information
[//]: # (
. e.g. detailed explanation, stacktraces, related issues, suggestions
. how to fix, links for us to have more context, eg. Stackoverflow, Gitter etc
)
<!-- Love allure-report? Please consider supporting our collective:
👉 https://opencollective.com/allure-report/donate --> | priority | packages tab is not shown in demo report due to invalid initializer note for support questions please use stackoverflow or gitter this repository s issues are reserved for feature requests and bug reports in case of any problems with allure jenkins plugin please use the following repository to create an issue make sure you have a clear name for your issue the name should start with a capital letter and no dot is required in the end of the sentence an example of good issue names the report is broken in add an ability to disable default plugins support emoji in test descriptions i m submitting a bug report feature request support request please do not submit support request here see note at the top of this template what is the current behavior if the current behavior is a bug please provide the steps to reproduce and if possible a minimal demo of the problem screenshot is attached what is the expected behavior correct identifier packages tab is shown what is the motivation use case for changing the behavior please tell us about your environment allure version test framework testng allure adaptor allure testng generate report using allure maven other information e g detailed explanation stacktraces related issues suggestions how to fix links for us to have more context eg stackoverflow gitter etc love allure report please consider supporting our collective 👉 | 1 |
628,798 | 20,014,459,241 | IssuesEvent | 2022-02-01 10:35:21 | kubermatic/kubeone | https://api.github.com/repos/kubermatic/kubeone | opened | Test Cilium in kube-proxy mode -- Test Release 1.4 | priority/high sig/cluster-management | Instructions:
* Download the latest KubeOne 1.4.0 release candidate
* Follow the [Create a Kubernetes cluster tutorial](https://docs.kubermatic.com/kubeone/master/tutorials/creating_clusters/) to create your cluster
* Make sure to have Cilium enabled:
```yaml
clusterNetwork:
cni:
cilium:
enableHubble: true
```
* Wait for machine-controller-managed nodes to join the cluster
* Ensure all pods are Running
* Create a LoadBalancer Service and point to some test pod (e.g. Nginx pod). The LB should be reachable and serve the content as expected
* Ensure Hubble is running and reachable (you might need to port-forward to it)
This test can be done on a single cloud provider, on a single operating system, with a single Kubernetes version (e.g. 1.23.3 on Ubuntu on AWS). | 1.0 | Test Cilium in kube-proxy mode -- Test Release 1.4 - Instructions:
* Download the latest KubeOne 1.4.0 release candidate
* Follow the [Create a Kubernetes cluster tutorial](https://docs.kubermatic.com/kubeone/master/tutorials/creating_clusters/) to create your cluster
* Make sure to have Cilium enabled:
```yaml
clusterNetwork:
cni:
cilium:
enableHubble: true
```
* Wait for machine-controller-managed nodes to join the cluster
* Ensure all pods are Running
* Create a LoadBalancer Service and point to some test pod (e.g. Nginx pod). The LB should be reachable and serve the content as expected
* Ensure Hubble is running and reachable (you might need to port-forward to it)
This test can be done on a single cloud provider, on a single operating system, with a single Kubernetes version (e.g. 1.23.3 on Ubuntu on AWS). | priority | test cilium in kube proxy mode test release instructions download the latest kubeone release candidate follow the to create your cluster make sure to have cilium enabled yaml clusternetwork cni cilium enablehubble true wait for machine controller managed nodes to join the cluster ensure all pods are running create a loadbalancer service and point to some test pod e g nginx pod the lb should be reachable and serve the content as expected ensure hubble is running and reachable you might need to port forward to it this test can be done on a single cloud provider on a single operating system with a single kubernetes version e g on ubuntu on aws | 1 |
86,263 | 3,704,411,907 | IssuesEvent | 2016-03-01 00:04:24 | Macainian/Django-Maced | https://api.github.com/repos/Macainian/Django-Maced | opened | When items are modified, they only propagate select name/value changes. | Bug High Priority | This is an extension of issue #52 | 1.0 | When items are modified, they only propagate select name/value changes. - This is an extension of issue #52 | priority | when items are modified they only propagate select name value changes this is an extension of issue | 1 |
441,199 | 12,709,385,628 | IssuesEvent | 2020-06-23 12:16:34 | AlonDiskin/Visuals | https://api.github.com/repos/AlonDiskin/Visuals | closed | Videos and recycle bin state update error | bug high priority | reproduce steps:
- move some videos to recycle bin
- restore all trash items from recycle bin browser screen
- trash more videos from videos browser
- videos browser not update items as removed
- the now empty recycle bin browser do not show the new items as expected | 1.0 | Videos and recycle bin state update error - reproduce steps:
- move some videos to recycle bin
- restore all trash items from recycle bin browser screen
- trash more videos from videos browser
- videos browser not update items as removed
- the now empty recycle bin browser do not show the new items as expected | priority | videos and recycle bin state update error reproduce steps move some videos to recycle bin restore all trash items from recycle bin browser screen trash more videos from videos browser videos browser not update items as removed the now empty recycle bin browser do not show the new items as expected | 1 |
9,932 | 2,608,937,862 | IssuesEvent | 2015-02-26 11:05:57 | CSSE1001/MyPyTutor | https://api.github.com/repos/CSSE1001/MyPyTutor | closed | Verify installer behaviour for directories with spaces on Windows | bug priority: high | The installer (new and old) need to be checked on Windows to ensure that they correctly handle directories with spaces.
I've had trouble with this issue in the past, especially with the use of `os.execv`; Windows sometimes seems to like screwing up the components of the vector, even though they're given as a list. | 1.0 | Verify installer behaviour for directories with spaces on Windows - The installer (new and old) need to be checked on Windows to ensure that they correctly handle directories with spaces.
I've had trouble with this issue in the past, especially with the use of `os.execv`; Windows sometimes seems to like screwing up the components of the vector, even though they're given as a list. | priority | verify installer behaviour for directories with spaces on windows the installer new and old need to be checked on windows to ensure that they correctly handle directories with spaces i ve had trouble with this issue in the past especially with the use of os execv windows sometimes seems to like screwing up the components of the vector even though they re given as a list | 1 |
144,859 | 5,547,067,858 | IssuesEvent | 2017-03-23 03:45:42 | CS2103JAN2017-T16-B4/main | https://api.github.com/repos/CS2103JAN2017-T16-B4/main | closed | As a user I want to edit the deadline, name and schedule of any task | priority.high type.story | Update any task as required, as well as marking it as done and deleting it | 1.0 | As a user I want to edit the deadline, name and schedule of any task - Update any task as required, as well as marking it as done and deleting it | priority | as a user i want to edit the deadline name and schedule of any task update any task as required as well as marking it as done and deleting it | 1 |
266,999 | 8,378,159,735 | IssuesEvent | 2018-10-06 11:03:58 | rit-sse/OneRepoToRuleThemAll | https://api.github.com/repos/rit-sse/OneRepoToRuleThemAll | closed | Pin Down Node Version | High Priority | In our `Dockerfile` we say `FROM node` which will install the latest node version. It appears that this is installing node v10.
We should pin down the version (eg. `FROM node:carbon`) to the LTS release so that we're running on a consistent node version, not running on odd numbered Node releases, and can have a consistent upgrade process for our dependencies (not having them break under us).
Node v10 will become LTS this month (October 2018), so we can wait until it does to pin down the version.
We can also delete the `Dockerfile.edge` file as it's out-of-date and no longer needed. | 1.0 | Pin Down Node Version - In our `Dockerfile` we say `FROM node` which will install the latest node version. It appears that this is installing node v10.
We should pin down the version (eg. `FROM node:carbon`) to the LTS release so that we're running on a consistent node version, not running on odd numbered Node releases, and can have a consistent upgrade process for our dependencies (not having them break under us).
Node v10 will become LTS this month (October 2018), so we can wait until it does to pin down the version.
We can also delete the `Dockerfile.edge` file as it's out-of-date and no longer needed. | priority | pin down node version in our dockerfile we say from node which will install the latest node version it appears that this is installing node we should pin down the version eg from node carbon to the lts release so that we re running on a consistent node version not running on odd numbered node releases and can have a consistent upgrade process for our dependencies not having them break under us node will become lts this month october so we can wait until it does to pin down the version we can also delete the dockerfile edge file as it s out of date and no longer needed | 1 |
732,954 | 25,282,044,689 | IssuesEvent | 2022-11-16 16:25:50 | leka/LekaOS | https://api.github.com/repos/leka/LekaOS | closed | [Story] - LekaOS BLE v1.0.0 | 01 - type: story 90 - priority: high | # Introduction
BLE communication with the iPad is one of the most important feature of LekaOS. Without it the robot is useless and its presence allow us to:
- control the robot
- debug/log what's going on
- tests new functionalities (FOTA, scheduling, etc.)
## Roadmap
1. Setup a basic working example using mbed
1. Test the integration with timed tasks and scheduling
1. Implement [Leka Communication Specifications](https://github.com/leka/LKAlphaComSpecs)
## 1. Setup a basic working example using mbed
This first step is really easy: use the examples from mbed-os to create a simple working example of BLE for Leka to pave the way for future developments
This must include:
- [x] read, write and notifications services/characteristics
- [x] the use of predefined services (battery) and proprietary services (temperature, firmware version, basic command)
## 2. Test integration with timed tasks and scheduling
Through out our development we will need to start tasks and stop them. Gaining a deeper understanding about how this will work between BLE and mbed os is paramount as all future development will rely on that.
More information in #1 - https://github.com/leka/LekaOS/issues/1#issuecomment-557007922
This must include:
- [x] starting a long running thread
- [x] stoping a long running thread anytime
- [x] starting again from the beginning
- [x] starting again where we left off (if possible)
## 3. Implement Leka Communication Specifications
We'll review, update and implement the Leka Communication Specifications in parallel with the iOS app.
The main goal is to have the two perfectly synchronised.
This must include:
- [x] implement all the services/characteristics
- [x] implement the command analyser
- [x] a lot of testing
## 4. Mandatory commands
- [x] implement reset command and test in case of infinite loop | 1.0 | [Story] - LekaOS BLE v1.0.0 - # Introduction
BLE communication with the iPad is one of the most important feature of LekaOS. Without it the robot is useless and its presence allow us to:
- control the robot
- debug/log what's going on
- tests new functionalities (FOTA, scheduling, etc.)
## Roadmap
1. Setup a basic working example using mbed
1. Test the integration with timed tasks and scheduling
1. Implement [Leka Communication Specifications](https://github.com/leka/LKAlphaComSpecs)
## 1. Setup a basic working example using mbed
This first step is really easy: use the examples from mbed-os to create a simple working example of BLE for Leka to pave the way for future developments
This must include:
- [x] read, write and notifications services/characteristics
- [x] the use of predefined services (battery) and proprietary services (temperature, firmware version, basic command)
## 2. Test integration with timed tasks and scheduling
Through out our development we will need to start tasks and stop them. Gaining a deeper understanding about how this will work between BLE and mbed os is paramount as all future development will rely on that.
More information in #1 - https://github.com/leka/LekaOS/issues/1#issuecomment-557007922
This must include:
- [x] starting a long running thread
- [x] stoping a long running thread anytime
- [x] starting again from the beginning
- [x] starting again where we left off (if possible)
## 3. Implement Leka Communication Specifications
We'll review, update and implement the Leka Communication Specifications in parallel with the iOS app.
The main goal is to have the two perfectly synchronised.
This must include:
- [x] implement all the services/characteristics
- [x] implement the command analyser
- [x] a lot of testing
## 4. Mandatory commands
- [x] implement reset command and test in case of infinite loop | priority | lekaos ble introduction ble communication with the ipad is one of the most important feature of lekaos without it the robot is useless and its presence allow us to control the robot debug log what s going on tests new functionalities fota scheduling etc roadmap setup a basic working example using mbed test the integration with timed tasks and scheduling implement setup a basic working example using mbed this first step is really easy use the examples from mbed os to create a simple working example of ble for leka to pave the way for future developments this must include read write and notifications services characteristics the use of predefined services battery and proprietary services temperature firmware version basic command test integration with timed tasks and scheduling through out our development we will need to start tasks and stop them gaining a deeper understanding about how this will work between ble and mbed os is paramount as all future development will rely on that more information in this must include starting a long running thread stoping a long running thread anytime starting again from the beginning starting again where we left off if possible implement leka communication specifications we ll review update and implement the leka communication specifications in parallel with the ios app the main goal is to have the two perfectly synchronised this must include implement all the services characteristics implement the command analyser a lot of testing mandatory commands implement reset command and test in case of infinite loop | 1 |
788,885 | 27,771,807,270 | IssuesEvent | 2023-03-16 14:53:35 | janus-idp/software-templates | https://api.github.com/repos/janus-idp/software-templates | opened | Various updates in the 6 Janus GPTs | kind/enhancement priority/high | 1. Rename the GPTs titles.
ie
From: `.NET Frontend Golden Path Template`
To: `Create a .NET Frontend application with CI/CD`
2. Show the CI step before the CD step. The CI step will become step 3 and Argo will be step 4.
3. Update the current description in the CI section.
From: `This action will create a simple CI based on chosen method`
To: `This action will create a CI pipeline for your application based on chosen method`
4. In the ArgoCD step:
- Select Quay registry as default
- Move the image url before the namespace
5. Make sure the ports default are correct. It seems that most templates are using 5000 as default which is incorrect. | 1.0 | Various updates in the 6 Janus GPTs - 1. Rename the GPTs titles.
ie
From: `.NET Frontend Golden Path Template`
To: `Create a .NET Frontend application with CI/CD`
2. Show the CI step before the CD step. The CI step will become step 3 and Argo will be step 4.
3. Update the current description in the CI section.
From: `This action will create a simple CI based on chosen method`
To: `This action will create a CI pipeline for your application based on chosen method`
4. In the ArgoCD step:
- Select Quay registry as default
- Move the image url before the namespace
5. Make sure the ports default are correct. It seems that most templates are using 5000 as default which is incorrect. | priority | various updates in the janus gpts rename the gpts titles ie from net frontend golden path template to create a net frontend application with ci cd show the ci step before the cd step the ci step will become step and argo will be step update the current description in the ci section from this action will create a simple ci based on chosen method to this action will create a ci pipeline for your application based on chosen method in the argocd step select quay registry as default move the image url before the namespace make sure the ports default are correct it seems that most templates are using as default which is incorrect | 1 |
197,744 | 6,963,453,986 | IssuesEvent | 2017-12-08 17:26:59 | python/mypy | https://api.github.com/repos/python/mypy | closed | Slow incremental run when single module has errors | bug priority-0-high topic-daemon topic-incremental | `mypy -i` is slower than expected in the following scenario:
* Check out fresh mypy repository
* Create tiny module `mypy.x` that generates an error: `echo '1 + ""' > mypy/x.py`
* Run `mypy -i mypy`
* This takes a while, generates an error for `mypy/x.py` as expected
* Run `mypy -i mypy`
* This is as slow as the previous run, but caching should make this much faster
| 1.0 | Slow incremental run when single module has errors - `mypy -i` is slower than expected in the following scenario:
* Check out fresh mypy repository
* Create tiny module `mypy.x` that generates an error: `echo '1 + ""' > mypy/x.py`
* Run `mypy -i mypy`
* This takes a while, generates an error for `mypy/x.py` as expected
* Run `mypy -i mypy`
* This is as slow as the previous run, but caching should make this much faster
| priority | slow incremental run when single module has errors mypy i is slower than expected in the following scenario check out fresh mypy repository create tiny module mypy x that generates an error echo mypy x py run mypy i mypy this takes a while generates an error for mypy x py as expected run mypy i mypy this is as slow as the previous run but caching should make this much faster | 1 |
529,381 | 15,387,560,930 | IssuesEvent | 2021-03-03 09:41:22 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | Coverity - High Impact Outstanding issues in MDEvent files | Framework High Priority Stale | This issue was originally [TRAC 9938](http://trac.mantidproject.org/mantid/ticket/9938)
There are 17 Coverity high impact outstanding issued in the `MDEvent` module. Distribute as you see fit.
| 1.0 | Coverity - High Impact Outstanding issues in MDEvent files - This issue was originally [TRAC 9938](http://trac.mantidproject.org/mantid/ticket/9938)
There are 17 Coverity high impact outstanding issued in the `MDEvent` module. Distribute as you see fit.
| priority | coverity high impact outstanding issues in mdevent files this issue was originally there are coverity high impact outstanding issued in the mdevent module distribute as you see fit | 1 |
509,616 | 14,740,532,189 | IssuesEvent | 2021-01-07 09:14:23 | canonical-web-and-design/vanilla-framework | https://api.github.com/repos/canonical-web-and-design/vanilla-framework | closed | Fieldset: Transparent border on fieldset results in border colour spilling outside element | Bug 🐛 Priority: High | **Describe the bug**
The transparent border currently applied to fieldsets causes a strange bug in chrome where the border color fills the window.
Replacing the fieldset with a div and copyign all fieldset styles onto the div doesn't seem to trigger the bug, so it seems highly specific to the fieldset or an inherent property of it:

**To Reproduce**
Paste this [markup](https://pastebin.canonical.com/p/2pnRKXPPkR/) in any site using vanilla.
- Device: Mac
- OS: Latest
- Browser Version 87.0.4280.67 (Official Build) (x86_64) | 1.0 | Fieldset: Transparent border on fieldset results in border colour spilling outside element - **Describe the bug**
The transparent border currently applied to fieldsets causes a strange bug in chrome where the border color fills the window.
Replacing the fieldset with a div and copyign all fieldset styles onto the div doesn't seem to trigger the bug, so it seems highly specific to the fieldset or an inherent property of it:

**To Reproduce**
Paste this [markup](https://pastebin.canonical.com/p/2pnRKXPPkR/) in any site using vanilla.
- Device: Mac
- OS: Latest
- Browser Version 87.0.4280.67 (Official Build) (x86_64) | priority | fieldset transparent border on fieldset results in border colour spilling outside element describe the bug the transparent border currently applied to fieldsets causes a strange bug in chrome where the border color fills the window replacing the fieldset with a div and copyign all fieldset styles onto the div doesn t seem to trigger the bug so it seems highly specific to the fieldset or an inherent property of it to reproduce paste this in any site using vanilla device mac os latest browser version official build | 1 |
155,746 | 5,960,139,006 | IssuesEvent | 2017-05-29 13:16:15 | Aurorastation/Aurora.3 | https://api.github.com/repos/Aurorastation/Aurora.3 | opened | SMC: Mapped in power connections do not appear to be updated properly | bug:confirmed flag:development flag:high priority | Okay. Long story short.
We've been running into issues in new map meme where connections that are valid, and exist since round start, are not flagged properly until an admin invokes `make-powernets`. Example cases of this:
* Tesla setup. This is all mapped in to be 100% functional (as far as wiring goes). However, after starting the engine, power does not flow until `make-powernets` or a local powernet update (I presume, haven't tried) is invoked.
* #2345 . Where mapped in z-wires do not move power until `make-powernets` is invoked. In this case, not even remaking the wires helps reliably, thus it's possible that local updates fail at times. Whether or not this is specific to z-wires should be confirmed.
* Possibly #2520 . Needs confirmation.
@Lohikar since this is something you've worked on, could you investigate during the week? | 1.0 | SMC: Mapped in power connections do not appear to be updated properly - Okay. Long story short.
We've been running into issues in new map meme where connections that are valid, and exist since round start, are not flagged properly until an admin invokes `make-powernets`. Example cases of this:
* Tesla setup. This is all mapped in to be 100% functional (as far as wiring goes). However, after starting the engine, power does not flow until `make-powernets` or a local powernet update (I presume, haven't tried) is invoked.
* #2345 . Where mapped in z-wires do not move power until `make-powernets` is invoked. In this case, not even remaking the wires helps reliably, thus it's possible that local updates fail at times. Whether or not this is specific to z-wires should be confirmed.
* Possibly #2520 . Needs confirmation.
@Lohikar since this is something you've worked on, could you investigate during the week? | priority | smc mapped in power connections do not appear to be updated properly okay long story short we ve been running into issues in new map meme where connections that are valid and exist since round start are not flagged properly until an admin invokes make powernets example cases of this tesla setup this is all mapped in to be functional as far as wiring goes however after starting the engine power does not flow until make powernets or a local powernet update i presume haven t tried is invoked where mapped in z wires do not move power until make powernets is invoked in this case not even remaking the wires helps reliably thus it s possible that local updates fail at times whether or not this is specific to z wires should be confirmed possibly needs confirmation lohikar since this is something you ve worked on could you investigate during the week | 1 |
272,410 | 8,508,152,747 | IssuesEvent | 2018-10-30 21:06:11 | isawnyu/isaw.web | https://api.github.com/repos/isawnyu/isaw.web | opened | Unable to edit images | bug high priority | I'm getting a blank dialog box this afternoon when I attempt to edit an image. It looks like this:
<img width="985" alt="screen shot 2018-10-30 at 4 57 29 pm" src="https://user-images.githubusercontent.com/7882896/47750546-00067c00-dc66-11e8-8be3-ac5886561082.png">
I am using Firefox (Mac). | 1.0 | Unable to edit images - I'm getting a blank dialog box this afternoon when I attempt to edit an image. It looks like this:
<img width="985" alt="screen shot 2018-10-30 at 4 57 29 pm" src="https://user-images.githubusercontent.com/7882896/47750546-00067c00-dc66-11e8-8be3-ac5886561082.png">
I am using Firefox (Mac). | priority | unable to edit images i m getting a blank dialog box this afternoon when i attempt to edit an image it looks like this img width alt screen shot at pm src i am using firefox mac | 1 |
739,995 | 25,731,498,609 | IssuesEvent | 2022-12-07 20:41:38 | fecgov/fec-cms | https://api.github.com/repos/fecgov/fec-cms | closed | Update meeting page template to change default open meeting and executive session time | Work: Front-end High priority | ### Summary
**What we're after:**
_OCS has indicated that meeting times will shift to start meetings at 10:30. We need to update the default time in the meeting page template so that users interested in Commission meetings have accurate information._
### Completion criteria
- [ ] Default time is changed on the meeting page template to 10:30. | 1.0 | Update meeting page template to change default open meeting and executive session time - ### Summary
**What we're after:**
_OCS has indicated that meeting times will shift to start meetings at 10:30. We need to update the default time in the meeting page template so that users interested in Commission meetings have accurate information._
### Completion criteria
- [ ] Default time is changed on the meeting page template to 10:30. | priority | update meeting page template to change default open meeting and executive session time summary what we re after ocs has indicated that meeting times will shift to start meetings at we need to update the default time in the meeting page template so that users interested in commission meetings have accurate information completion criteria default time is changed on the meeting page template to | 1 |
740,879 | 25,771,981,009 | IssuesEvent | 2022-12-09 08:48:15 | sorrowcode/taesch- | https://api.github.com/repos/sorrowcode/taesch- | closed | feature - Verbindung NearShopsScreen und MapScreen | USP - 5 feature 0 - highest priority | Als Nutzer möchte ich, dass ich beim Aufrufen eines Ladens eine Möglichkeit habe, dass ich zur Karte geführt werde, damit ich sehen kann, wo genau der Laden in meiner Nähe sich befindet.
Hier soll eine Strategie ausgearbeitet und implementiert werden, wodurch die Möglichkeit besteht, dynamisch zwischen NearShopsScreen und MapScreen zu navigieren und möglicherweise einzelne Läden anzeigen zu lassen. | 1.0 | feature - Verbindung NearShopsScreen und MapScreen - Als Nutzer möchte ich, dass ich beim Aufrufen eines Ladens eine Möglichkeit habe, dass ich zur Karte geführt werde, damit ich sehen kann, wo genau der Laden in meiner Nähe sich befindet.
Hier soll eine Strategie ausgearbeitet und implementiert werden, wodurch die Möglichkeit besteht, dynamisch zwischen NearShopsScreen und MapScreen zu navigieren und möglicherweise einzelne Läden anzeigen zu lassen. | priority | feature verbindung nearshopsscreen und mapscreen als nutzer möchte ich dass ich beim aufrufen eines ladens eine möglichkeit habe dass ich zur karte geführt werde damit ich sehen kann wo genau der laden in meiner nähe sich befindet hier soll eine strategie ausgearbeitet und implementiert werden wodurch die möglichkeit besteht dynamisch zwischen nearshopsscreen und mapscreen zu navigieren und möglicherweise einzelne läden anzeigen zu lassen | 1 |
586,988 | 17,601,238,335 | IssuesEvent | 2021-08-17 12:08:16 | gitpod-io/gitpod | https://api.github.com/repos/gitpod-io/gitpod | opened | Agent Smith signature check fails because of permission issues | type: bug component: agent-smith priority: highest (user impact) | ### Bug description
Agent Smith cannot check signatures of running processes at all because of permission issues. All violations that are found are found because if blacklisted commands, not exec signatures.
Log [link](https://cloudlogging.app.goo.gl/KcAc7153HjjXA6uE8)
Example log output:
```
{"@type":"type.googleapis.com/google.devtools.clouderrorreporting.v1beta1.ReportedErrorEvent","error":"open /proc/560401/exe: permission denied","level":"warning","message":"cannot open executable to check signatures","path":"id","serviceContext":{"service":"agent-smith","version":""},"severity":"WARNING","time":"2021-08-17T10:59:18Z"}
```
### Steps to reproduce
- Create a preview environment
- start a workspace
- check log of (relevant) agent smith daemon
### Expected behavior
Agent smith should be able to check signatures of all processes running inside a workspace.
### Example repository
_No response_
### Anything else?
_No response_ | 1.0 | Agent Smith signature check fails because of permission issues - ### Bug description
Agent Smith cannot check signatures of running processes at all because of permission issues. All violations that are found are found because if blacklisted commands, not exec signatures.
Log [link](https://cloudlogging.app.goo.gl/KcAc7153HjjXA6uE8)
Example log output:
```
{"@type":"type.googleapis.com/google.devtools.clouderrorreporting.v1beta1.ReportedErrorEvent","error":"open /proc/560401/exe: permission denied","level":"warning","message":"cannot open executable to check signatures","path":"id","serviceContext":{"service":"agent-smith","version":""},"severity":"WARNING","time":"2021-08-17T10:59:18Z"}
```
### Steps to reproduce
- Create a preview environment
- start a workspace
- check log of (relevant) agent smith daemon
### Expected behavior
Agent smith should be able to check signatures of all processes running inside a workspace.
### Example repository
_No response_
### Anything else?
_No response_ | priority | agent smith signature check fails because of permission issues bug description agent smith cannot check signatures of running processes at all because of permission issues all violations that are found are found because if blacklisted commands not exec signatures log example log output type type googleapis com google devtools clouderrorreporting reportederrorevent error open proc exe permission denied level warning message cannot open executable to check signatures path id servicecontext service agent smith version severity warning time steps to reproduce create a preview environment start a workspace check log of relevant agent smith daemon expected behavior agent smith should be able to check signatures of all processes running inside a workspace example repository no response anything else no response | 1 |
800,379 | 28,363,711,686 | IssuesEvent | 2023-04-12 12:36:37 | flowforge/flowforge | https://api.github.com/repos/flowforge/flowforge | closed | Cancel button application/settings edit page doesn't work | bug area:frontend priority:high | ### Current Behavior
```
ntime-core.esm-bundler.js:169 [Vue warn]: Property "cancelEditName" was accessed during render but is not defined on instance.
at <ProjectSettings application= Object instances= Array(1) is-visiting-admin=false ... >
at <RouterView application= Object instances= Array(1) is-visiting-admin=false ... >
at <ProjectPage onVnodeUnmounted=fn<onVnodeUnmounted> ref=Ref< Proxy(Object) > >
at <RouterView>
at <FfLayoutPlatform key=0 >
at <App>
w
```
### Expected Behavior
_No response_
### Steps To Reproduce
_No response_
### Environment
- FlowForge version:
- Node.js version:
- npm version:
- Platform/OS:
- Browser:
| 1.0 | Cancel button application/settings edit page doesn't work - ### Current Behavior
```
ntime-core.esm-bundler.js:169 [Vue warn]: Property "cancelEditName" was accessed during render but is not defined on instance.
at <ProjectSettings application= Object instances= Array(1) is-visiting-admin=false ... >
at <RouterView application= Object instances= Array(1) is-visiting-admin=false ... >
at <ProjectPage onVnodeUnmounted=fn<onVnodeUnmounted> ref=Ref< Proxy(Object) > >
at <RouterView>
at <FfLayoutPlatform key=0 >
at <App>
w
```
### Expected Behavior
_No response_
### Steps To Reproduce
_No response_
### Environment
- FlowForge version:
- Node.js version:
- npm version:
- Platform/OS:
- Browser:
| priority | cancel button application settings edit page doesn t work current behavior ntime core esm bundler js property canceleditname was accessed during render but is not defined on instance at at at ref ref at at at w expected behavior no response steps to reproduce no response environment flowforge version node js version npm version platform os browser | 1 |
772,523 | 27,125,714,699 | IssuesEvent | 2023-02-16 05:01:02 | FastcampusMini/mini-project | https://api.github.com/repos/FastcampusMini/mini-project | closed | Basket API | For: API Priority: High Status: In Progress Type: Feature | ## Title
장바구니 기능
## Description
장바구니
## Tasks
- [x] 장바구니 repo, service, dto 구현
| 1.0 | Basket API - ## Title
장바구니 기능
## Description
장바구니
## Tasks
- [x] 장바구니 repo, service, dto 구현
| priority | basket api title 장바구니 기능 description 장바구니 tasks 장바구니 repo service dto 구현 | 1 |
777,656 | 27,289,606,690 | IssuesEvent | 2023-02-23 15:43:58 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | PFacct issue in cluster | Type: Bug Priority: High | PacketFence 12.2.
Switching from standalone to cluster when creating a cluster, the load balancer does not want to start because PF acct is already listening.
| 1.0 | PFacct issue in cluster - PacketFence 12.2.
Switching from standalone to cluster when creating a cluster, the load balancer does not want to start because PF acct is already listening.
| priority | pfacct issue in cluster packetfence switching from standalone to cluster when creating a cluster the load balancer does not want to start because pf acct is already listening | 1 |
329,670 | 10,022,953,619 | IssuesEvent | 2019-07-16 17:58:54 | ArtskydJ/comicsrss.com | https://api.github.com/repos/ArtskydJ/comicsrss.com | closed | Some gocomics strips are not showing up | high-priority | From @megadr01d, originally posted in #86
Hope you don't mind... I removed your other comment since I see these as two separate issues.
-----------------------------------
Why are some GoComics comics are not available?
Like these two, which I'd love to see in the future:
https://www.gocomics.com/how-to-cat
https://www.gocomics.com/redmeat | 1.0 | Some gocomics strips are not showing up - From @megadr01d, originally posted in #86
Hope you don't mind... I removed your other comment since I see these as two separate issues.
-----------------------------------
Why are some GoComics comics are not available?
Like these two, which I'd love to see in the future:
https://www.gocomics.com/how-to-cat
https://www.gocomics.com/redmeat | priority | some gocomics strips are not showing up from originally posted in hope you don t mind i removed your other comment since i see these as two separate issues why are some gocomics comics are not available like these two which i d love to see in the future | 1 |
530,309 | 15,420,784,113 | IssuesEvent | 2021-03-05 12:05:53 | apluslms/a-plus | https://api.github.com/repos/apluslms/a-plus | closed | Latest course instance redirection must pick the latest visible instance | O1 needs area: UX student effort: hours experience: good first issue priority: high requester: Aalto teacher type: bug | Pull request #772 added the redirection view to the latest course instance: https://plus.cs.aalto.fi/o1/
However, if the latest instance is hidden from students, then this redirection becomes useless and confusing to students. It is common that the latest instance is hidden before it has begun because it is still work in progress. The redirection should always pick the latest VISIBLE instance. | 1.0 | Latest course instance redirection must pick the latest visible instance - Pull request #772 added the redirection view to the latest course instance: https://plus.cs.aalto.fi/o1/
However, if the latest instance is hidden from students, then this redirection becomes useless and confusing to students. It is common that the latest instance is hidden before it has begun because it is still work in progress. The redirection should always pick the latest VISIBLE instance. | priority | latest course instance redirection must pick the latest visible instance pull request added the redirection view to the latest course instance however if the latest instance is hidden from students then this redirection becomes useless and confusing to students it is common that the latest instance is hidden before it has begun because it is still work in progress the redirection should always pick the latest visible instance | 1 |
566,102 | 16,796,056,019 | IssuesEvent | 2021-06-16 03:49:18 | parallel-finance/parallel | https://api.github.com/repos/parallel-finance/parallel | closed | tokio-runtime-worker runtime::timestamp: `pallet_timestamp::UnixTime::now` is called at genesis, invalid value returned: 0 | high priority | <img width="1297" alt="image" src="https://user-images.githubusercontent.com/33961674/121112081-81711280-c842-11eb-9f7d-3391524261f4.png">
| 1.0 | tokio-runtime-worker runtime::timestamp: `pallet_timestamp::UnixTime::now` is called at genesis, invalid value returned: 0 - <img width="1297" alt="image" src="https://user-images.githubusercontent.com/33961674/121112081-81711280-c842-11eb-9f7d-3391524261f4.png">
| priority | tokio runtime worker runtime timestamp pallet timestamp unixtime now is called at genesis invalid value returned img width alt image src | 1 |
138,583 | 5,344,628,530 | IssuesEvent | 2017-02-17 15:01:22 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | Feature Request: torch 'module' object has no attribute '__version__' | enhancement high priority | could you please add `__version__` to the `torch` module! all the cool modules have it! ;) | 1.0 | Feature Request: torch 'module' object has no attribute '__version__' - could you please add `__version__` to the `torch` module! all the cool modules have it! ;) | priority | feature request torch module object has no attribute version could you please add version to the torch module all the cool modules have it | 1 |
761,872 | 26,700,578,959 | IssuesEvent | 2023-01-27 14:05:46 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | Error when Renaming Workspaces with Plots | High Priority Bug | **Describe the bug**
Workbench throws errors when a workspace is renamed with a plot of its data open.
**To Reproduce**
1. Load data
2. Plot a spectrum
3. Rename the workspace
**Expected behaviour**
Title of the plot is changed and no errors are thrown.
**Platform/Version (please complete the following information):**
- OS: All
- Mantid Version: 6.6 Nightly (2023-01-24)
**Additional context**
```
Error occurred in handler:
Traceback (most recent call last):
File "/home/conor/repos/mantid/qt/applications/workbench/workbench/plotting/figuremanager.py", line 62, in wrapper
func(*args, **kwargs)
File "/home/conor/repos/mantid/qt/applications/workbench/workbench/plotting/figuremanager.py", line 162, in renameHandle
self.canvas.manager.set_window_title(_replace_workspace_name_in_string(oldName, newName, self.canvas.get_window_title()))
AttributeError: 'MantidFigureCanvas' object has no attribute 'get_window_title'
```
| 1.0 | Error when Renaming Workspaces with Plots - **Describe the bug**
Workbench throws errors when a workspace is renamed with a plot of its data open.
**To Reproduce**
1. Load data
2. Plot a spectrum
3. Rename the workspace
**Expected behaviour**
Title of the plot is changed and no errors are thrown.
**Platform/Version (please complete the following information):**
- OS: All
- Mantid Version: 6.6 Nightly (2023-01-24)
**Additional context**
```
Error occurred in handler:
Traceback (most recent call last):
File "/home/conor/repos/mantid/qt/applications/workbench/workbench/plotting/figuremanager.py", line 62, in wrapper
func(*args, **kwargs)
File "/home/conor/repos/mantid/qt/applications/workbench/workbench/plotting/figuremanager.py", line 162, in renameHandle
self.canvas.manager.set_window_title(_replace_workspace_name_in_string(oldName, newName, self.canvas.get_window_title()))
AttributeError: 'MantidFigureCanvas' object has no attribute 'get_window_title'
```
| priority | error when renaming workspaces with plots describe the bug workbench throws errors when a workspace is renamed with a plot of its data open to reproduce load data plot a spectrum rename the workspace expected behaviour title of the plot is changed and no errors are thrown platform version please complete the following information os all mantid version nightly additional context error occurred in handler traceback most recent call last file home conor repos mantid qt applications workbench workbench plotting figuremanager py line in wrapper func args kwargs file home conor repos mantid qt applications workbench workbench plotting figuremanager py line in renamehandle self canvas manager set window title replace workspace name in string oldname newname self canvas get window title attributeerror mantidfigurecanvas object has no attribute get window title | 1 |
829,179 | 31,857,819,731 | IssuesEvent | 2023-09-15 08:45:36 | rcpch/rcpch-audit-engine | https://api.github.com/repos/rcpch/rcpch-audit-engine | closed | Current code commit/version should be linked in topmost nav (left) | before-launch feature request priority: high | For ease of administrator/developer debugging, it would be a nice-to-have if the current commit hash of the running version of the software was visible in the top left nav.
* [ ] Should show **short** commit hash of the current version
* [ ] Should only be shown to Admins and Developers
* [ ] Should click through to the commit on Github for quickly viewing the relevant code
* [ ] Version number +/- Git tag +/- Github Release should also be displayed once we have semver established
Sketch of what this would look like

| 1.0 | Current code commit/version should be linked in topmost nav (left) - For ease of administrator/developer debugging, it would be a nice-to-have if the current commit hash of the running version of the software was visible in the top left nav.
* [ ] Should show **short** commit hash of the current version
* [ ] Should only be shown to Admins and Developers
* [ ] Should click through to the commit on Github for quickly viewing the relevant code
* [ ] Version number +/- Git tag +/- Github Release should also be displayed once we have semver established
Sketch of what this would look like

| priority | current code commit version should be linked in topmost nav left for ease of administrator developer debugging it would be a nice to have if the current commit hash of the running version of the software was visible in the top left nav should show short commit hash of the current version should only be shown to admins and developers should click through to the commit on github for quickly viewing the relevant code version number git tag github release should also be displayed once we have semver established sketch of what this would look like | 1 |
472,092 | 13,616,165,547 | IssuesEvent | 2020-09-23 15:17:51 | OpenLightingProject/ola | https://api.github.com/repos/OpenLightingProject/ola | closed | New warnings with GCC9 | Component-Plugin Difficulty-Easy Language-C++ Maintainability OpSys-Linux Priority-High bug | I received [Debian bug 925793](https://bugs.debian.org/925793), which produces a few new warnings of this type:
```
libs/acn/HeaderSetTest.cpp: In member function 'void HeaderSetTest::testTransportHeader()':
libs/acn/HeaderSetTest.cpp:78:29: error: implicitly-declared 'ola::acn::TransportHeader::TransportHeader(const ola::acn::TransportHeader&)' is deprecated [-Werror=deprecated-copy]
78 | TransportHeader header2 = header;
| ^~~~~~
In file included from ./libs/acn/HeaderSet.h:30,
from libs/acn/HeaderSetTest.cpp:28:
../libs/acn/TransportHeader.h:58:8: note: because 'ola::acn::TransportHeader' has user-provided 'void ola::acn::TransportHeader::operator=(const ola::acn::TransportHeader&)'
58 | void operator=(const TransportHeader &other) {
| ^~~~~~~~
```
Basically, what happens here is that the compiler complains about a class which does have an assignment operator, but no copy constructor. You should probably add them. | 1.0 | New warnings with GCC9 - I received [Debian bug 925793](https://bugs.debian.org/925793), which produces a few new warnings of this type:
```
libs/acn/HeaderSetTest.cpp: In member function 'void HeaderSetTest::testTransportHeader()':
libs/acn/HeaderSetTest.cpp:78:29: error: implicitly-declared 'ola::acn::TransportHeader::TransportHeader(const ola::acn::TransportHeader&)' is deprecated [-Werror=deprecated-copy]
78 | TransportHeader header2 = header;
| ^~~~~~
In file included from ./libs/acn/HeaderSet.h:30,
from libs/acn/HeaderSetTest.cpp:28:
../libs/acn/TransportHeader.h:58:8: note: because 'ola::acn::TransportHeader' has user-provided 'void ola::acn::TransportHeader::operator=(const ola::acn::TransportHeader&)'
58 | void operator=(const TransportHeader &other) {
| ^~~~~~~~
```
Basically, what happens here is that the compiler complains about a class which does have an assignment operator, but no copy constructor. You should probably add them. | priority | new warnings with i received which produces a few new warnings of this type libs acn headersettest cpp in member function void headersettest testtransportheader libs acn headersettest cpp error implicitly declared ola acn transportheader transportheader const ola acn transportheader is deprecated transportheader header in file included from libs acn headerset h from libs acn headersettest cpp libs acn transportheader h note because ola acn transportheader has user provided void ola acn transportheader operator const ola acn transportheader void operator const transportheader other basically what happens here is that the compiler complains about a class which does have an assignment operator but no copy constructor you should probably add them | 1 |
286,709 | 8,792,000,176 | IssuesEvent | 2018-12-21 14:43:02 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | opened | support symbolic random functions (Vector<Expression>) | priority: high team: automotive type: feature request | Unlike the existing Vector<Expression>, where each element can always be evaluated separately, I think in the random case I need to call `Evaluate(vector<Expression>, generator)` and have the same random value used for all of the outputs?
in other words... if I have
```
v = Variable(RANDOM_UNIFORM)
y1 = 5*v + 1
y2 = 3*v + 3
y = Eigen::Vector2<Expression>(y1, y2)
```
i would like to be able to call `Evaluate(y1, generator)` and `Evaluate(y2, generator)` separately, which would result in independent draws from the distribution of `v`.
but I would also like to be able to call `Evaluate(y, generator)`, and have `v` drawn only once, but used for both `y1` and `y2`. | 1.0 | support symbolic random functions (Vector<Expression>) - Unlike the existing Vector<Expression>, where each element can always be evaluated separately, I think in the random case I need to call `Evaluate(vector<Expression>, generator)` and have the same random value used for all of the outputs?
in other words... if I have
```
v = Variable(RANDOM_UNIFORM)
y1 = 5*v + 1
y2 = 3*v + 3
y = Eigen::Vector2<Expression>(y1, y2)
```
i would like to be able to call `Evaluate(y1, generator)` and `Evaluate(y2, generator)` separately, which would result in independent draws from the distribution of `v`.
but I would also like to be able to call `Evaluate(y, generator)`, and have `v` drawn only once, but used for both `y1` and `y2`. | priority | support symbolic random functions vector unlike the existing vector where each element can always be evaluated separately i think in the random case i need to call evaluate vector generator and have the same random value used for all of the outputs in other words if i have v variable random uniform v v y eigen i would like to be able to call evaluate generator and evaluate generator separately which would result in independent draws from the distribution of v but i would also like to be able to call evaluate y generator and have v drawn only once but used for both and | 1 |
423,427 | 12,296,028,115 | IssuesEvent | 2020-05-11 05:57:13 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | color.firefox.com - see bug description | browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal type-webrender-enabled | <!-- @browser: Firefox 78.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; rv:78.0) Gecko/20100101 Firefox/78.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52670 -->
<!-- @extra_labels: type-webrender-enabled -->
**URL**: https://color.firefox.com/
**Browser / Version**: Firefox 78.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: Background seen below the toolbar
**Steps to Reproduce**:
Just noticed that there's a line (maybe one pixel high) below my toolbar (on every website) when setting a background image with a different color than my toolbar is.
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/5/435ee96a-d92d-456d-b553-c6a057259ed9.jpeg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: true</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200509091423</li><li>channel: nightly</li><li>GPUs: <ul>
<li>active: true</li><li>description: Intel(R) Iris(R) Graphics 540</li><li>deviceID: 0x1926</li><li>vendorID: 0x8086</li><li>driverVersion: 26.20.100.7529</li>
</ul></li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/5/b66ac394-49ab-48be-b5f9-fe79cfddd30c)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | color.firefox.com - see bug description - <!-- @browser: Firefox 78.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; rv:78.0) Gecko/20100101 Firefox/78.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52670 -->
<!-- @extra_labels: type-webrender-enabled -->
**URL**: https://color.firefox.com/
**Browser / Version**: Firefox 78.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: Background seen below the toolbar
**Steps to Reproduce**:
Just noticed that there's a line (maybe one pixel high) below my toolbar (on every website) when setting a background image with a different color than my toolbar is.
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/5/435ee96a-d92d-456d-b553-c6a057259ed9.jpeg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: true</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200509091423</li><li>channel: nightly</li><li>GPUs: <ul>
<li>active: true</li><li>description: Intel(R) Iris(R) Graphics 540</li><li>deviceID: 0x1926</li><li>vendorID: 0x8086</li><li>driverVersion: 26.20.100.7529</li>
</ul></li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/5/b66ac394-49ab-48be-b5f9-fe79cfddd30c)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | color firefox com see bug description url browser version firefox operating system windows tested another browser no problem type something else description background seen below the toolbar steps to reproduce just noticed that there s a line maybe one pixel high below my toolbar on every website when setting a background image with a different color than my toolbar is view the screenshot img alt screenshot src browser configuration gfx webrender all true gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly gpus active true description intel r iris r graphics deviceid vendorid driverversion hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
830,331 | 32,002,753,395 | IssuesEvent | 2023-09-21 13:14:17 | electric-sql/electric | https://api.github.com/repos/electric-sql/electric | closed | [VAX-1004] Supported dev OSes | bug linear High priority | Given the wide usage of Ubuntu 22.04, it would seem a reasonable dev system to work with. Are you guys all working/testing on MacOS or something? Several steps in the quickstart just seem to fail, and trying to hack around them requires installing and sourcing an `emsdk`, installing `node_modules` with `PUPPETEER_SKIP_DOWNLOAD=true npm i @electric-sql/prisma-generator`, doing `yarn install`, etc. - all things not listed at all in the quickstart.
I realise things are moving fast, but a quick test in a fresh Ubuntu 22.04 VM for the quickstart seems pretty reasonable to me... The issue being, it is very much not "quick" to get started, and wouldn't be possible for someone who hasn't spent a lot of time trying to get things working.
<sub>[VAX-1004](https://linear.app/electric-sql/issue/VAX-1004/supported-dev-oses)</sub> | 1.0 | [VAX-1004] Supported dev OSes - Given the wide usage of Ubuntu 22.04, it would seem a reasonable dev system to work with. Are you guys all working/testing on MacOS or something? Several steps in the quickstart just seem to fail, and trying to hack around them requires installing and sourcing an `emsdk`, installing `node_modules` with `PUPPETEER_SKIP_DOWNLOAD=true npm i @electric-sql/prisma-generator`, doing `yarn install`, etc. - all things not listed at all in the quickstart.
I realise things are moving fast, but a quick test in a fresh Ubuntu 22.04 VM for the quickstart seems pretty reasonable to me... The issue being, it is very much not "quick" to get started, and wouldn't be possible for someone who hasn't spent a lot of time trying to get things working.
<sub>[VAX-1004](https://linear.app/electric-sql/issue/VAX-1004/supported-dev-oses)</sub> | priority | supported dev oses given the wide usage of ubuntu it would seem a reasonable dev system to work with are you guys all working testing on macos or something several steps in the quickstart just seem to fail and trying to hack around them requires installing and sourcing an emsdk installing node modules with puppeteer skip download true npm i electric sql prisma generator doing yarn install etc all things not listed at all in the quickstart i realise things are moving fast but a quick test in a fresh ubuntu vm for the quickstart seems pretty reasonable to me the issue being it is very much not quick to get started and wouldn t be possible for someone who hasn t spent a lot of time trying to get things working | 1 |
451,270 | 13,032,575,573 | IssuesEvent | 2020-07-28 04:39:42 | rica441/takutaku | https://api.github.com/repos/rica441/takutaku | closed | かんたんログイン機能 | feature high priority | ## なにをやるのか
- ユーザー情報を登録せずにゲストユーザーとしてログインする
## なぜそれをやるのか
- いろんな人にtekutekuを使ってみて欲しいから
- フィードバックをもらいたい
## 参考リンク
- https://www.isoroot.jp/blog/2451/
| 1.0 | かんたんログイン機能 - ## なにをやるのか
- ユーザー情報を登録せずにゲストユーザーとしてログインする
## なぜそれをやるのか
- いろんな人にtekutekuを使ってみて欲しいから
- フィードバックをもらいたい
## 参考リンク
- https://www.isoroot.jp/blog/2451/
| priority | かんたんログイン機能 なにをやるのか ユーザー情報を登録せずにゲストユーザーとしてログインする なぜそれをやるのか いろんな人にtekutekuを使ってみて欲しいから フィードバックをもらいたい 参考リンク | 1 |
552,680 | 16,246,837,957 | IssuesEvent | 2021-05-07 15:23:37 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YCQL] Index not chosen based on ordering and certain ordering allowed that shouldn't be | priority/high | Instead of optimizing for the ordering an index is chosen only based on the criteria which means it's impossible to make some queries. Additionally, queries that involve the order of the primary index are allowed despite using a secondary index.
Here's an example schema:
```cql
CREATE KEYSPACE test;
USE test;
CREATE TABLE test (
id int,
scope text,
key text,
value text,
PRIMARY KEY (id)
) WITH default_time_to_live = 0
AND transactions = {'enabled': 'true'};
CREATE INDEX test_index_1 ON test (scope, key, value)
WITH CLUSTERING ORDER BY (key ASC, value ASC)
AND transactions = {'enabled': 'true'};
CREATE INDEX test_index_2 ON test (scope, key, id)
WITH CLUSTERING ORDER BY (key ASC, id ASC)
AND transactions = {'enabled': 'true'};
INSERT into test (id, scope, key, value) VALUES (4, 'test', 'bar', 'baz');
INSERT into test (id, scope, key, value) VALUES (3, 'test', 'foo', 'baz');
INSERT into test (id, scope, key, value) VALUES (2, 'test', 'foo', 'bar');
INSERT into test (id, scope, key, value) VALUES (1, 'test', 'foo', 'bar');
```
Now let's say you want to paginate over all of the "foo" `key` and "bar" `value`. You need to order by id because `value` is mutable and could cause the pagination to duplicate/change as your paginating. So your query is:
```
EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id LIMIT 1;
InvalidRequest: Error from server: code=2200 [Invalid query] message="Invalid Arguments. Order by currently only support the ordering of columns following their declared order in the PRIMARY KEY
EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id LIMIT 1;
^^^
(ql error -304)"
```
Which errors. This is because it chooses `test_index_1` instead of choosing `test_index_2`.
If we add `value` to the end of the order by, however:
```
EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id, value LIMIT 1;
QUERY PLAN
------------------------------------------------------
Index Only Scan using test.test_index_1 on test.test
Key Conditions: (scope = 'test')
Filter: (key = 'foo') AND (value = 'bar')
```
Then doesn't error but it doesn't follow the order in `test_index_1 (key, value, id)`. The `id` at the end is assumed based on the docs saying:
> Any primary key column of the table not indexed explicitly in `index_columns` is added as a clustering column to the index implicitly. | 1.0 | [YCQL] Index not chosen based on ordering and certain ordering allowed that shouldn't be - Instead of optimizing for the ordering an index is chosen only based on the criteria which means it's impossible to make some queries. Additionally, queries that involve the order of the primary index are allowed despite using a secondary index.
Here's an example schema:
```cql
CREATE KEYSPACE test;
USE test;
CREATE TABLE test (
id int,
scope text,
key text,
value text,
PRIMARY KEY (id)
) WITH default_time_to_live = 0
AND transactions = {'enabled': 'true'};
CREATE INDEX test_index_1 ON test (scope, key, value)
WITH CLUSTERING ORDER BY (key ASC, value ASC)
AND transactions = {'enabled': 'true'};
CREATE INDEX test_index_2 ON test (scope, key, id)
WITH CLUSTERING ORDER BY (key ASC, id ASC)
AND transactions = {'enabled': 'true'};
INSERT into test (id, scope, key, value) VALUES (4, 'test', 'bar', 'baz');
INSERT into test (id, scope, key, value) VALUES (3, 'test', 'foo', 'baz');
INSERT into test (id, scope, key, value) VALUES (2, 'test', 'foo', 'bar');
INSERT into test (id, scope, key, value) VALUES (1, 'test', 'foo', 'bar');
```
Now let's say you want to paginate over all of the "foo" `key` and "bar" `value`. You need to order by id because `value` is mutable and could cause the pagination to duplicate/change as your paginating. So your query is:
```
EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id LIMIT 1;
InvalidRequest: Error from server: code=2200 [Invalid query] message="Invalid Arguments. Order by currently only support the ordering of columns following their declared order in the PRIMARY KEY
EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id LIMIT 1;
^^^
(ql error -304)"
```
Which errors. This is because it chooses `test_index_1` instead of choosing `test_index_2`.
If we add `value` to the end of the order by, however:
```
EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id, value LIMIT 1;
QUERY PLAN
------------------------------------------------------
Index Only Scan using test.test_index_1 on test.test
Key Conditions: (scope = 'test')
Filter: (key = 'foo') AND (value = 'bar')
```
Then doesn't error but it doesn't follow the order in `test_index_1 (key, value, id)`. The `id` at the end is assumed based on the docs saying:
> Any primary key column of the table not indexed explicitly in `index_columns` is added as a clustering column to the index implicitly. | priority | index not chosen based on ordering and certain ordering allowed that shouldn t be instead of optimizing for the ordering an index is chosen only based on the criteria which means it s impossible to make some queries additionally queries that involve the order of the primary index are allowed despite using a secondary index here s an example schema cql create keyspace test use test create table test id int scope text key text value text primary key id with default time to live and transactions enabled true create index test index on test scope key value with clustering order by key asc value asc and transactions enabled true create index test index on test scope key id with clustering order by key asc id asc and transactions enabled true insert into test id scope key value values test bar baz insert into test id scope key value values test foo baz insert into test id scope key value values test foo bar insert into test id scope key value values test foo bar now let s say you want to paginate over all of the foo key and bar value you need to order by id because value is mutable and could cause the pagination to duplicate change as your paginating so your query is explain select id from test where scope test and key foo and value bar order by key id limit invalidrequest error from server code message invalid arguments order by currently only support the ordering of columns following their declared order in the primary key explain select id from test where scope test and key foo and value bar order by key id limit ql error which errors this is because it chooses test index instead of choosing test index if we add value to the end of the order by however explain select id from test where scope test and key foo and value bar order by key id value limit query plan index only scan using test test index on test test key conditions scope test filter key foo and value bar then doesn t error but it doesn t follow the order in test index key value id the id at the end is assumed based on the docs saying any primary key column of the table not indexed explicitly in index columns is added as a clustering column to the index implicitly | 1 |
598,144 | 18,237,879,504 | IssuesEvent | 2021-10-01 09:14:46 | AbsaOSS/enceladus | https://api.github.com/repos/AbsaOSS/enceladus | opened | Add the possibility to limit the output files sizes | feature Conformance under discussion Standardization priority: high | ## Background
Sometimes the output files are too big and other applications have troubles operating with them.
## Feature
Make it possible to specify the maximum size of output parquet files.
## Proposed Solution
Solution Ideas:
1. Have an optional setting for maximal file size.
2. If the limit would be overshot, repartition.
3. This should be done both for _Standardization_ and _Conformance_.
| 1.0 | Add the possibility to limit the output files sizes - ## Background
Sometimes the output files are too big and other applications have troubles operating with them.
## Feature
Make it possible to specify the maximum size of output parquet files.
## Proposed Solution
Solution Ideas:
1. Have an optional setting for maximal file size.
2. If the limit would be overshot, repartition.
3. This should be done both for _Standardization_ and _Conformance_.
| priority | add the possibility to limit the output files sizes background sometimes the output files are too big and other applications have troubles operating with them feature make it possible to specify the maximum size of output parquet files proposed solution solution ideas have an optional setting for maximal file size if the limit would be overshot repartition this should be done both for standardization and conformance | 1 |
439,719 | 12,685,663,232 | IssuesEvent | 2020-06-20 06:01:37 | sodafoundation/dashboard | https://api.github.com/repos/sodafoundation/dashboard | closed | Dashboard still shows OpenSDS Logo and references to OpenSDS | High Priority enhancement | **Issue/Feature Description:**
Change OpenSDS logo to SODA and all occurrences of OpenSDS to SODA
**Why this issue to fixed / feature is needed(give scenarios or use cases):**
Since the project is now renamed to SODA all occurrences of the OpenSDS logo and OpenSDS title have to be changed to SODA. The main places that the changes have to be made are:
- Login Page Logo
- Home Page Menu Logo
- Browser Tab title
- Favicon
| 1.0 | Dashboard still shows OpenSDS Logo and references to OpenSDS - **Issue/Feature Description:**
Change OpenSDS logo to SODA and all occurrences of OpenSDS to SODA
**Why this issue to fixed / feature is needed(give scenarios or use cases):**
Since the project is now renamed to SODA all occurrences of the OpenSDS logo and OpenSDS title have to be changed to SODA. The main places that the changes have to be made are:
- Login Page Logo
- Home Page Menu Logo
- Browser Tab title
- Favicon
| priority | dashboard still shows opensds logo and references to opensds issue feature description change opensds logo to soda and all occurrences of opensds to soda why this issue to fixed feature is needed give scenarios or use cases since the project is now renamed to soda all occurrences of the opensds logo and opensds title have to be changed to soda the main places that the changes have to be made are login page logo home page menu logo browser tab title favicon | 1 |
262,284 | 8,263,191,459 | IssuesEvent | 2018-09-14 00:50:23 | ampproject/amphtml | https://api.github.com/repos/ampproject/amphtml | closed | Cleanup amp-image-slider loading issue | Category: Presentation P1: High Priority Type: Feature Request | Currently, there is no code that handles loading for `amp-image-slider`. By default, the slider should construct its structure and load the images normally, which shows the placeholder and the loading 3 dots of `amp-img`.
This is actually one of the proposal: __Proposal A__: leave as is.
(LEFT is the placeholder for left image. The following are under "Fast 3G" so should be much faster on actual devices)

However, due to ownership related issues, the 3 dots would only show when `layer` experiment is enabled. Otherwise, the loading process of images is never displayed.
Another proposal (__Proposal B__) to the loading state is to whitelist the slider itself for loading and display 3 dots of the slider, which would mask on top of the loading images until `LOAD_END` is signaled. This is attempted in #17420

| 1.0 | Cleanup amp-image-slider loading issue - Currently, there is no code that handles loading for `amp-image-slider`. By default, the slider should construct its structure and load the images normally, which shows the placeholder and the loading 3 dots of `amp-img`.
This is actually one of the proposal: __Proposal A__: leave as is.
(LEFT is the placeholder for left image. The following are under "Fast 3G" so should be much faster on actual devices)

However, due to ownership related issues, the 3 dots would only show when `layer` experiment is enabled. Otherwise, the loading process of images is never displayed.
Another proposal (__Proposal B__) to the loading state is to whitelist the slider itself for loading and display 3 dots of the slider, which would mask on top of the loading images until `LOAD_END` is signaled. This is attempted in #17420

| priority | cleanup amp image slider loading issue currently there is no code that handles loading for amp image slider by default the slider should construct its structure and load the images normally which shows the placeholder and the loading dots of amp img this is actually one of the proposal proposal a leave as is left is the placeholder for left image the following are under fast so should be much faster on actual devices however due to ownership related issues the dots would only show when layer experiment is enabled otherwise the loading process of images is never displayed another proposal proposal b to the loading state is to whitelist the slider itself for loading and display dots of the slider which would mask on top of the loading images until load end is signaled this is attempted in | 1 |
671,530 | 22,764,866,088 | IssuesEvent | 2022-07-08 02:33:02 | sacloud/terraform-provider-sakuracloud | https://api.github.com/repos/sacloud/terraform-provider-sakuracloud | closed | sakuracloud_archiveのos_typeパラメータがスコープに対応していない | bug v2 priority/high | os_typeを指定すると対応するパブリックアーカイブを検索するが、その際にscopeが考慮されないためパブリックアーカイブと同一のタグを持つアーカイブがあると意図したアーカイブを参照できない。
| 1.0 | sakuracloud_archiveのos_typeパラメータがスコープに対応していない - os_typeを指定すると対応するパブリックアーカイブを検索するが、その際にscopeが考慮されないためパブリックアーカイブと同一のタグを持つアーカイブがあると意図したアーカイブを参照できない。
| priority | sakuracloud archiveのos typeパラメータがスコープに対応していない os typeを指定すると対応するパブリックアーカイブを検索するが、その際にscopeが考慮されないためパブリックアーカイブと同一のタグを持つアーカイブがあると意図したアーカイブを参照できない。 | 1 |
230,957 | 7,621,839,427 | IssuesEvent | 2018-05-03 09:56:33 | HBHWoolacotts/RPii | https://api.github.com/repos/HBHWoolacotts/RPii | closed | Unable to view uploaded Files in Service Jobs (Forbidden) | Label: General RP Bugs and Support Priority - High | Since the server upgrade, we cannot access uploaded files in service jobs. When we try to view them by pressing the little paperclip, it says Forbidden:


| 1.0 | Unable to view uploaded Files in Service Jobs (Forbidden) - Since the server upgrade, we cannot access uploaded files in service jobs. When we try to view them by pressing the little paperclip, it says Forbidden:


| priority | unable to view uploaded files in service jobs forbidden since the server upgrade we cannot access uploaded files in service jobs when we try to view them by pressing the little paperclip it says forbidden | 1 |
366,443 | 10,820,907,083 | IssuesEvent | 2019-11-08 17:23:39 | Automattic/simplenote-macos | https://api.github.com/repos/Automattic/simplenote-macos | opened | Crash: Emojis Again! | [priority] high bug | ### Details:
1. Add a new note
2. Enter the following emojis: `☺️🖖🏿`
3. Insert the following emoji in between: `😃`
As a result the app will crash. We'll end up with a malformed unicode sequence, which effectively breaks Apple's JSON Parser.
| 1.0 | Crash: Emojis Again! - ### Details:
1. Add a new note
2. Enter the following emojis: `☺️🖖🏿`
3. Insert the following emoji in between: `😃`
As a result the app will crash. We'll end up with a malformed unicode sequence, which effectively breaks Apple's JSON Parser.
| priority | crash emojis again details add a new note enter the following emojis ☺️🖖🏿 insert the following emoji in between 😃 as a result the app will crash we ll end up with a malformed unicode sequence which effectively breaks apple s json parser | 1 |
486,499 | 14,010,133,872 | IssuesEvent | 2020-10-29 04:15:18 | CIA-Homebrew/BJCP-Scoresheet | https://api.github.com/repos/CIA-Homebrew/BJCP-Scoresheet | closed | Multiple scoresheet PDFs should be downloaded as a zip | HIGH PRIORITY back end enhancement | Currently, downloading multiple scoresheets will start multiple individual downloads. When more than one download is requested, there should be a zip handler that puts all pdfs in a single zip
Consider using this lib: https://github.com/archiverjs/node-archiver | 1.0 | Multiple scoresheet PDFs should be downloaded as a zip - Currently, downloading multiple scoresheets will start multiple individual downloads. When more than one download is requested, there should be a zip handler that puts all pdfs in a single zip
Consider using this lib: https://github.com/archiverjs/node-archiver | priority | multiple scoresheet pdfs should be downloaded as a zip currently downloading multiple scoresheets will start multiple individual downloads when more than one download is requested there should be a zip handler that puts all pdfs in a single zip consider using this lib | 1 |
504,337 | 14,616,802,865 | IssuesEvent | 2020-12-22 13:50:19 | SAP/ownid-webapp | https://api.github.com/repos/SAP/ownid-webapp | opened | IOS 14 cookies | Priority: High Type: Bug | If the user disabled FIDO, what will happen after 8 days?
Login flow ends with "Account recovery needed" message.
Tested on PROD (demo.ownid.com) with iPhone without FIDO | 1.0 | IOS 14 cookies - If the user disabled FIDO, what will happen after 8 days?
Login flow ends with "Account recovery needed" message.
Tested on PROD (demo.ownid.com) with iPhone without FIDO | priority | ios cookies if the user disabled fido what will happen after days login flow ends with account recovery needed message tested on prod demo ownid com with iphone without fido | 1 |
359,885 | 10,682,144,325 | IssuesEvent | 2019-10-22 03:59:45 | ele-l10n-cjk/wingpanel-indicator-inputmethod | https://api.github.com/repos/ele-l10n-cjk/wingpanel-indicator-inputmethod | opened | Human readable engine names | Priority: High | "mozc-jp" should be "Mozc", "libpynin" should be "Intelligent Pynin", and so on
| 1.0 | Human readable engine names - "mozc-jp" should be "Mozc", "libpynin" should be "Intelligent Pynin", and so on
| priority | human readable engine names mozc jp should be mozc libpynin should be intelligent pynin and so on | 1 |
373,032 | 11,032,092,490 | IssuesEvent | 2019-12-06 19:20:10 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | Giving Reputation to a User with an ASCII username from steam (displayed as ????? ingame) will crash the game of the person giving the reputation | High Priority | Log:
[crash (1).txt](https://github.com/StrangeLoopGames/EcoIssues/files/3933667/crash.1.txt)
| 1.0 | Giving Reputation to a User with an ASCII username from steam (displayed as ????? ingame) will crash the game of the person giving the reputation - Log:
[crash (1).txt](https://github.com/StrangeLoopGames/EcoIssues/files/3933667/crash.1.txt)
| priority | giving reputation to a user with an ascii username from steam displayed as ingame will crash the game of the person giving the reputation log | 1 |
180,350 | 6,648,819,156 | IssuesEvent | 2017-09-28 10:46:53 | minio/minio | https://api.github.com/repos/minio/minio | closed | Prepare first Minio Azure VM | priority: high | Create a json template to download Minio binary and configure it with a azure storage account entered by the user before VM deployment. | 1.0 | Prepare first Minio Azure VM - Create a json template to download Minio binary and configure it with a azure storage account entered by the user before VM deployment. | priority | prepare first minio azure vm create a json template to download minio binary and configure it with a azure storage account entered by the user before vm deployment | 1 |
702,440 | 24,122,321,872 | IssuesEvent | 2022-09-20 19:55:28 | misskey-dev/misskey | https://api.github.com/repos/misskey-dev/misskey | opened | Jestが動作しない | :question:needs more investigation ⚠️bug? 🔥high priority | `SyntaxError: The requested module '@/config.js' does not provide an export named 'Config'`とか出る | 1.0 | Jestが動作しない - `SyntaxError: The requested module '@/config.js' does not provide an export named 'Config'`とか出る | priority | jestが動作しない syntaxerror the requested module config js does not provide an export named config とか出る | 1 |
216,004 | 7,300,015,087 | IssuesEvent | 2018-02-26 22:03:47 | opl-/school-ebook-store | https://api.github.com/repos/opl-/school-ebook-store | closed | [Backend] System logowania | priority-high | Pozwala na łatwe przechowywanie zakupionych książek i zapamiętywanie ustawień użytkownika. | 1.0 | [Backend] System logowania - Pozwala na łatwe przechowywanie zakupionych książek i zapamiętywanie ustawień użytkownika. | priority | system logowania pozwala na łatwe przechowywanie zakupionych książek i zapamiętywanie ustawień użytkownika | 1 |
439,470 | 12,683,117,186 | IssuesEvent | 2020-06-19 18:58:15 | RE-SS3D/SS3D | https://api.github.com/repos/RE-SS3D/SS3D | closed | Add client-side interactions & Examine | High Priority | Allow for interactions that only run client-side, so that we can add interactions such as 'examine'
### Summary
- Any given interaction should be able to 'mark' that it is client-side.
- As a test implement the 'examine' interaction. Should be as extensible as possible.
- It'd probably be a good idea to consult on specifically how the examine interaction is extensible | 1.0 | Add client-side interactions & Examine - Allow for interactions that only run client-side, so that we can add interactions such as 'examine'
### Summary
- Any given interaction should be able to 'mark' that it is client-side.
- As a test implement the 'examine' interaction. Should be as extensible as possible.
- It'd probably be a good idea to consult on specifically how the examine interaction is extensible | priority | add client side interactions examine allow for interactions that only run client side so that we can add interactions such as examine summary any given interaction should be able to mark that it is client side as a test implement the examine interaction should be as extensible as possible it d probably be a good idea to consult on specifically how the examine interaction is extensible | 1 |
678,127 | 23,189,531,641 | IssuesEvent | 2022-08-01 11:23:50 | kubermatic/kubermatic | https://api.github.com/repos/kubermatic/kubermatic | closed | User cannot delete Cluster Template | kind/bug priority/high | ### What happened?
As a project user I was trying to remove one of existing cluster templates. The API request return 200 but probably due to an existing finalizer (kubermatic.k8c.io/cleanup-credentials-secrets) it stays in the terminating state forever.
### Expected behavior
After deleting `clustertemplate` it should be removed from the cluster.
### How to reproduce the issue?
1. Create a project or use one of existing ones.
2. Go to `Cluster Templates` page and create a new cluster template.
3. After the new cluster template is saved go back to the cluster templates page again and try to remove it.
### How is your environment configured?
- KKP version: 2.21 (dev on 29.07.2022) | 1.0 | User cannot delete Cluster Template - ### What happened?
As a project user I was trying to remove one of existing cluster templates. The API request return 200 but probably due to an existing finalizer (kubermatic.k8c.io/cleanup-credentials-secrets) it stays in the terminating state forever.
### Expected behavior
After deleting `clustertemplate` it should be removed from the cluster.
### How to reproduce the issue?
1. Create a project or use one of existing ones.
2. Go to `Cluster Templates` page and create a new cluster template.
3. After the new cluster template is saved go back to the cluster templates page again and try to remove it.
### How is your environment configured?
- KKP version: 2.21 (dev on 29.07.2022) | priority | user cannot delete cluster template what happened as a project user i was trying to remove one of existing cluster templates the api request return but probably due to an existing finalizer kubermatic io cleanup credentials secrets it stays in the terminating state forever expected behavior after deleting clustertemplate it should be removed from the cluster how to reproduce the issue create a project or use one of existing ones go to cluster templates page and create a new cluster template after the new cluster template is saved go back to the cluster templates page again and try to remove it how is your environment configured kkp version dev on | 1 |
653,398 | 21,581,373,702 | IssuesEvent | 2022-05-02 19:07:02 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | WorkQueue fails to create subscription due to ORACLE unique constraint violation | BUG WMAgent WorkQueue Highest Priority | **Impact of the bug**
WorkQueue, WMBS
**Describe the bug**
While checking the status of few workflows upon a P&R request, I've found plenty of Oracle errors: `ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated` at the two newly deployed CERN agents. The agent version deployed is **v2.0.2.patch1**. This is an error we do not observe neither in the FNAL agents (they are using MAriaDB anyway) nor at the agents currently on drain.
Taking as an example a workflow that is been distributed amongst those all agents:
[cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873](https://cmsweb.cern.ch/reqmgr2/fetch?rid=cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873)
**agent url:** cmsgwms-submit4.fnal.gov,cmsgwms-submit5.fnal.gov,cmsgwms-submit6.fnal.gov,cmsgwms-submit7.fnal.gov,vocms0252.cern.ch,vocms0253.cern.ch,vocms0282.cern.ch,vocms0283.cern.ch
We observe the following at `vocms0282 - v2.0.2.patch1`: [1], and no errors at all in both `vocms0252`.
In addition to that I see quite low number of WMBS jobs created for those two new CERN agents. Something like 866 jobs for `vocms0282` vs 43K jobs for `cmsgwms-submit7.fnal.gov`. And hence the low number of idle condor jobs for the two new CERN agents:
```
vocms0282: $ condor_q --totals
Total for query: 6187 jobs; 0 completed, 0 removed, 4224 idle, 1963 running, 0 held, 0 suspended
vocms0283: $ condor_q --totals
Total for query: 30827 jobs; 0 completed, 0 removed, 3695 idle, 27132 running, 0 held, 0 suspended
```
**How to reproduce it**
Steps to reproduce the behavior:
**Expected behavior**
A clear and concise description of what you expected to happen.
**Additional context and error message**
Add any other context about the problem here, like error message and/or traceback. You might want to use triple back ticks to properly format it.
[1]
```
2022-04-13 23:44:52,787:139862773548800:ERROR:WMBSHelper:Failed to create subscription. Error: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
[SQL: INSERT INTO dbsbuffer_dataset_subscription
(id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime)
SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve,
:move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime
FROM DUAL
WHERE NOT EXISTS
( SELECT *
FROM dbsbuffer_dataset_subscription
WHERE dataset_id = :id
AND site = :site
AND custodial = :custodial
AND auto_approve = :auto_approve
AND move = :move
AND priority = :priority
AND dataset_lifetime = :dataset_lifetime )
]
[parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}]
(Background on this error at: http://sqlalche.me/e/gkpj)
Traceback (most recent call last):
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context
self.dialect.do_execute(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute
cursor.execute(statement, parameters)
cx_Oracle.IntegrityError: ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 413, in createSubscriptionAndAddFiles
self.createSubscription(self.topLevelTask, self.topLevelFileset)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 266, in createSubscription
self._createDatasetSubscriptionsInDBSBuffer()
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 520, in _createDatasetSubscriptionsInDBSBuffer
dbsDataset.addSubscription(subInfo[dataset])
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/DBSBufferDataset.py", line 140, in addSubscription
action.execute(self['id'], subscriptionInformation,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/MySQL/NewSubscription.py", line 66, in execute
self.dbi.processData(self.sql, binds=binds, conn=conn, transaction=transaction)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 172, in processData
r = self.executebinds(s, b, connection=connection,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 64, in executebinds
resultProxy = connection.execute(s, b)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 982, in execute
return self._execute_text(object_, multiparams, params)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1149, in _execute_text
ret = self._execute_context(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1247, in _execute_context
self._handle_dbapi_exception(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1466, in _handle_dbapi_exception
util.raise_from_cause(sqlalchemy_exception, exc_info)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 383, in raise_from_cause
reraise(type(exception), exception, tb=exc_tb, cause=cause)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 128, in reraise
raise value.with_traceback(tb)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context
self.dialect.do_execute(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute
cursor.execute(statement, parameters)
sqlalchemy.exc.IntegrityError: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
[SQL: INSERT INTO dbsbuffer_dataset_subscription
(id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime)
SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve,
:move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime
FROM DUAL
WHERE NOT EXISTS
( SELECT *
FROM dbsbuffer_dataset_subscription
WHERE dataset_id = :id
AND site = :site
AND custodial = :custodial
AND auto_approve = :auto_approve
AND move = :move
AND priority = :priority
AND dataset_lifetime = :dataset_lifetime )
]
[parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}]
(Background on this error at: http://sqlalche.me/e/gkpj)
2022-04-13 23:44:52,789:139862773548800:ERROR:WorkQueue:Failed to create subscription for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 with block name None
Error: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
[SQL: INSERT INTO dbsbuffer_dataset_subscription
(id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime)
SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve,
:move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime
FROM DUAL
WHERE NOT EXISTS
( SELECT *
FROM dbsbuffer_dataset_subscription
WHERE dataset_id = :id
AND site = :site
AND custodial = :custodial
AND auto_approve = :auto_approve
AND move = :move
AND priority = :priority
AND dataset_lifetime = :dataset_lifetime )
]
[parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}]
(Background on this error at: http://sqlalche.me/e/gkpj)
Traceback (most recent call last):
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context
self.dialect.do_execute(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute
cursor.execute(statement, parameters)
cx_Oracle.IntegrityError: ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WorkQueue.py", line 352, in getWork
match['Subscription'] = self._wmbsPreparation(match,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WorkQueue.py", line 453, in _wmbsPreparation
sub, match['NumOfFilesAdded'] = wmbsHelper.createSubscriptionAndAddFiles(block=dbsBlock)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 418, in createSubscriptionAndAddFiles
raise ex
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 413, in createSubscriptionAndAddFiles
self.createSubscription(self.topLevelTask, self.topLevelFileset)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 266, in createSubscription
self._createDatasetSubscriptionsInDBSBuffer()
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 520, in _createDatasetSubscriptionsInDBSBuffer
dbsDataset.addSubscription(subInfo[dataset])
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/DBSBufferDataset.py", line 140, in addSubscription
action.execute(self['id'], subscriptionInformation,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/MySQL/NewSubscription.py", line 66, in execute
self.dbi.processData(self.sql, binds=binds, conn=conn, transaction=transaction)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 172, in processData
r = self.executebinds(s, b, connection=connection,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 64, in executebinds
resultProxy = connection.execute(s, b)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 982, in execute
return self._execute_text(object_, multiparams, params)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1149, in _execute_text
ret = self._execute_context(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1247, in _execute_context
self._handle_dbapi_exception(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1466, in _handle_dbapi_exception
util.raise_from_cause(sqlalchemy_exception, exc_info)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 383, in raise_from_cause
reraise(type(exception), exception, tb=exc_tb, cause=cause)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 128, in reraise
raise value.with_traceback(tb)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context
self.dialect.do_execute(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute
cursor.execute(statement, parameters)
sqlalchemy.exc.IntegrityError: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
[SQL: INSERT INTO dbsbuffer_dataset_subscription
(id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime)
SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve,
:move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime
FROM DUAL
WHERE NOT EXISTS
( SELECT *
FROM dbsbuffer_dataset_subscription
WHERE dataset_id = :id
AND site = :site
AND custodial = :custodial
AND auto_approve = :auto_approve
AND move = :move
AND priority = :priority
AND dataset_lifetime = :dataset_lifetime )
]
[parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}]
(Background on this error at: http://sqlalche.me/e/gkpj)
```
[2]
```
2022-04-11 10:24:24,733:140665460737792:INFO:WorkQueue:Running WMBS preparation for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 with ParentQueueId 84cadfd2061398ef871a19d9bb5d7466,
with common location ['T2_US_Florida']
2022-04-11 10:24:24,744:140665460737792:INFO:Fileset:Fileset created: cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873-TSG-Run3Winter22GS-00002_0-27765ea9ee953e8b90915707e32347a3
2022-04-11 10:24:26,349:140665460737792:INFO:Rucio:WMCore Rucio initialization parameters: {'account': 'wmcore_transferor', 'rucio_host': None, 'auth_host': None, 'ca_cert': None, 'auth_type': None, 'creds': None, 'timeout': 600, 'user_agent': 'wmcore-client'}
2022-04-11 10:24:26,352:140665460737792:INFO:Rucio:Rucio client initialization parameters: {'host': 'http://cms-rucio.cern.ch', 'auth_host': 'https://cms-rucio-auth.cern.ch', 'auth_type': 'x509', 'account': 'wmcore_transferor', 'user_agent': 'wmcore-client/1.25.5', 'ca_cert': '/etc/grid-security/certificates/', 'creds': {'client_cert': '/data/certs/servicecert.pem', 'client_key': '/data/certs/servicekey.pem'}, 'timeout': 600, 'request_retries': 3}
2022-04-11 10:24:26,462:140665460737792:INFO:Rucio:Pileup container location for /MinBias_TuneCP5_13p6TeV-pythia8/Run3Winter22GS-122X_mcRun3_2021_realistic_v9-v1/GEN-SIM from single RSE locks at: ['T1_US_FNAL_Disk', 'T2_US_Florida', 'T1_ES_PIC_Disk']
2022-04-11 10:24:28,757:140665460737792:INFO:SandboxCreator:Created sandbox cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873-Sandbox.tar.bz2 with size 3197422
2022-04-11 10:24:28,776:140665460737792:INFO:Workflow:Workflow id 47920 created for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873
2022-04-11 10:24:28,788:140665460737792:INFO:WMBSHelper:Top level subscription 67000 created for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873
2022-04-11 10:24:28,793:140665460737792:INFO:Fileset:Fileset created: /cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873/TSG-Run3Winter22GS-00002_0/unmerged-RAWSIMoutputGEN-SIM
```
[3]
```
``` | 1.0 | WorkQueue fails to create subscription due to ORACLE unique constraint violation - **Impact of the bug**
WorkQueue, WMBS
**Describe the bug**
While checking the status of few workflows upon a P&R request, I've found plenty of Oracle errors: `ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated` at the two newly deployed CERN agents. The agent version deployed is **v2.0.2.patch1**. This is an error we do not observe neither in the FNAL agents (they are using MAriaDB anyway) nor at the agents currently on drain.
Taking as an example a workflow that is been distributed amongst those all agents:
[cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873](https://cmsweb.cern.ch/reqmgr2/fetch?rid=cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873)
**agent url:** cmsgwms-submit4.fnal.gov,cmsgwms-submit5.fnal.gov,cmsgwms-submit6.fnal.gov,cmsgwms-submit7.fnal.gov,vocms0252.cern.ch,vocms0253.cern.ch,vocms0282.cern.ch,vocms0283.cern.ch
We observe the following at `vocms0282 - v2.0.2.patch1`: [1], and no errors at all in both `vocms0252`.
In addition to that I see quite low number of WMBS jobs created for those two new CERN agents. Something like 866 jobs for `vocms0282` vs 43K jobs for `cmsgwms-submit7.fnal.gov`. And hence the low number of idle condor jobs for the two new CERN agents:
```
vocms0282: $ condor_q --totals
Total for query: 6187 jobs; 0 completed, 0 removed, 4224 idle, 1963 running, 0 held, 0 suspended
vocms0283: $ condor_q --totals
Total for query: 30827 jobs; 0 completed, 0 removed, 3695 idle, 27132 running, 0 held, 0 suspended
```
**How to reproduce it**
Steps to reproduce the behavior:
**Expected behavior**
A clear and concise description of what you expected to happen.
**Additional context and error message**
Add any other context about the problem here, like error message and/or traceback. You might want to use triple back ticks to properly format it.
[1]
```
2022-04-13 23:44:52,787:139862773548800:ERROR:WMBSHelper:Failed to create subscription. Error: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
[SQL: INSERT INTO dbsbuffer_dataset_subscription
(id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime)
SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve,
:move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime
FROM DUAL
WHERE NOT EXISTS
( SELECT *
FROM dbsbuffer_dataset_subscription
WHERE dataset_id = :id
AND site = :site
AND custodial = :custodial
AND auto_approve = :auto_approve
AND move = :move
AND priority = :priority
AND dataset_lifetime = :dataset_lifetime )
]
[parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}]
(Background on this error at: http://sqlalche.me/e/gkpj)
Traceback (most recent call last):
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context
self.dialect.do_execute(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute
cursor.execute(statement, parameters)
cx_Oracle.IntegrityError: ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 413, in createSubscriptionAndAddFiles
self.createSubscription(self.topLevelTask, self.topLevelFileset)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 266, in createSubscription
self._createDatasetSubscriptionsInDBSBuffer()
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 520, in _createDatasetSubscriptionsInDBSBuffer
dbsDataset.addSubscription(subInfo[dataset])
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/DBSBufferDataset.py", line 140, in addSubscription
action.execute(self['id'], subscriptionInformation,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/MySQL/NewSubscription.py", line 66, in execute
self.dbi.processData(self.sql, binds=binds, conn=conn, transaction=transaction)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 172, in processData
r = self.executebinds(s, b, connection=connection,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 64, in executebinds
resultProxy = connection.execute(s, b)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 982, in execute
return self._execute_text(object_, multiparams, params)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1149, in _execute_text
ret = self._execute_context(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1247, in _execute_context
self._handle_dbapi_exception(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1466, in _handle_dbapi_exception
util.raise_from_cause(sqlalchemy_exception, exc_info)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 383, in raise_from_cause
reraise(type(exception), exception, tb=exc_tb, cause=cause)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 128, in reraise
raise value.with_traceback(tb)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context
self.dialect.do_execute(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute
cursor.execute(statement, parameters)
sqlalchemy.exc.IntegrityError: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
[SQL: INSERT INTO dbsbuffer_dataset_subscription
(id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime)
SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve,
:move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime
FROM DUAL
WHERE NOT EXISTS
( SELECT *
FROM dbsbuffer_dataset_subscription
WHERE dataset_id = :id
AND site = :site
AND custodial = :custodial
AND auto_approve = :auto_approve
AND move = :move
AND priority = :priority
AND dataset_lifetime = :dataset_lifetime )
]
[parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}]
(Background on this error at: http://sqlalche.me/e/gkpj)
2022-04-13 23:44:52,789:139862773548800:ERROR:WorkQueue:Failed to create subscription for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 with block name None
Error: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
[SQL: INSERT INTO dbsbuffer_dataset_subscription
(id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime)
SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve,
:move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime
FROM DUAL
WHERE NOT EXISTS
( SELECT *
FROM dbsbuffer_dataset_subscription
WHERE dataset_id = :id
AND site = :site
AND custodial = :custodial
AND auto_approve = :auto_approve
AND move = :move
AND priority = :priority
AND dataset_lifetime = :dataset_lifetime )
]
[parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}]
(Background on this error at: http://sqlalche.me/e/gkpj)
Traceback (most recent call last):
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context
self.dialect.do_execute(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute
cursor.execute(statement, parameters)
cx_Oracle.IntegrityError: ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WorkQueue.py", line 352, in getWork
match['Subscription'] = self._wmbsPreparation(match,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WorkQueue.py", line 453, in _wmbsPreparation
sub, match['NumOfFilesAdded'] = wmbsHelper.createSubscriptionAndAddFiles(block=dbsBlock)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 418, in createSubscriptionAndAddFiles
raise ex
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 413, in createSubscriptionAndAddFiles
self.createSubscription(self.topLevelTask, self.topLevelFileset)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 266, in createSubscription
self._createDatasetSubscriptionsInDBSBuffer()
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 520, in _createDatasetSubscriptionsInDBSBuffer
dbsDataset.addSubscription(subInfo[dataset])
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/DBSBufferDataset.py", line 140, in addSubscription
action.execute(self['id'], subscriptionInformation,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/MySQL/NewSubscription.py", line 66, in execute
self.dbi.processData(self.sql, binds=binds, conn=conn, transaction=transaction)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 172, in processData
r = self.executebinds(s, b, connection=connection,
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 64, in executebinds
resultProxy = connection.execute(s, b)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 982, in execute
return self._execute_text(object_, multiparams, params)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1149, in _execute_text
ret = self._execute_context(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1247, in _execute_context
self._handle_dbapi_exception(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1466, in _handle_dbapi_exception
util.raise_from_cause(sqlalchemy_exception, exc_info)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 383, in raise_from_cause
reraise(type(exception), exception, tb=exc_tb, cause=cause)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 128, in reraise
raise value.with_traceback(tb)
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context
self.dialect.do_execute(
File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute
cursor.execute(statement, parameters)
sqlalchemy.exc.IntegrityError: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated
[SQL: INSERT INTO dbsbuffer_dataset_subscription
(id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime)
SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve,
:move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime
FROM DUAL
WHERE NOT EXISTS
( SELECT *
FROM dbsbuffer_dataset_subscription
WHERE dataset_id = :id
AND site = :site
AND custodial = :custodial
AND auto_approve = :auto_approve
AND move = :move
AND priority = :priority
AND dataset_lifetime = :dataset_lifetime )
]
[parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}]
(Background on this error at: http://sqlalche.me/e/gkpj)
```
[2]
```
2022-04-11 10:24:24,733:140665460737792:INFO:WorkQueue:Running WMBS preparation for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 with ParentQueueId 84cadfd2061398ef871a19d9bb5d7466,
with common location ['T2_US_Florida']
2022-04-11 10:24:24,744:140665460737792:INFO:Fileset:Fileset created: cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873-TSG-Run3Winter22GS-00002_0-27765ea9ee953e8b90915707e32347a3
2022-04-11 10:24:26,349:140665460737792:INFO:Rucio:WMCore Rucio initialization parameters: {'account': 'wmcore_transferor', 'rucio_host': None, 'auth_host': None, 'ca_cert': None, 'auth_type': None, 'creds': None, 'timeout': 600, 'user_agent': 'wmcore-client'}
2022-04-11 10:24:26,352:140665460737792:INFO:Rucio:Rucio client initialization parameters: {'host': 'http://cms-rucio.cern.ch', 'auth_host': 'https://cms-rucio-auth.cern.ch', 'auth_type': 'x509', 'account': 'wmcore_transferor', 'user_agent': 'wmcore-client/1.25.5', 'ca_cert': '/etc/grid-security/certificates/', 'creds': {'client_cert': '/data/certs/servicecert.pem', 'client_key': '/data/certs/servicekey.pem'}, 'timeout': 600, 'request_retries': 3}
2022-04-11 10:24:26,462:140665460737792:INFO:Rucio:Pileup container location for /MinBias_TuneCP5_13p6TeV-pythia8/Run3Winter22GS-122X_mcRun3_2021_realistic_v9-v1/GEN-SIM from single RSE locks at: ['T1_US_FNAL_Disk', 'T2_US_Florida', 'T1_ES_PIC_Disk']
2022-04-11 10:24:28,757:140665460737792:INFO:SandboxCreator:Created sandbox cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873-Sandbox.tar.bz2 with size 3197422
2022-04-11 10:24:28,776:140665460737792:INFO:Workflow:Workflow id 47920 created for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873
2022-04-11 10:24:28,788:140665460737792:INFO:WMBSHelper:Top level subscription 67000 created for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873
2022-04-11 10:24:28,793:140665460737792:INFO:Fileset:Fileset created: /cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873/TSG-Run3Winter22GS-00002_0/unmerged-RAWSIMoutputGEN-SIM
```
[3]
```
``` | priority | workqueue fails to create subscription due to oracle unique constraint violation impact of the bug workqueue wmbs describe the bug while checking the status of few workflows upon a p r request i ve found plenty of oracle errors ora unique constraint cms wmbs uq dbs dat sub violated at the two newly deployed cern agents the agent version deployed is this is an error we do not observe neither in the fnal agents they are using mariadb anyway nor at the agents currently on drain taking as an example a workflow that is been distributed amongst those all agents agent url cmsgwms fnal gov cmsgwms fnal gov cmsgwms fnal gov cmsgwms fnal gov cern ch cern ch cern ch cern ch we observe the following at and no errors at all in both in addition to that i see quite low number of wmbs jobs created for those two new cern agents something like jobs for vs jobs for cmsgwms fnal gov and hence the low number of idle condor jobs for the two new cern agents condor q totals total for query jobs completed removed idle running held suspended condor q totals total for query jobs completed removed idle running held suspended how to reproduce it steps to reproduce the behavior expected behavior a clear and concise description of what you expected to happen additional context and error message add any other context about the problem here like error message and or traceback you might want to use triple back ticks to properly format it error wmbshelper failed to create subscription error cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated sql insert into dbsbuffer dataset subscription id dataset id site custodial auto approve move priority subscribed phedex group delete blocks dataset lifetime select dbsbuffer dataset sub seq nextval id site custodial auto approve move priority phedex group delete blocks dataset lifetime from dual where not exists select from dbsbuffer dataset subscription where dataset id id and site site and custodial custodial and auto approve auto approve and move move and priority priority and dataset lifetime dataset lifetime background on this error at traceback most recent call last file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self dialect do execute file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated the above exception was the direct cause of the following exception traceback most recent call last file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscriptionandaddfiles self createsubscription self topleveltask self toplevelfileset file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscription self createdatasetsubscriptionsindbsbuffer file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createdatasetsubscriptionsindbsbuffer dbsdataset addsubscription subinfo file data srv wmagent sw cms lib site packages wmcomponent dbsbufferdataset py line in addsubscription action execute self subscriptioninformation file data srv wmagent sw cms lib site packages wmcomponent mysql newsubscription py line in execute self dbi processdata self sql binds binds conn conn transaction transaction file data srv wmagent sw cms lib site packages wmcore database dbcore py line in processdata r self executebinds s b connection connection file data srv wmagent sw cms lib site packages wmcore database dbcore py line in executebinds resultproxy connection execute s b file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute return self execute text object multiparams params file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute text ret self execute context file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self handle dbapi exception file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in handle dbapi exception util raise from cause sqlalchemy exception exc info file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy util compat py line in raise from cause reraise type exception exception tb exc tb cause cause file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy util compat py line in reraise raise value with traceback tb file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self dialect do execute file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters sqlalchemy exc integrityerror cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated sql insert into dbsbuffer dataset subscription id dataset id site custodial auto approve move priority subscribed phedex group delete blocks dataset lifetime select dbsbuffer dataset sub seq nextval id site custodial auto approve move priority phedex group delete blocks dataset lifetime from dual where not exists select from dbsbuffer dataset subscription where dataset id id and site site and custodial custodial and auto approve auto approve and move move and priority priority and dataset lifetime dataset lifetime background on this error at error workqueue failed to create subscription for cmsunified task tsg t with block name none error cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated sql insert into dbsbuffer dataset subscription id dataset id site custodial auto approve move priority subscribed phedex group delete blocks dataset lifetime select dbsbuffer dataset sub seq nextval id site custodial auto approve move priority phedex group delete blocks dataset lifetime from dual where not exists select from dbsbuffer dataset subscription where dataset id id and site site and custodial custodial and auto approve auto approve and move move and priority priority and dataset lifetime dataset lifetime background on this error at traceback most recent call last file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self dialect do execute file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated the above exception was the direct cause of the following exception traceback most recent call last file data srv wmagent sw cms lib site packages wmcore workqueue workqueue py line in getwork match self wmbspreparation match file data srv wmagent sw cms lib site packages wmcore workqueue workqueue py line in wmbspreparation sub match wmbshelper createsubscriptionandaddfiles block dbsblock file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscriptionandaddfiles raise ex file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscriptionandaddfiles self createsubscription self topleveltask self toplevelfileset file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscription self createdatasetsubscriptionsindbsbuffer file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createdatasetsubscriptionsindbsbuffer dbsdataset addsubscription subinfo file data srv wmagent sw cms lib site packages wmcomponent dbsbufferdataset py line in addsubscription action execute self subscriptioninformation file data srv wmagent sw cms lib site packages wmcomponent mysql newsubscription py line in execute self dbi processdata self sql binds binds conn conn transaction transaction file data srv wmagent sw cms lib site packages wmcore database dbcore py line in processdata r self executebinds s b connection connection file data srv wmagent sw cms lib site packages wmcore database dbcore py line in executebinds resultproxy connection execute s b file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute return self execute text object multiparams params file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute text ret self execute context file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self handle dbapi exception file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in handle dbapi exception util raise from cause sqlalchemy exception exc info file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy util compat py line in raise from cause reraise type exception exception tb exc tb cause cause file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy util compat py line in reraise raise value with traceback tb file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self dialect do execute file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters sqlalchemy exc integrityerror cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated sql insert into dbsbuffer dataset subscription id dataset id site custodial auto approve move priority subscribed phedex group delete blocks dataset lifetime select dbsbuffer dataset sub seq nextval id site custodial auto approve move priority phedex group delete blocks dataset lifetime from dual where not exists select from dbsbuffer dataset subscription where dataset id id and site site and custodial custodial and auto approve auto approve and move move and priority priority and dataset lifetime dataset lifetime background on this error at info workqueue running wmbs preparation for cmsunified task tsg t with parentqueueid with common location info fileset fileset created cmsunified task tsg t tsg info rucio wmcore rucio initialization parameters account wmcore transferor rucio host none auth host none ca cert none auth type none creds none timeout user agent wmcore client info rucio rucio client initialization parameters host auth host auth type account wmcore transferor user agent wmcore client ca cert etc grid security certificates creds client cert data certs servicecert pem client key data certs servicekey pem timeout request retries info rucio pileup container location for minbias realistic gen sim from single rse locks at info sandboxcreator created sandbox cmsunified task tsg t sandbox tar with size info workflow workflow id created for cmsunified task tsg t info wmbshelper top level subscription created for cmsunified task tsg t info fileset fileset created cmsunified task tsg t tsg unmerged rawsimoutputgen sim | 1 |
315,145 | 9,606,963,978 | IssuesEvent | 2019-05-11 15:00:31 | thechutrain/rc-coffee-chats | https://api.github.com/repos/thechutrain/rc-coffee-chats | closed | Change the fallback user to someone else during NGW | high priority | Facilitators will be busy during NGW organizing events, so it would be best to have the fallback user be a current or a Recurser who would be fine without a coffee chat. | 1.0 | Change the fallback user to someone else during NGW - Facilitators will be busy during NGW organizing events, so it would be best to have the fallback user be a current or a Recurser who would be fine without a coffee chat. | priority | change the fallback user to someone else during ngw facilitators will be busy during ngw organizing events so it would be best to have the fallback user be a current or a recurser who would be fine without a coffee chat | 1 |
650,764 | 21,416,721,590 | IssuesEvent | 2022-04-22 11:37:24 | lucypoulton/acebook | https://api.github.com/repos/lucypoulton/acebook | closed | Users can comment on posts | size: large type: feature priority: high | - [ ] Users can interact with a post and leave a comment.
- [ ] User comments can be seen underneath the original post
- [ ] Posted comments have the Commenters/Users name next to them | 1.0 | Users can comment on posts - - [ ] Users can interact with a post and leave a comment.
- [ ] User comments can be seen underneath the original post
- [ ] Posted comments have the Commenters/Users name next to them | priority | users can comment on posts users can interact with a post and leave a comment user comments can be seen underneath the original post posted comments have the commenters users name next to them | 1 |
98,455 | 4,021,535,905 | IssuesEvent | 2016-05-16 22:22:03 | theBoardlist/website | https://api.github.com/repos/theBoardlist/website | opened | Unable to add email address to account with no email | bug High Priority | Steps to reproduce:
1. Go to a profile page for an account with no email address associated with it; e.g., Amy Banse
2. Click edit > add email address
3. Enter an email that isn't already associated with an account and click SAVE
Expected result:
Email added to account, profile updated
Actual result:
No email added, profile not updated. We get the screen that we normally see if we're trying to add an email address that's already in the system.

| 1.0 | Unable to add email address to account with no email - Steps to reproduce:
1. Go to a profile page for an account with no email address associated with it; e.g., Amy Banse
2. Click edit > add email address
3. Enter an email that isn't already associated with an account and click SAVE
Expected result:
Email added to account, profile updated
Actual result:
No email added, profile not updated. We get the screen that we normally see if we're trying to add an email address that's already in the system.

| priority | unable to add email address to account with no email steps to reproduce go to a profile page for an account with no email address associated with it e g amy banse click edit add email address enter an email that isn t already associated with an account and click save expected result email added to account profile updated actual result no email added profile not updated we get the screen that we normally see if we re trying to add an email address that s already in the system | 1 |
605,614 | 18,737,748,197 | IssuesEvent | 2021-11-04 09:52:32 | betagouv/service-national-universel | https://api.github.com/repos/betagouv/service-national-universel | closed | fix(inscription): préparer dashboard pour inscription 2022 | enhancement priority-HIGH inscription | ### Fonctionnalité liée à un problème ?
_No response_
### Fonctionnalité
il faut que le dashboard, onglet inscirpiton. faire reference a la cohorte 2022.
### Commentaires
_No response_ | 1.0 | fix(inscription): préparer dashboard pour inscription 2022 - ### Fonctionnalité liée à un problème ?
_No response_
### Fonctionnalité
il faut que le dashboard, onglet inscirpiton. faire reference a la cohorte 2022.
### Commentaires
_No response_ | priority | fix inscription préparer dashboard pour inscription fonctionnalité liée à un problème no response fonctionnalité il faut que le dashboard onglet inscirpiton faire reference a la cohorte commentaires no response | 1 |
400,490 | 11,776,002,985 | IssuesEvent | 2020-03-16 12:27:09 | HE-Arc/CSRuby | https://api.github.com/repos/HE-Arc/CSRuby | closed | Mise en place de l'inscription | core high priority | Inscription
- email
- profilename
- password
- password confirmation | 1.0 | Mise en place de l'inscription - Inscription
- email
- profilename
- password
- password confirmation | priority | mise en place de l inscription inscription email profilename password password confirmation | 1 |
534,175 | 15,611,414,397 | IssuesEvent | 2021-03-19 14:17:53 | MaibornWolff/codecharta | https://api.github.com/repos/MaibornWolff/codecharta | opened | Indicate total nodes and excluded / flattened nodes in file explorer | UX / UI difficulty:low feature pr-visualization priority:high | # Feature request
## Description
As a user, I want to see the total number of nodes, excluded nodes and flattened nodes so that the numbers correspond to the matching nodes.


## Acceptance criteria
- If the search bar is empty, show the total number of nodes, excluded nodes and flattened nodes
- As soon as the search is triggered show the number of matching nodes and the total nodes together as indicated in the screenshot (`matchingNodes/totalNodes`)
| 1.0 | Indicate total nodes and excluded / flattened nodes in file explorer - # Feature request
## Description
As a user, I want to see the total number of nodes, excluded nodes and flattened nodes so that the numbers correspond to the matching nodes.


## Acceptance criteria
- If the search bar is empty, show the total number of nodes, excluded nodes and flattened nodes
- As soon as the search is triggered show the number of matching nodes and the total nodes together as indicated in the screenshot (`matchingNodes/totalNodes`)
| priority | indicate total nodes and excluded flattened nodes in file explorer feature request description as a user i want to see the total number of nodes excluded nodes and flattened nodes so that the numbers correspond to the matching nodes acceptance criteria if the search bar is empty show the total number of nodes excluded nodes and flattened nodes as soon as the search is triggered show the number of matching nodes and the total nodes together as indicated in the screenshot matchingnodes totalnodes | 1 |
496,626 | 14,350,593,228 | IssuesEvent | 2020-11-29 21:33:29 | swharden/ScottPlot | https://api.github.com/repos/swharden/ScottPlot | closed | Road to ScottPlot 4.1 | HIGH PRIORITY | ### Update on November 17, 2020
**⚠️ I am reducing effort on issues and PRs while I work on this daily.**
* See [changelog.md](https://github.com/swharden/ScottPlot/blob/master/dev/changelog.md) for a summary of major changes.
* See [roadmap.md](https://github.com/swharden/ScottPlot/blob/master/dev/roadmap.md) for a big-picture discussion of plans and goals
* I merged #605 which accomplishes most of the goals described on this page
* The master branch is now ScottPlot `4.1-beta`. Old source code has its own [4.0-stable](https://github.com/swharden/ScottPlot/tree/4.0-stable) branch.
### Plottables to Refactor
- [x] Candlestick / OHLC - [PlottableOHLC.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableOHLC.cs)
- [x] Annotation - [PlottableText.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableText.cs)
- [x] Bar - [PlottableBar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableBar.cs)
- [x] ErrorBar - [PlottableErrorBars.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableErrorBars.cs)
- [x] Function - [PlottableFunction.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableFunction.cs)
- [x] HLine - [PlottableAxisLine.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisLine.cs)
- [x] HSpan - [PlottableAxisSpan.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisSpan.cs)
- [x] VLine - [PlottableAxisLine.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisLine.cs)
- [x] VSpan - [PlottableAxisSpan.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisSpan.cs)
- [x] Heatmap - [PlottableHeatmap.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableHeatmap.cs)
- [x] Image - [PlottableImage.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableImage.cs)
- [x] Pie - [PlottablePie](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePie.cs)
- [x] Polygon - [PlottablePolygon.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePolygon.cs)
- [x] Polygons - [PlottablePolygons.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePolygons.cs)
- [x] Populations - [PlottablePopulations.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePopulations.cs)
- [x] Radar - [PlottableRadar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableRadar.cs)
- [x] ScaleBar - [PlottableScaleBar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScaleBar.cs)
- [x] Scatter / ScatterHighlight - [PlottableScatter.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScatter.cs) [PlottableScatterHighlight.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScatterHighlight.cs)
- [x] Signal / SignalConst / SignalXY (#585)
- [x] Text - [PlottableText.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableText.cs)
- [x] VectorField - [PlottableVectorField.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableVectorField.cs)
### Renderables to Refactor
- [x] Axis labels
- [x] Axis tick marks
- [x] Axis tick labels
- [x] Grid lines
- [x] Figure and Data background
- [x] Error message
- [x] Benchmark message
### Refactor / Large Features
- [x] Delete the `Plottable` class and replace it with `IPlottable`
- [x] Change plottable namespaces (`PlottableScatter` becomes `Plottable.Scatter`)
- [x] support multiple X and Y axes ✨
### Remaining Tasks
- [x] Make GetAxisLimits() return AxisLimits object
- [x] Create a control to test a timer-based render method
- [x] Create new plottables specifically for growing data (mark public data fields readonly and provide setters to replace data arrays)
- [x] Improve data validation for plottables (create separate pre-render and deep validation methods in an interface)
### Extra release of 4.0
- [x] Add `RenderLock()` and `RenderUnlock()` fields to help with multi-threaded plot manipulation
- [x] Add NuGet message noting transition to 4.1 and to check preleases
- [x] Review all the items on #412 since many have now been completed or made obsolete | 1.0 | Road to ScottPlot 4.1 - ### Update on November 17, 2020
**⚠️ I am reducing effort on issues and PRs while I work on this daily.**
* See [changelog.md](https://github.com/swharden/ScottPlot/blob/master/dev/changelog.md) for a summary of major changes.
* See [roadmap.md](https://github.com/swharden/ScottPlot/blob/master/dev/roadmap.md) for a big-picture discussion of plans and goals
* I merged #605 which accomplishes most of the goals described on this page
* The master branch is now ScottPlot `4.1-beta`. Old source code has its own [4.0-stable](https://github.com/swharden/ScottPlot/tree/4.0-stable) branch.
### Plottables to Refactor
- [x] Candlestick / OHLC - [PlottableOHLC.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableOHLC.cs)
- [x] Annotation - [PlottableText.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableText.cs)
- [x] Bar - [PlottableBar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableBar.cs)
- [x] ErrorBar - [PlottableErrorBars.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableErrorBars.cs)
- [x] Function - [PlottableFunction.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableFunction.cs)
- [x] HLine - [PlottableAxisLine.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisLine.cs)
- [x] HSpan - [PlottableAxisSpan.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisSpan.cs)
- [x] VLine - [PlottableAxisLine.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisLine.cs)
- [x] VSpan - [PlottableAxisSpan.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisSpan.cs)
- [x] Heatmap - [PlottableHeatmap.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableHeatmap.cs)
- [x] Image - [PlottableImage.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableImage.cs)
- [x] Pie - [PlottablePie](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePie.cs)
- [x] Polygon - [PlottablePolygon.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePolygon.cs)
- [x] Polygons - [PlottablePolygons.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePolygons.cs)
- [x] Populations - [PlottablePopulations.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePopulations.cs)
- [x] Radar - [PlottableRadar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableRadar.cs)
- [x] ScaleBar - [PlottableScaleBar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScaleBar.cs)
- [x] Scatter / ScatterHighlight - [PlottableScatter.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScatter.cs) [PlottableScatterHighlight.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScatterHighlight.cs)
- [x] Signal / SignalConst / SignalXY (#585)
- [x] Text - [PlottableText.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableText.cs)
- [x] VectorField - [PlottableVectorField.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableVectorField.cs)
### Renderables to Refactor
- [x] Axis labels
- [x] Axis tick marks
- [x] Axis tick labels
- [x] Grid lines
- [x] Figure and Data background
- [x] Error message
- [x] Benchmark message
### Refactor / Large Features
- [x] Delete the `Plottable` class and replace it with `IPlottable`
- [x] Change plottable namespaces (`PlottableScatter` becomes `Plottable.Scatter`)
- [x] support multiple X and Y axes ✨
### Remaining Tasks
- [x] Make GetAxisLimits() return AxisLimits object
- [x] Create a control to test a timer-based render method
- [x] Create new plottables specifically for growing data (mark public data fields readonly and provide setters to replace data arrays)
- [x] Improve data validation for plottables (create separate pre-render and deep validation methods in an interface)
### Extra release of 4.0
- [x] Add `RenderLock()` and `RenderUnlock()` fields to help with multi-threaded plot manipulation
- [x] Add NuGet message noting transition to 4.1 and to check preleases
- [x] Review all the items on #412 since many have now been completed or made obsolete | priority | road to scottplot update on november ⚠️ i am reducing effort on issues and prs while i work on this daily see for a summary of major changes see for a big picture discussion of plans and goals i merged which accomplishes most of the goals described on this page the master branch is now scottplot beta old source code has its own branch plottables to refactor candlestick ohlc annotation bar errorbar function hline hspan vline vspan heatmap image pie polygon polygons populations radar scalebar scatter scatterhighlight signal signalconst signalxy text vectorfield renderables to refactor axis labels axis tick marks axis tick labels grid lines figure and data background error message benchmark message refactor large features delete the plottable class and replace it with iplottable change plottable namespaces plottablescatter becomes plottable scatter support multiple x and y axes ✨ remaining tasks make getaxislimits return axislimits object create a control to test a timer based render method create new plottables specifically for growing data mark public data fields readonly and provide setters to replace data arrays improve data validation for plottables create separate pre render and deep validation methods in an interface extra release of add renderlock and renderunlock fields to help with multi threaded plot manipulation add nuget message noting transition to and to check preleases review all the items on since many have now been completed or made obsolete | 1 |
522,217 | 15,158,189,119 | IssuesEvent | 2021-02-12 00:32:28 | NOAA-GSL/MATS | https://api.github.com/repos/NOAA-GSL/MATS | closed | METexpress surface app does not plot 2 curves correctly | Priority: High Project: MATS Status: Closed Type: Bug | ---
Author Name: **bonny.strong** (@bonnystrong)
Original Redmine Issue: 64018, https://vlab.ncep.noaa.gov/redmine/issues/64018
Original Date: 2019-05-15
Original Assignee: molly.b.smith
---
I tried plotting 2 different curves from different MET databases. I'm attaching screenshots where I plotted each curve separately, then when I tried to plot both together. They have data over different time intervals, but when I did the plot together I changed the time interval to a range that included both. But the plot changed the time to only that of the second curve, then reported no data for the first curve.
Also, in producing this issue, I've noticed that there is nothing on the graph itself that identifies the database that was used. I think we need to have this, or the titles are not definitive.
---
- [METe-srf-plot1-defaults.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22717/METe-srf-plot1-defaults.JPG) (bonny.strong)
- [METe-src-plot2.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22718/METe-src-plot2.JPG) (bonny.strong)
- [METe-src-plot1and2.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22719/METe-src-plot1and2.JPG) (bonny.strong)
- [Screen Shot 2019-05-21 at 12.00.42 PM.png](https://vlab.ncep.noaa.gov/redmine/attachments/download/22810/Screen%20Shot%202019-05-21%20at%2012.00.42%20PM.png) (molly.b.smith) | 1.0 | METexpress surface app does not plot 2 curves correctly - ---
Author Name: **bonny.strong** (@bonnystrong)
Original Redmine Issue: 64018, https://vlab.ncep.noaa.gov/redmine/issues/64018
Original Date: 2019-05-15
Original Assignee: molly.b.smith
---
I tried plotting 2 different curves from different MET databases. I'm attaching screenshots where I plotted each curve separately, then when I tried to plot both together. They have data over different time intervals, but when I did the plot together I changed the time interval to a range that included both. But the plot changed the time to only that of the second curve, then reported no data for the first curve.
Also, in producing this issue, I've noticed that there is nothing on the graph itself that identifies the database that was used. I think we need to have this, or the titles are not definitive.
---
- [METe-srf-plot1-defaults.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22717/METe-srf-plot1-defaults.JPG) (bonny.strong)
- [METe-src-plot2.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22718/METe-src-plot2.JPG) (bonny.strong)
- [METe-src-plot1and2.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22719/METe-src-plot1and2.JPG) (bonny.strong)
- [Screen Shot 2019-05-21 at 12.00.42 PM.png](https://vlab.ncep.noaa.gov/redmine/attachments/download/22810/Screen%20Shot%202019-05-21%20at%2012.00.42%20PM.png) (molly.b.smith) | priority | metexpress surface app does not plot curves correctly author name bonny strong bonnystrong original redmine issue original date original assignee molly b smith i tried plotting different curves from different met databases i m attaching screenshots where i plotted each curve separately then when i tried to plot both together they have data over different time intervals but when i did the plot together i changed the time interval to a range that included both but the plot changed the time to only that of the second curve then reported no data for the first curve also in producing this issue i ve noticed that there is nothing on the graph itself that identifies the database that was used i think we need to have this or the titles are not definitive bonny strong bonny strong bonny strong molly b smith | 1 |
73,000 | 3,398,554,695 | IssuesEvent | 2015-12-02 04:53:45 | R4stl1n/allianceauth | https://api.github.com/repos/R4stl1n/allianceauth | closed | Empty EveCorperationInfo query result breaks the HR app | bug High Priority | Probably need to put a try/catch in there.
http://i.imgur.com/Z6K5eSS.png | 1.0 | Empty EveCorperationInfo query result breaks the HR app - Probably need to put a try/catch in there.
http://i.imgur.com/Z6K5eSS.png | priority | empty evecorperationinfo query result breaks the hr app probably need to put a try catch in there | 1 |
237,779 | 7,764,155,657 | IssuesEvent | 2018-06-01 19:12:00 | fedora-infra/bodhi | https://api.github.com/repos/fedora-infra/bodhi | closed | Bugs should only be modified if they are Fedora/EPEL-only | High priority RFE bugzilla | Bodhi occasionally tries to update bugs it should not, such as this one:
https://bugzilla.redhat.com/show_bug.cgi?id=1383657#c6
This is a top-level CVE bug and its state should not be modified by Bodhi since it tracks the issue in products other than Fedora and EPEL. Bodhi could simply check that the component is not 'vulnerability' and product 'Security Response' and skip making any state changes if it is.
| 1.0 | Bugs should only be modified if they are Fedora/EPEL-only - Bodhi occasionally tries to update bugs it should not, such as this one:
https://bugzilla.redhat.com/show_bug.cgi?id=1383657#c6
This is a top-level CVE bug and its state should not be modified by Bodhi since it tracks the issue in products other than Fedora and EPEL. Bodhi could simply check that the component is not 'vulnerability' and product 'Security Response' and skip making any state changes if it is.
| priority | bugs should only be modified if they are fedora epel only bodhi occasionally tries to update bugs it should not such as this one this is a top level cve bug and its state should not be modified by bodhi since it tracks the issue in products other than fedora and epel bodhi could simply check that the component is not vulnerability and product security response and skip making any state changes if it is | 1 |
358,897 | 10,651,603,088 | IssuesEvent | 2019-10-17 10:45:06 | AY1920S1-CS2113T-F11-1/main | https://api.github.com/repos/AY1920S1-CS2113T-F11-1/main | closed | Enhance Reminders for Duke to fit requirements for SpongeBob | priority.High type.Task | Do things like adding reminding user of events and deadlines in the next one week (can modify this depending on your requirement) | 1.0 | Enhance Reminders for Duke to fit requirements for SpongeBob - Do things like adding reminding user of events and deadlines in the next one week (can modify this depending on your requirement) | priority | enhance reminders for duke to fit requirements for spongebob do things like adding reminding user of events and deadlines in the next one week can modify this depending on your requirement | 1 |
690,454 | 23,660,592,958 | IssuesEvent | 2022-08-26 15:12:25 | python/mypy | https://api.github.com/repos/python/mypy | closed | Regression on `master`: mypy unconditionally crashes when used with `--enable-error-code` for multiple error codes | crash priority-0-high | **Crash Report**
On the `master` branch, mypy unconditionally crashes if used with the command-line options `--enable-error-code ignore-without-code`. I've bisected the regression to:
- #13502
**Traceback**
```
Traceback (most recent call last):
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 196, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 86, in _run_code
exec(code, run_globals)
File "C:\Users\alexw\coding\mypy\venv\Scripts\mypy.exe\__main__.py", line 7, in <module>
File "C:\Users\alexw\coding\mypy\mypy\__main__.py", line 15, in console_entry
main()
File "C:\Users\alexw\coding\mypy\mypy\main.py", line 95, in main
res, messages, blockers = run_build(sources, options, fscache, t0, stdout, stderr)
File "C:\Users\alexw\coding\mypy\mypy\main.py", line 174, in run_build
res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 186, in build
result = _build(
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 269, in _build
graph = dispatch(sources, manager, stdout)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 2873, in dispatch
process_graph(graph, manager)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 3257, in process_graph
process_stale_scc(graph, scc, manager)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 3378, in process_stale_scc
graph[id].write_cache()
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 2447, in write_cache
new_interface_hash, self.meta = write_cache(
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 1639, in write_cache
meta_str = json_dumps(meta, manager.options.debug_cache)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 1510, in json_dumps
return json.dumps(obj, sort_keys=True)
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\__init__.py", line 238, in dumps
**kw).encode(obj)
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 199, in encode
chunks = self.iterencode(o, _one_shot=True)
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 257, in iterencode
return _iterencode(o, 0)
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 179, in default
raise TypeError(f'Object of type {o.__class__.__name__} '
TypeError: Object of type ErrorCode is not JSON serializable
```
**To Reproduce**
- `cd` into a local clone of mypy, activate a local environment with an editable install of mypy.
- Checkout the `master` branch.
- Run `mypy --enable-error-code ignore-without-code` on any repository. I've reproduced it using the selfcheck (`python -m mypy --config-file mypy_self_check.ini -p mypy --enable-error-code ignore-without-code`) and also on my local clone of `flake8-pyi`, which has the `ignore-without-code` error code enabled as part of its default mypy configuration. | 1.0 | Regression on `master`: mypy unconditionally crashes when used with `--enable-error-code` for multiple error codes - **Crash Report**
On the `master` branch, mypy unconditionally crashes if used with the command-line options `--enable-error-code ignore-without-code`. I've bisected the regression to:
- #13502
**Traceback**
```
Traceback (most recent call last):
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 196, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 86, in _run_code
exec(code, run_globals)
File "C:\Users\alexw\coding\mypy\venv\Scripts\mypy.exe\__main__.py", line 7, in <module>
File "C:\Users\alexw\coding\mypy\mypy\__main__.py", line 15, in console_entry
main()
File "C:\Users\alexw\coding\mypy\mypy\main.py", line 95, in main
res, messages, blockers = run_build(sources, options, fscache, t0, stdout, stderr)
File "C:\Users\alexw\coding\mypy\mypy\main.py", line 174, in run_build
res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 186, in build
result = _build(
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 269, in _build
graph = dispatch(sources, manager, stdout)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 2873, in dispatch
process_graph(graph, manager)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 3257, in process_graph
process_stale_scc(graph, scc, manager)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 3378, in process_stale_scc
graph[id].write_cache()
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 2447, in write_cache
new_interface_hash, self.meta = write_cache(
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 1639, in write_cache
meta_str = json_dumps(meta, manager.options.debug_cache)
File "C:\Users\alexw\coding\mypy\mypy\build.py", line 1510, in json_dumps
return json.dumps(obj, sort_keys=True)
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\__init__.py", line 238, in dumps
**kw).encode(obj)
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 199, in encode
chunks = self.iterencode(o, _one_shot=True)
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 257, in iterencode
return _iterencode(o, 0)
File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 179, in default
raise TypeError(f'Object of type {o.__class__.__name__} '
TypeError: Object of type ErrorCode is not JSON serializable
```
**To Reproduce**
- `cd` into a local clone of mypy, activate a local environment with an editable install of mypy.
- Checkout the `master` branch.
- Run `mypy --enable-error-code ignore-without-code` on any repository. I've reproduced it using the selfcheck (`python -m mypy --config-file mypy_self_check.ini -p mypy --enable-error-code ignore-without-code`) and also on my local clone of `flake8-pyi`, which has the `ignore-without-code` error code enabled as part of its default mypy configuration. | priority | regression on master mypy unconditionally crashes when used with enable error code for multiple error codes crash report on the master branch mypy unconditionally crashes if used with the command line options enable error code ignore without code i ve bisected the regression to traceback traceback most recent call last file c users alexw appdata local programs python lib runpy py line in run module as main return run code code main globals none file c users alexw appdata local programs python lib runpy py line in run code exec code run globals file c users alexw coding mypy venv scripts mypy exe main py line in file c users alexw coding mypy mypy main py line in console entry main file c users alexw coding mypy mypy main py line in main res messages blockers run build sources options fscache stdout stderr file c users alexw coding mypy mypy main py line in run build res build build sources options none flush errors fscache stdout stderr file c users alexw coding mypy mypy build py line in build result build file c users alexw coding mypy mypy build py line in build graph dispatch sources manager stdout file c users alexw coding mypy mypy build py line in dispatch process graph graph manager file c users alexw coding mypy mypy build py line in process graph process stale scc graph scc manager file c users alexw coding mypy mypy build py line in process stale scc graph write cache file c users alexw coding mypy mypy build py line in write cache new interface hash self meta write cache file c users alexw coding mypy mypy build py line in write cache meta str json dumps meta manager options debug cache file c users alexw coding mypy mypy build py line in json dumps return json dumps obj sort keys true file c users alexw appdata local programs python lib json init py line in dumps kw encode obj file c users alexw appdata local programs python lib json encoder py line in encode chunks self iterencode o one shot true file c users alexw appdata local programs python lib json encoder py line in iterencode return iterencode o file c users alexw appdata local programs python lib json encoder py line in default raise typeerror f object of type o class name typeerror object of type errorcode is not json serializable to reproduce cd into a local clone of mypy activate a local environment with an editable install of mypy checkout the master branch run mypy enable error code ignore without code on any repository i ve reproduced it using the selfcheck python m mypy config file mypy self check ini p mypy enable error code ignore without code and also on my local clone of pyi which has the ignore without code error code enabled as part of its default mypy configuration | 1 |
239,807 | 7,800,063,868 | IssuesEvent | 2018-06-09 04:14:23 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0006880:
it should be possible to save drafts/templates without subject/recipients | Felamimail Mantis high priority | **Reported by pschuele on 3 Aug 2012 20:08**
**Version:** Milan (2012.03.5)
it should be possible to save drafts/templates without subject/recipients
| 1.0 | 0006880:
it should be possible to save drafts/templates without subject/recipients - **Reported by pschuele on 3 Aug 2012 20:08**
**Version:** Milan (2012.03.5)
it should be possible to save drafts/templates without subject/recipients
| priority | it should be possible to save drafts templates without subject recipients reported by pschuele on aug version milan it should be possible to save drafts templates without subject recipients | 1 |
328,254 | 9,991,754,586 | IssuesEvent | 2019-07-11 11:55:43 | telstra/open-kilda | https://api.github.com/repos/telstra/open-kilda | closed | [Stats] As a user I can see table stats in OpenTSDB | area/stats feature priority/2-high | In [statsresponse](http://flowgrammable.org/sdn/openflow/message-layer/statsresponse/#ofp_1_3_0) Table message wee need to collect `active-count`, `lookup-count` and `matched-count` per table in a same cadense we do flow stats.
Also, we should record a diff: `lookup-count` minus `matched-count`. Matric name should be `missed`
metric details:
- name: `switch.table.*` without `-count` (e.g. `switch.table.active`, etc.)
- `switch.table.active`
- `switch.table.lookup`
- `switch.table.matched`
- `switch.table.missed`
- tags: switchid, tableid | 1.0 | [Stats] As a user I can see table stats in OpenTSDB - In [statsresponse](http://flowgrammable.org/sdn/openflow/message-layer/statsresponse/#ofp_1_3_0) Table message wee need to collect `active-count`, `lookup-count` and `matched-count` per table in a same cadense we do flow stats.
Also, we should record a diff: `lookup-count` minus `matched-count`. Matric name should be `missed`
metric details:
- name: `switch.table.*` without `-count` (e.g. `switch.table.active`, etc.)
- `switch.table.active`
- `switch.table.lookup`
- `switch.table.matched`
- `switch.table.missed`
- tags: switchid, tableid | priority | as a user i can see table stats in opentsdb in table message wee need to collect active count lookup count and matched count per table in a same cadense we do flow stats also we should record a diff lookup count minus matched count matric name should be missed metric details name switch table without count e g switch table active etc switch table active switch table lookup switch table matched switch table missed tags switchid tableid | 1 |
109,777 | 4,413,816,943 | IssuesEvent | 2016-08-13 02:32:14 | thalida/thalida.com | https://api.github.com/repos/thalida/thalida.com | closed | Selected highlighted text doesn't show selected state. | platform: all priority: high serverity: bug status: ready | <img width="503" alt="screen shot 2016-08-12 at 3 26 12 pm" src="https://cloud.githubusercontent.com/assets/3401715/17634539/22b0d216-60a1-11e6-8553-d9b535b2fd54.png">
| 1.0 | Selected highlighted text doesn't show selected state. - <img width="503" alt="screen shot 2016-08-12 at 3 26 12 pm" src="https://cloud.githubusercontent.com/assets/3401715/17634539/22b0d216-60a1-11e6-8553-d9b535b2fd54.png">
| priority | selected highlighted text doesn t show selected state img width alt screen shot at pm src | 1 |
61,994 | 3,164,000,989 | IssuesEvent | 2015-09-20 20:12:59 | audreyr/cookiecutter | https://api.github.com/repos/audreyr/cookiecutter | closed | Out of date Click dependency | high-priority | Click is up to version 5.1. Should cookiecutter support the newer release of Click? | 1.0 | Out of date Click dependency - Click is up to version 5.1. Should cookiecutter support the newer release of Click? | priority | out of date click dependency click is up to version should cookiecutter support the newer release of click | 1 |
341,471 | 10,297,895,466 | IssuesEvent | 2019-08-28 09:59:04 | UniversityOfHelsinkiCS/fuksilaiterekisteri | https://api.github.com/repos/UniversityOfHelsinkiCS/fuksilaiterekisteri | closed | prepare registration for oodi downtime exceptions | enhancement high priority | when oodi is down for maintenance, registration is 'down for maintenance'
next downtime is (presumably) on 1st of Sep | 1.0 | prepare registration for oodi downtime exceptions - when oodi is down for maintenance, registration is 'down for maintenance'
next downtime is (presumably) on 1st of Sep | priority | prepare registration for oodi downtime exceptions when oodi is down for maintenance registration is down for maintenance next downtime is presumably on of sep | 1 |
716,126 | 24,622,255,598 | IssuesEvent | 2022-10-16 03:41:26 | HaDuve/TravelCostNative | https://api.github.com/repos/HaDuve/TravelCostNative | closed | Fix the bug about the expense context time frame | Bug 1 - High Priority | It should be today, not last 24 hours.
Same with week month etc. | 1.0 | Fix the bug about the expense context time frame - It should be today, not last 24 hours.
Same with week month etc. | priority | fix the bug about the expense context time frame it should be today not last hours same with week month etc | 1 |
602,419 | 18,468,743,286 | IssuesEvent | 2021-10-17 11:11:30 | AY2122S1-CS2103T-F13-2/tp | https://api.github.com/repos/AY2122S1-CS2103T-F13-2/tp | opened | Modify Help Command | priority.High type.Story | As a nurse, I can view all commands available to me, so that I can better take advantage of the full functionality of the application. | 1.0 | Modify Help Command - As a nurse, I can view all commands available to me, so that I can better take advantage of the full functionality of the application. | priority | modify help command as a nurse i can view all commands available to me so that i can better take advantage of the full functionality of the application | 1 |
241,834 | 7,834,916,458 | IssuesEvent | 2018-06-16 20:19:10 | alakajam-team/alakajam | https://api.github.com/repos/alakajam-team/alakajam | closed | "Save comment & ratings" only saves the ratings if there is a comment | bug high priority | * `npm run watch`
* Go to http://localhost:8000/2nd-alakajam/11/game-9/
* Put 10 stars on Overall
* Click "Save comment & ratings"
Expected: ratings are saved and we see 10 stars.
Actual: ratings are not saved, we still see empty stars.
If a nonempty comment is supplied, both are saved correctly. | 1.0 | "Save comment & ratings" only saves the ratings if there is a comment - * `npm run watch`
* Go to http://localhost:8000/2nd-alakajam/11/game-9/
* Put 10 stars on Overall
* Click "Save comment & ratings"
Expected: ratings are saved and we see 10 stars.
Actual: ratings are not saved, we still see empty stars.
If a nonempty comment is supplied, both are saved correctly. | priority | save comment ratings only saves the ratings if there is a comment npm run watch go to put stars on overall click save comment ratings expected ratings are saved and we see stars actual ratings are not saved we still see empty stars if a nonempty comment is supplied both are saved correctly | 1 |
83,995 | 3,645,586,391 | IssuesEvent | 2016-02-15 15:16:08 | onyxfish/csvkit | https://api.github.com/repos/onyxfish/csvkit | closed | Unicode issues causing tests to fail in Python 2 | High Priority | I changed a passing test to read from a file containing a UTF-8 character, and now tests are failing. Not sure how to correct. See https://travis-ci.org/onyxfish/csvkit/jobs/108149327
/cc @onyxfish | 1.0 | Unicode issues causing tests to fail in Python 2 - I changed a passing test to read from a file containing a UTF-8 character, and now tests are failing. Not sure how to correct. See https://travis-ci.org/onyxfish/csvkit/jobs/108149327
/cc @onyxfish | priority | unicode issues causing tests to fail in python i changed a passing test to read from a file containing a utf character and now tests are failing not sure how to correct see cc onyxfish | 1 |
559,498 | 16,564,721,620 | IssuesEvent | 2021-05-29 06:45:17 | hatnote/montage | https://api.github.com/repos/hatnote/montage | closed | Unable to run Montage in Firefox | priority high type bug | On clicking the link 'Log in using Wikimedia account', Firefox throws the error message:
**Internal server error**
_<ExceptionInfo [AttributeError: 'JSONCookie' object has no attribute 'set_expires'] (27 frames, last=Callpoint('complete_login', 162, 'montage.public_endpoints', './montage/public_endpoints.py', 121, ' cookie.set_expires()'))>
Error type: http://docs.python.org/2/library/exceptions.html#exceptions.AttributeError_
Same result on several machines. Firefox 62.0.3 for Mac. | 1.0 | Unable to run Montage in Firefox - On clicking the link 'Log in using Wikimedia account', Firefox throws the error message:
**Internal server error**
_<ExceptionInfo [AttributeError: 'JSONCookie' object has no attribute 'set_expires'] (27 frames, last=Callpoint('complete_login', 162, 'montage.public_endpoints', './montage/public_endpoints.py', 121, ' cookie.set_expires()'))>
Error type: http://docs.python.org/2/library/exceptions.html#exceptions.AttributeError_
Same result on several machines. Firefox 62.0.3 for Mac. | priority | unable to run montage in firefox on clicking the link log in using wikimedia account firefox throws the error message internal server error error type same result on several machines firefox for mac | 1 |
109,080 | 4,369,800,252 | IssuesEvent | 2016-08-04 02:07:45 | jsdsa/jsdsa | https://api.github.com/repos/jsdsa/jsdsa | closed | List of sorting algorithms | Algorithms High priority | ### Simple Sorts
- [x] Insertion sort
- [x] Selection sort
### Efficient Sorts
- [x] Heap sort
- [x] Merge sort
- [x] Quick sort
### Bubble sort and variants
- [x] Bubble sort
- [x] Comb sort
- [x] Shell sort
### Distribution sort
- [x] Bucket sort
- [x] Counting sort
- [x] Radix sort
Optimizations, alternatives, variants to be added later. | 1.0 | List of sorting algorithms - ### Simple Sorts
- [x] Insertion sort
- [x] Selection sort
### Efficient Sorts
- [x] Heap sort
- [x] Merge sort
- [x] Quick sort
### Bubble sort and variants
- [x] Bubble sort
- [x] Comb sort
- [x] Shell sort
### Distribution sort
- [x] Bucket sort
- [x] Counting sort
- [x] Radix sort
Optimizations, alternatives, variants to be added later. | priority | list of sorting algorithms simple sorts insertion sort selection sort efficient sorts heap sort merge sort quick sort bubble sort and variants bubble sort comb sort shell sort distribution sort bucket sort counting sort radix sort optimizations alternatives variants to be added later | 1 |
335,735 | 10,165,696,369 | IssuesEvent | 2019-08-07 14:23:44 | OpenLiveWriter/OpenLiveWriter | https://api.github.com/repos/OpenLiveWriter/OpenLiveWriter | closed | CI builds continuously spawn new instances | bug high-priority | Builds from AppVeyor continuously spawn new instances of themselves. Renaming Update.exe and starting OpenLiveWriter.exe directly seems workaround the issue, however this is a considerable roadblock for many users that wish to try out dev builds.
OpenLiveWriter.ApplicationMain.LaunchAdditionalInstance seems to be called each time an extra instance is spawned. Below is a stack trace of one of these calls;
```
at OpenLiveWriter.ApplicationMain.LaunchAdditionalInstance(String[] args)
at OpenLiveWriter.ApplicationMain.LaunchAction(String[] args, Boolean isFirstInstance)
at OpenLiveWriter.CoreServices.SingleInstanceApplicationManager.LaunchActionThreadWithState.ThreadProc()
at System.Threading.ThreadHelper.ThreadStart_Context(Object state)
at System.Threading.ExecutionContext.RunInternal(ExecutionContext executionContext, ContextCallback callback, Object state, Boolean preserveSyncCtx)
at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state, Boolean preserveSyncCtx)
at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
at System.Threading.ThreadHelper.ThreadStart()
```
## Workarounds
Numerous workarounds have been outlined in issue #786 for specifically installing the #810 fix. Installing a recent dev build would involve applying these instructions with a more recent build artifact rather than the build for #810.
- http://kgbthatsme.blogspot.com/2019/05/open-live-writer-works-again-with.html
- http://www.thanislim.com/2019/03/how-to-fix-open-live-writer-error-400.html
In summary:
1. Rename `%localappdata%\OpenLiveWriter\Update.exe` to another file name.
2. Change all Open Live Writer shortcuts to point towards `%localappdata%\OpenLiveWriter\app-x.y.z\OpenLiveWriter.exe` where x.y.z is the installed app version.
Please keep in mind that these workarounds break auto-updating.
| 1.0 | CI builds continuously spawn new instances - Builds from AppVeyor continuously spawn new instances of themselves. Renaming Update.exe and starting OpenLiveWriter.exe directly seems workaround the issue, however this is a considerable roadblock for many users that wish to try out dev builds.
OpenLiveWriter.ApplicationMain.LaunchAdditionalInstance seems to be called each time an extra instance is spawned. Below is a stack trace of one of these calls;
```
at OpenLiveWriter.ApplicationMain.LaunchAdditionalInstance(String[] args)
at OpenLiveWriter.ApplicationMain.LaunchAction(String[] args, Boolean isFirstInstance)
at OpenLiveWriter.CoreServices.SingleInstanceApplicationManager.LaunchActionThreadWithState.ThreadProc()
at System.Threading.ThreadHelper.ThreadStart_Context(Object state)
at System.Threading.ExecutionContext.RunInternal(ExecutionContext executionContext, ContextCallback callback, Object state, Boolean preserveSyncCtx)
at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state, Boolean preserveSyncCtx)
at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
at System.Threading.ThreadHelper.ThreadStart()
```
## Workarounds
Numerous workarounds have been outlined in issue #786 for specifically installing the #810 fix. Installing a recent dev build would involve applying these instructions with a more recent build artifact rather than the build for #810.
- http://kgbthatsme.blogspot.com/2019/05/open-live-writer-works-again-with.html
- http://www.thanislim.com/2019/03/how-to-fix-open-live-writer-error-400.html
In summary:
1. Rename `%localappdata%\OpenLiveWriter\Update.exe` to another file name.
2. Change all Open Live Writer shortcuts to point towards `%localappdata%\OpenLiveWriter\app-x.y.z\OpenLiveWriter.exe` where x.y.z is the installed app version.
Please keep in mind that these workarounds break auto-updating.
| priority | ci builds continuously spawn new instances builds from appveyor continuously spawn new instances of themselves renaming update exe and starting openlivewriter exe directly seems workaround the issue however this is a considerable roadblock for many users that wish to try out dev builds openlivewriter applicationmain launchadditionalinstance seems to be called each time an extra instance is spawned below is a stack trace of one of these calls at openlivewriter applicationmain launchadditionalinstance string args at openlivewriter applicationmain launchaction string args boolean isfirstinstance at openlivewriter coreservices singleinstanceapplicationmanager launchactionthreadwithstate threadproc at system threading threadhelper threadstart context object state at system threading executioncontext runinternal executioncontext executioncontext contextcallback callback object state boolean preservesyncctx at system threading executioncontext run executioncontext executioncontext contextcallback callback object state boolean preservesyncctx at system threading executioncontext run executioncontext executioncontext contextcallback callback object state at system threading threadhelper threadstart workarounds numerous workarounds have been outlined in issue for specifically installing the fix installing a recent dev build would involve applying these instructions with a more recent build artifact rather than the build for in summary rename localappdata openlivewriter update exe to another file name change all open live writer shortcuts to point towards localappdata openlivewriter app x y z openlivewriter exe where x y z is the installed app version please keep in mind that these workarounds break auto updating | 1 |
397,338 | 11,727,101,959 | IssuesEvent | 2020-03-10 15:27:43 | AY1920S2-CS2103T-W16-2/main | https://api.github.com/repos/AY1920S2-CS2103T-W16-2/main | opened | As a exerciser I want to see a history of all my past exercises | core priority.High type.Epic | so I can see how regularly I have been exercising | 1.0 | As a exerciser I want to see a history of all my past exercises - so I can see how regularly I have been exercising | priority | as a exerciser i want to see a history of all my past exercises so i can see how regularly i have been exercising | 1 |
375,200 | 11,101,396,505 | IssuesEvent | 2019-12-16 21:20:53 | carbon-design-system/design-language-website | https://api.github.com/repos/carbon-design-system/design-language-website | closed | Color picker down. The tabs do not change color choices | Severity 1 🚨 priority: high type: bug 🐛 | the screen goes black when selecting PMS, RGB or CMYK.
https://www.ibm.com/design/language/color/#specifications | 1.0 | Color picker down. The tabs do not change color choices - the screen goes black when selecting PMS, RGB or CMYK.
https://www.ibm.com/design/language/color/#specifications | priority | color picker down the tabs do not change color choices the screen goes black when selecting pms rgb or cmyk | 1 |
97,509 | 3,994,729,999 | IssuesEvent | 2016-05-10 13:27:46 | GoogleCloudPlatform/gcloud-eclipse-tools | https://api.github.com/repos/GoogleCloudPlatform/gcloud-eclipse-tools | opened | call setErrorMessage when project creation fails in StandardProjectWizard | enhancement high priority | find
// todo if fail, call setErrorMessage()
note that this happens a lot while testing new code, adn we need it fo that if nothing else. | 1.0 | call setErrorMessage when project creation fails in StandardProjectWizard - find
// todo if fail, call setErrorMessage()
note that this happens a lot while testing new code, adn we need it fo that if nothing else. | priority | call seterrormessage when project creation fails in standardprojectwizard find todo if fail call seterrormessage note that this happens a lot while testing new code adn we need it fo that if nothing else | 1 |
505,795 | 14,645,854,324 | IssuesEvent | 2020-12-26 10:34:14 | GrottoCenter/Grottocenter3 | https://api.github.com/repos/GrottoCenter/Grottocenter3 | closed | Magazine | Priority: High Type: Bug | Vincent a créé un type Magazine alors que dans le modèle nous avons Collection qui correspond au type prévu dans DCMI.
Il faudrait affecter l'id de Collection à tout ce qui est déclaré comme Magazine.
C'est relativement urgent car les utilisateurs sont en train de recréer les Collections. | 1.0 | Magazine - Vincent a créé un type Magazine alors que dans le modèle nous avons Collection qui correspond au type prévu dans DCMI.
Il faudrait affecter l'id de Collection à tout ce qui est déclaré comme Magazine.
C'est relativement urgent car les utilisateurs sont en train de recréer les Collections. | priority | magazine vincent a créé un type magazine alors que dans le modèle nous avons collection qui correspond au type prévu dans dcmi il faudrait affecter l id de collection à tout ce qui est déclaré comme magazine c est relativement urgent car les utilisateurs sont en train de recréer les collections | 1 |
525,130 | 15,238,489,435 | IssuesEvent | 2021-02-19 02:03:08 | TerriaJS/nationalmap | https://api.github.com/repos/TerriaJS/nationalmap | closed | Add ~globalDisclaimer~ welcome message to v8 | High priority | This is about welcome message - not globalDisclaimer
~This is what we have for DEA, so follow the format and change the message:~
```json
"globalDisclaimer": {
"buttonTitle": "I agree",
"confirmationRequired": true,
"devHostRegex": "\\b(staging|preview|test|dev)\\.",
"enableOnLocalhost": true,
"message": "The information displayed on the DEA Maps (the \"Service\") is for general informational purposes only, and is not intended to provide any commercial, financial, or legal advice.\n\nThe Service provides access to a range of data sets and tools, some of which are in development (beta). All data and tools are provided on an \"as is\" and \"with all faults\" basis without any warranty whatsoever. Geoscience Australia and Data61 do not warrant that these data sets and tools shall meet any requirements or expectations, or that the will be fit for any intended purposes.\n\nGeoscience Australia and Data61 assumes no responsibility for errors or omissions in the contents of the Service and reserves the right to make additions, deletions, or modification to the contents on the Service at any time without prior notice.\n\nGeoscience Australia and Data61 does not guarantee the accuracy, relevance, timeliness, or completeness of any information or data available through the Service or on linked external websites.\n\n[See full terms and conditions here.](./about#terms-and-conditions)",
"prodHostRegex": "prod\\.saas\\.terria\\.io$",
"title": "Disclaimer"
},
``` | 1.0 | Add ~globalDisclaimer~ welcome message to v8 - This is about welcome message - not globalDisclaimer
~This is what we have for DEA, so follow the format and change the message:~
```json
"globalDisclaimer": {
"buttonTitle": "I agree",
"confirmationRequired": true,
"devHostRegex": "\\b(staging|preview|test|dev)\\.",
"enableOnLocalhost": true,
"message": "The information displayed on the DEA Maps (the \"Service\") is for general informational purposes only, and is not intended to provide any commercial, financial, or legal advice.\n\nThe Service provides access to a range of data sets and tools, some of which are in development (beta). All data and tools are provided on an \"as is\" and \"with all faults\" basis without any warranty whatsoever. Geoscience Australia and Data61 do not warrant that these data sets and tools shall meet any requirements or expectations, or that the will be fit for any intended purposes.\n\nGeoscience Australia and Data61 assumes no responsibility for errors or omissions in the contents of the Service and reserves the right to make additions, deletions, or modification to the contents on the Service at any time without prior notice.\n\nGeoscience Australia and Data61 does not guarantee the accuracy, relevance, timeliness, or completeness of any information or data available through the Service or on linked external websites.\n\n[See full terms and conditions here.](./about#terms-and-conditions)",
"prodHostRegex": "prod\\.saas\\.terria\\.io$",
"title": "Disclaimer"
},
``` | priority | add globaldisclaimer welcome message to this is about welcome message not globaldisclaimer this is what we have for dea so follow the format and change the message json globaldisclaimer buttontitle i agree confirmationrequired true devhostregex b staging preview test dev enableonlocalhost true message the information displayed on the dea maps the service is for general informational purposes only and is not intended to provide any commercial financial or legal advice n nthe service provides access to a range of data sets and tools some of which are in development beta all data and tools are provided on an as is and with all faults basis without any warranty whatsoever geoscience australia and do not warrant that these data sets and tools shall meet any requirements or expectations or that the will be fit for any intended purposes n ngeoscience australia and assumes no responsibility for errors or omissions in the contents of the service and reserves the right to make additions deletions or modification to the contents on the service at any time without prior notice n ngeoscience australia and does not guarantee the accuracy relevance timeliness or completeness of any information or data available through the service or on linked external websites n n about terms and conditions prodhostregex prod saas terria io title disclaimer | 1 |
339,397 | 10,253,739,983 | IssuesEvent | 2019-08-21 12:03:30 | DimensionDev/Maskbook | https://api.github.com/repos/DimensionDev/Maskbook | opened | ux: new url permission scheme does not work with profile import | Priority: P2 (Most users) Severity: High Type: Bug Type: UI | importing an old profile does not trigger the webpage access permission request. | 1.0 | ux: new url permission scheme does not work with profile import - importing an old profile does not trigger the webpage access permission request. | priority | ux new url permission scheme does not work with profile import importing an old profile does not trigger the webpage access permission request | 1 |
377,488 | 11,171,737,926 | IssuesEvent | 2019-12-28 22:28:37 | Thorium-Sim/thorium | https://api.github.com/repos/Thorium-Sim/thorium | opened | Dilithium Stress | priority/high type/bug | ### Requested By: Your Friends from the Odyssey
### Priority: High
### Version: 2.1.0
The dilithium stress still doesn't match between core and the operations station. If you adjust it on our end it doesn't change it on theirs and vice versa. It also looks super whack and jank. (Refer to image)
### Steps to Reproduce
Put a dilithium screen in your configuration. Change the stress from your side. Observe as it does nothing (whether the flight is paused or not) | 1.0 | Dilithium Stress - ### Requested By: Your Friends from the Odyssey
### Priority: High
### Version: 2.1.0
The dilithium stress still doesn't match between core and the operations station. If you adjust it on our end it doesn't change it on theirs and vice versa. It also looks super whack and jank. (Refer to image)
### Steps to Reproduce
Put a dilithium screen in your configuration. Change the stress from your side. Observe as it does nothing (whether the flight is paused or not) | priority | dilithium stress requested by your friends from the odyssey priority high version the dilithium stress still doesn t match between core and the operations station if you adjust it on our end it doesn t change it on theirs and vice versa it also looks super whack and jank refer to image steps to reproduce put a dilithium screen in your configuration change the stress from your side observe as it does nothing whether the flight is paused or not | 1 |
472,561 | 13,627,009,296 | IssuesEvent | 2020-09-24 11:57:33 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio] Restoring a backup (with the users not backed up) fails | bug priority: high | ## Describe the bug
Restoring a backup (from a version with the users not backed up) fails
## To Reproduce
Steps to reproduce the behavior:
1. Create a backup using an earlier version of Crafter. For our example, we'll use 3.1.8
2. Using the 3.1.10 snapshot version of Crafter, restore the backup created using 3.1.8
3. Notice the error message `./crafter.sh: line 1126: /Users/vita/temp/test3/craftercms/crafter-authoring/mybackups/temp/users.sql: No such file or directory`
## Expected behavior
The restore should work
## Screenshots
{{If applicable, add screenshots to help explain your problem.}}
## Logs
```
2020-09-22 10:42:36.717 INFO 78765 --- [ main] ch.vorburger.mariadb4j.DB : Database startup complete.
2020-09-22 10:42:37.032 INFO 78765 --- [ main] c.v.m.s.boot.MariaDB4jApplication : Started MariaDB4jApplication in 4.989 seconds (JVM running for 5.485)
------------------------------------------------------------------------
Restoring embedded DB
------------------------------------------------------------------------
./crafter.sh: line 1126: /Users/vita/temp/test3/craftercms/crafter-authoring/mybackups/temp/users.sql: No such file or directory
Unable to continue, an error occurred or the script was forcefully stopped
```
## Specs
### Version
3.1.10 snapshot
### OS
OS X
### Browser
{{What browser did you use to produce the bug.}}
## Additional context
{{Add any other context about the problem here.}}
| 1.0 | [studio] Restoring a backup (with the users not backed up) fails - ## Describe the bug
Restoring a backup (from a version with the users not backed up) fails
## To Reproduce
Steps to reproduce the behavior:
1. Create a backup using an earlier version of Crafter. For our example, we'll use 3.1.8
2. Using the 3.1.10 snapshot version of Crafter, restore the backup created using 3.1.8
3. Notice the error message `./crafter.sh: line 1126: /Users/vita/temp/test3/craftercms/crafter-authoring/mybackups/temp/users.sql: No such file or directory`
## Expected behavior
The restore should work
## Screenshots
{{If applicable, add screenshots to help explain your problem.}}
## Logs
```
2020-09-22 10:42:36.717 INFO 78765 --- [ main] ch.vorburger.mariadb4j.DB : Database startup complete.
2020-09-22 10:42:37.032 INFO 78765 --- [ main] c.v.m.s.boot.MariaDB4jApplication : Started MariaDB4jApplication in 4.989 seconds (JVM running for 5.485)
------------------------------------------------------------------------
Restoring embedded DB
------------------------------------------------------------------------
./crafter.sh: line 1126: /Users/vita/temp/test3/craftercms/crafter-authoring/mybackups/temp/users.sql: No such file or directory
Unable to continue, an error occurred or the script was forcefully stopped
```
## Specs
### Version
3.1.10 snapshot
### OS
OS X
### Browser
{{What browser did you use to produce the bug.}}
## Additional context
{{Add any other context about the problem here.}}
| priority | restoring a backup with the users not backed up fails describe the bug restoring a backup from a version with the users not backed up fails to reproduce steps to reproduce the behavior create a backup using an earlier version of crafter for our example we ll use using the snapshot version of crafter restore the backup created using notice the error message crafter sh line users vita temp craftercms crafter authoring mybackups temp users sql no such file or directory expected behavior the restore should work screenshots if applicable add screenshots to help explain your problem logs info ch vorburger db database startup complete info c v m s boot started in seconds jvm running for restoring embedded db crafter sh line users vita temp craftercms crafter authoring mybackups temp users sql no such file or directory unable to continue an error occurred or the script was forcefully stopped specs version snapshot os os x browser what browser did you use to produce the bug additional context add any other context about the problem here | 1 |
194,465 | 6,895,025,129 | IssuesEvent | 2017-11-23 12:17:24 | ballerinalang/composer | https://api.github.com/repos/ballerinalang/composer | opened | Default config is wrong for ftp service | 0.95.1 Priority/Highest Severity/Critical | Please see the below image. The default config is wrong.
Current:
```
import ballerina.net.ftp;
service<fs> service1 {
resource echo1 (fs:FileSystemEvent m) {
}
}
```
Expected:
```
import ballerina.net.ftp;
service<ftp> service1 {
resource echo1 (ftp:FTPServerEvent m) {
}
}
```

| 1.0 | Default config is wrong for ftp service - Please see the below image. The default config is wrong.
Current:
```
import ballerina.net.ftp;
service<fs> service1 {
resource echo1 (fs:FileSystemEvent m) {
}
}
```
Expected:
```
import ballerina.net.ftp;
service<ftp> service1 {
resource echo1 (ftp:FTPServerEvent m) {
}
}
```

| priority | default config is wrong for ftp service please see the below image the default config is wrong current import ballerina net ftp service resource fs filesystemevent m expected import ballerina net ftp service resource ftp ftpserverevent m | 1 |
489,012 | 14,100,190,562 | IssuesEvent | 2020-11-06 03:27:34 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Unable to list Roles via console with Oracle19c | Affected/5.11.0-Beta Priority/Highest Severity/Blocker bug identity-core | **Describe the issue:**
Role list will be empty with Oracle19c, in Console > Manage > Roles
**How to reproduce:**
1. Configure Oracle as Primary user store
2. Login to console with superuser credentials
3. View roles (Manage)
**Expected behavior:**
Added roles should be display properly
**Environment information**
- Product Version: 5.11 Beta 3
- OS: Windows
- Database: Oracle 19c
- Userstore: JDBC
---


| 1.0 | Unable to list Roles via console with Oracle19c - **Describe the issue:**
Role list will be empty with Oracle19c, in Console > Manage > Roles
**How to reproduce:**
1. Configure Oracle as Primary user store
2. Login to console with superuser credentials
3. View roles (Manage)
**Expected behavior:**
Added roles should be display properly
**Environment information**
- Product Version: 5.11 Beta 3
- OS: Windows
- Database: Oracle 19c
- Userstore: JDBC
---


| priority | unable to list roles via console with describe the issue role list will be empty with in console manage roles how to reproduce configure oracle as primary user store login to console with superuser credentials view roles manage expected behavior added roles should be display properly environment information product version beta os windows database oracle userstore jdbc | 1 |
603,974 | 18,674,989,732 | IssuesEvent | 2021-10-31 12:05:18 | AY2122S1-CS2103-T14-2/tp | https://api.github.com/repos/AY2122S1-CS2103-T14-2/tp | closed | [PE-D] Snoozed reminders re-activate after closing and re-opening app. | priority.High type.Task | Steps to re-produce:
1. Add a reminder for current day
2. Snooze reminder
3. Type exit command
4. Re-open ePoch
5. Find that occurrences has been reduced, reminder not snoozed, and CCA label is not shown properly.
Before re-opening:

After re-opening:

<!--session: 1635494417648-992e5fad-48d0-4008-813a-369898477680-->
<!--Version: Web v3.4.1-->
-------------
Labels: `type.FunctionalityBug` `severity.High`
original: juliussneezer04/ped#9 | 1.0 | [PE-D] Snoozed reminders re-activate after closing and re-opening app. - Steps to re-produce:
1. Add a reminder for current day
2. Snooze reminder
3. Type exit command
4. Re-open ePoch
5. Find that occurrences has been reduced, reminder not snoozed, and CCA label is not shown properly.
Before re-opening:

After re-opening:

<!--session: 1635494417648-992e5fad-48d0-4008-813a-369898477680-->
<!--Version: Web v3.4.1-->
-------------
Labels: `type.FunctionalityBug` `severity.High`
original: juliussneezer04/ped#9 | priority | snoozed reminders re activate after closing and re opening app steps to re produce add a reminder for current day snooze reminder type exit command re open epoch find that occurrences has been reduced reminder not snoozed and cca label is not shown properly before re opening after re opening labels type functionalitybug severity high original ped | 1 |
158,074 | 6,021,371,865 | IssuesEvent | 2017-06-07 18:32:36 | apollographql/apollo-android | https://api.github.com/repos/apollographql/apollo-android | closed | Trying to assign a Double value into Integer field. | Priority: High Type: Bug Type: compiler | Hey,
I don't know if by my error, or by some kind of bug in the compiler an erroneus code is being generated.
Following is one of the errors I'm getting:
`Error:(42, 39) error: incompatible types: double cannot be converted to Integer`
Generated file looks like this:
```
package type;
import java.lang.Boolean;
import java.lang.Integer;
import java.lang.String;
import javax.annotation.Generated;
import javax.annotation.Nullable;
@Generated("Apollo GraphQL")
public final class NotificationFilters {
private final @Nullable Boolean isNew;
private final @Nullable Integer limit;
private final @Nullable String cursor;
NotificationFilters(@Nullable Boolean isNew, @Nullable Integer limit, @Nullable String cursor) {
this.isNew = isNew;
this.limit = limit;
this.cursor = cursor;
}
public @Nullable Boolean isNew() {
return this.isNew;
}
public @Nullable Integer limit() {
return this.limit;
}
public @Nullable String cursor() {
return this.cursor;
}
public static Builder builder() {
return new Builder();
}
public static final class Builder {
private @Nullable Boolean isNew;
private @Nullable Integer limit = 10.0;
private @Nullable String cursor;
Builder() {
}
public Builder isNew(@Nullable Boolean isNew) {
this.isNew = isNew;
return this;
}
public Builder limit(@Nullable Integer limit) {
this.limit = limit;
return this;
}
public Builder cursor(@Nullable String cursor) {
this.cursor = cursor;
return this;
}
public NotificationFilters build() {
return new NotificationFilters(isNew, limit, cursor);
}
}
}
```
When you look at the generated Builder, you can indeed see that a double value is being assigned to an Integer. These errors happen for all models that have a default value set.
Following is part of the schema file:
```
{
"kind": "INPUT_OBJECT",
"name": "NotificationFilters",
"description": "",
"fields": null,
"inputFields": [
{
"name": "isNew",
"description": "",
"type": {
"kind": "SCALAR",
"name": "Boolean",
"ofType": null
},
"defaultValue": null
},
{
"name": "limit",
"description": "",
"type": {
"kind": "SCALAR",
"name": "Int",
"ofType": null
},
"defaultValue": "10"
},
{
"name": "cursor",
"description": "",
"type": {
"kind": "SCALAR",
"name": "String",
"ofType": null
},
"defaultValue": null
}
],
"interfaces": null,
"enumValues": null,
"possibleTypes": null
}
```
Any ideas or tips on what might be going wrong? I'm using the 0.3.1-SNAPSHOT.
| 1.0 | Trying to assign a Double value into Integer field. - Hey,
I don't know if by my error, or by some kind of bug in the compiler an erroneus code is being generated.
Following is one of the errors I'm getting:
`Error:(42, 39) error: incompatible types: double cannot be converted to Integer`
Generated file looks like this:
```
package type;
import java.lang.Boolean;
import java.lang.Integer;
import java.lang.String;
import javax.annotation.Generated;
import javax.annotation.Nullable;
@Generated("Apollo GraphQL")
public final class NotificationFilters {
private final @Nullable Boolean isNew;
private final @Nullable Integer limit;
private final @Nullable String cursor;
NotificationFilters(@Nullable Boolean isNew, @Nullable Integer limit, @Nullable String cursor) {
this.isNew = isNew;
this.limit = limit;
this.cursor = cursor;
}
public @Nullable Boolean isNew() {
return this.isNew;
}
public @Nullable Integer limit() {
return this.limit;
}
public @Nullable String cursor() {
return this.cursor;
}
public static Builder builder() {
return new Builder();
}
public static final class Builder {
private @Nullable Boolean isNew;
private @Nullable Integer limit = 10.0;
private @Nullable String cursor;
Builder() {
}
public Builder isNew(@Nullable Boolean isNew) {
this.isNew = isNew;
return this;
}
public Builder limit(@Nullable Integer limit) {
this.limit = limit;
return this;
}
public Builder cursor(@Nullable String cursor) {
this.cursor = cursor;
return this;
}
public NotificationFilters build() {
return new NotificationFilters(isNew, limit, cursor);
}
}
}
```
When you look at the generated Builder, you can indeed see that a double value is being assigned to an Integer. These errors happen for all models that have a default value set.
Following is part of the schema file:
```
{
"kind": "INPUT_OBJECT",
"name": "NotificationFilters",
"description": "",
"fields": null,
"inputFields": [
{
"name": "isNew",
"description": "",
"type": {
"kind": "SCALAR",
"name": "Boolean",
"ofType": null
},
"defaultValue": null
},
{
"name": "limit",
"description": "",
"type": {
"kind": "SCALAR",
"name": "Int",
"ofType": null
},
"defaultValue": "10"
},
{
"name": "cursor",
"description": "",
"type": {
"kind": "SCALAR",
"name": "String",
"ofType": null
},
"defaultValue": null
}
],
"interfaces": null,
"enumValues": null,
"possibleTypes": null
}
```
Any ideas or tips on what might be going wrong? I'm using the 0.3.1-SNAPSHOT.
| priority | trying to assign a double value into integer field hey i don t know if by my error or by some kind of bug in the compiler an erroneus code is being generated following is one of the errors i m getting error error incompatible types double cannot be converted to integer generated file looks like this package type import java lang boolean import java lang integer import java lang string import javax annotation generated import javax annotation nullable generated apollo graphql public final class notificationfilters private final nullable boolean isnew private final nullable integer limit private final nullable string cursor notificationfilters nullable boolean isnew nullable integer limit nullable string cursor this isnew isnew this limit limit this cursor cursor public nullable boolean isnew return this isnew public nullable integer limit return this limit public nullable string cursor return this cursor public static builder builder return new builder public static final class builder private nullable boolean isnew private nullable integer limit private nullable string cursor builder public builder isnew nullable boolean isnew this isnew isnew return this public builder limit nullable integer limit this limit limit return this public builder cursor nullable string cursor this cursor cursor return this public notificationfilters build return new notificationfilters isnew limit cursor when you look at the generated builder you can indeed see that a double value is being assigned to an integer these errors happen for all models that have a default value set following is part of the schema file kind input object name notificationfilters description fields null inputfields name isnew description type kind scalar name boolean oftype null defaultvalue null name limit description type kind scalar name int oftype null defaultvalue name cursor description type kind scalar name string oftype null defaultvalue null interfaces null enumvalues null possibletypes null any ideas or tips on what might be going wrong i m using the snapshot | 1 |
593,645 | 18,013,058,463 | IssuesEvent | 2021-09-16 10:52:13 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Prioritise methods when showing completions for module | Type/Improvement Priority/High Team/LanguageServer Points/1 Area/Completion | **Description:**
Check the below figure. It would be great if we optimise the completions to show the methods before the variables.
<img width="820" alt="Screenshot 2021-08-30 at 13 41 40" src="https://user-images.githubusercontent.com/7523274/131852202-6ee30149-5383-44e2-b683-e98efba80f33.png">
| 1.0 | Prioritise methods when showing completions for module - **Description:**
Check the below figure. It would be great if we optimise the completions to show the methods before the variables.
<img width="820" alt="Screenshot 2021-08-30 at 13 41 40" src="https://user-images.githubusercontent.com/7523274/131852202-6ee30149-5383-44e2-b683-e98efba80f33.png">
| priority | prioritise methods when showing completions for module description check the below figure it would be great if we optimise the completions to show the methods before the variables img width alt screenshot at src | 1 |
578,782 | 17,154,348,875 | IssuesEvent | 2021-07-14 03:38:03 | nerPatricia/imoveis-frontend | https://api.github.com/repos/nerPatricia/imoveis-frontend | closed | Update: Fazer o html do Editar Imoveis adicionar dados no form | high priority | utilizar formControlName, [formGroup], essas coisas de form reativo
<br>

| 1.0 | Update: Fazer o html do Editar Imoveis adicionar dados no form - utilizar formControlName, [formGroup], essas coisas de form reativo
<br>

| priority | update fazer o html do editar imoveis adicionar dados no form utilizar formcontrolname essas coisas de form reativo | 1 |
598,464 | 18,245,829,067 | IssuesEvent | 2021-10-01 18:15:11 | airbytehq/airbyte | https://api.github.com/repos/airbytehq/airbyte | closed | default displaySetupWizard to true for cloud | type/enhancement priority/high cloud-public-launch | @Jamakase noticed that this value wasn't being set properly.
Currently in OSS we set it to be `true` for the first workspace created/seeded on server startup and false otherwise. For cloud we need to be able to select this option via the api call. This requires adding an optional `displaySetupWizard` to `WorkspaceCreate` in the OSS API and setting that value in the call made form the Cloud API. | 1.0 | default displaySetupWizard to true for cloud - @Jamakase noticed that this value wasn't being set properly.
Currently in OSS we set it to be `true` for the first workspace created/seeded on server startup and false otherwise. For cloud we need to be able to select this option via the api call. This requires adding an optional `displaySetupWizard` to `WorkspaceCreate` in the OSS API and setting that value in the call made form the Cloud API. | priority | default displaysetupwizard to true for cloud jamakase noticed that this value wasn t being set properly currently in oss we set it to be true for the first workspace created seeded on server startup and false otherwise for cloud we need to be able to select this option via the api call this requires adding an optional displaysetupwizard to workspacecreate in the oss api and setting that value in the call made form the cloud api | 1 |
612,887 | 19,058,461,316 | IssuesEvent | 2021-11-26 02:02:01 | tracer-protocol/perpetual-pools-contracts | https://api.github.com/repos/tracer-protocol/perpetual-pools-contracts | closed | Minting and burning fee | enhancement priority: high review started | **Overview**
- Within V2, there will need to be a fee charged both the action of minting and burning. In other words, users will be charged a fee at the time of minting tokens, and also at the time of burning tokens.
- The fee will be a percentage. e.g. 1%. It should be a configurable parameter. The fee should be able to be changed by the contract owner after the time of deployment.
- The fee rate for minting will not necessarily be the same as the fee rate for burning.
- The proceeds of the fee will be added to the pool at which the user was minting into (or burning from).
---
**Minting Fee Logic**
- Hypothetically, let's say the mint fee rate is 1%.
- A user, Bob, commits to minting 100 USDC.
- At the time of committing to mint, 1 USDC will be taken away from the user, Bob. It is 1 USDC because it's 100 USDC * 1%.
- Bob's commit is now only 99 USDC. This will proceed as normal, and will mint tokens at the fair price.
Dev notes:
- In the above example, The user commits 100 Usdc, all 100 are transferred to the pool, but the amount of tokens stored as part of the commit would be 99, so when they go to claim, they will only get that 99 usdc worth of tokens.
---
**Burning Fee Logic**
- Hypothetically, let's say the burn fee rate is 2%.
- A user, Bob, has commited to burn his 200 Long Pool tokens.
- The 200 long pool tokens will sit in the shadow pool until the wait time is up.
- At the rebalancing event in which the commit will be executed, the SC should calculate the value of the 200 long pool tokens (e.g. if price is $2, then the overall value is $400).
- From this, the Smart contract takes $8 ($400 * 2%) from the user, and adds it back into the pool from which it was taken. All the users pool tokens are destroyed.
- 98% of the value is sent to the user.
---
### Extra notes
- Burning fees will require storing the fee rate at in a given update interval (since fee rate can change at any time) and using that to calculate how much quote token a user is entitled to during `updateAggregateBalance`
---
### Completion criteria (added by devs)
Where `n` is the fee rate as a percentage (`n == 1` means fee is 1%)
- When a user mints, `100 - n%` of the settlement tokens they minted with is stored in the user's commitment data.
- When a user mints then claims after update interval, the pool should still be in ownership of `n%` of the original commitment amount of settlement tokens.
- When a user mints, `n%` of the settlement tokens they minted with should be added to the balance of the side to which they are minting.
- e.g. if they mint $100 to the LONG side, with a 1% fee, directly after minting, $1 should be added to `LeveragedPool::longBalance`
- When a user burns, `100 - n%` of the settlement tokens they would normally be entitled to are in the user's aggregate balance after the update interval passes.
- When a user burns, upkeep happens, then they update their balance (usually through claiming), n% of the settlement tokens they gained should be kept in the pool and that amount added to the correct side of the pool (either `longBalance` or `shortBalance`)
- The same as above, but with `100 - n%` of the settlement tokens deposited to the user's wallet after claiming.
- If `n == 0`, then the user should get the exact amount they would be entitled to before the changes of this PR were implemented | 1.0 | Minting and burning fee - **Overview**
- Within V2, there will need to be a fee charged both the action of minting and burning. In other words, users will be charged a fee at the time of minting tokens, and also at the time of burning tokens.
- The fee will be a percentage. e.g. 1%. It should be a configurable parameter. The fee should be able to be changed by the contract owner after the time of deployment.
- The fee rate for minting will not necessarily be the same as the fee rate for burning.
- The proceeds of the fee will be added to the pool at which the user was minting into (or burning from).
---
**Minting Fee Logic**
- Hypothetically, let's say the mint fee rate is 1%.
- A user, Bob, commits to minting 100 USDC.
- At the time of committing to mint, 1 USDC will be taken away from the user, Bob. It is 1 USDC because it's 100 USDC * 1%.
- Bob's commit is now only 99 USDC. This will proceed as normal, and will mint tokens at the fair price.
Dev notes:
- In the above example, The user commits 100 Usdc, all 100 are transferred to the pool, but the amount of tokens stored as part of the commit would be 99, so when they go to claim, they will only get that 99 usdc worth of tokens.
---
**Burning Fee Logic**
- Hypothetically, let's say the burn fee rate is 2%.
- A user, Bob, has commited to burn his 200 Long Pool tokens.
- The 200 long pool tokens will sit in the shadow pool until the wait time is up.
- At the rebalancing event in which the commit will be executed, the SC should calculate the value of the 200 long pool tokens (e.g. if price is $2, then the overall value is $400).
- From this, the Smart contract takes $8 ($400 * 2%) from the user, and adds it back into the pool from which it was taken. All the users pool tokens are destroyed.
- 98% of the value is sent to the user.
---
### Extra notes
- Burning fees will require storing the fee rate at in a given update interval (since fee rate can change at any time) and using that to calculate how much quote token a user is entitled to during `updateAggregateBalance`
---
### Completion criteria (added by devs)
Where `n` is the fee rate as a percentage (`n == 1` means fee is 1%)
- When a user mints, `100 - n%` of the settlement tokens they minted with is stored in the user's commitment data.
- When a user mints then claims after update interval, the pool should still be in ownership of `n%` of the original commitment amount of settlement tokens.
- When a user mints, `n%` of the settlement tokens they minted with should be added to the balance of the side to which they are minting.
- e.g. if they mint $100 to the LONG side, with a 1% fee, directly after minting, $1 should be added to `LeveragedPool::longBalance`
- When a user burns, `100 - n%` of the settlement tokens they would normally be entitled to are in the user's aggregate balance after the update interval passes.
- When a user burns, upkeep happens, then they update their balance (usually through claiming), n% of the settlement tokens they gained should be kept in the pool and that amount added to the correct side of the pool (either `longBalance` or `shortBalance`)
- The same as above, but with `100 - n%` of the settlement tokens deposited to the user's wallet after claiming.
- If `n == 0`, then the user should get the exact amount they would be entitled to before the changes of this PR were implemented | priority | minting and burning fee overview within there will need to be a fee charged both the action of minting and burning in other words users will be charged a fee at the time of minting tokens and also at the time of burning tokens the fee will be a percentage e g it should be a configurable parameter the fee should be able to be changed by the contract owner after the time of deployment the fee rate for minting will not necessarily be the same as the fee rate for burning the proceeds of the fee will be added to the pool at which the user was minting into or burning from minting fee logic hypothetically let s say the mint fee rate is a user bob commits to minting usdc at the time of committing to mint usdc will be taken away from the user bob it is usdc because it s usdc bob s commit is now only usdc this will proceed as normal and will mint tokens at the fair price dev notes in the above example the user commits usdc all are transferred to the pool but the amount of tokens stored as part of the commit would be so when they go to claim they will only get that usdc worth of tokens burning fee logic hypothetically let s say the burn fee rate is a user bob has commited to burn his long pool tokens the long pool tokens will sit in the shadow pool until the wait time is up at the rebalancing event in which the commit will be executed the sc should calculate the value of the long pool tokens e g if price is then the overall value is from this the smart contract takes from the user and adds it back into the pool from which it was taken all the users pool tokens are destroyed of the value is sent to the user extra notes burning fees will require storing the fee rate at in a given update interval since fee rate can change at any time and using that to calculate how much quote token a user is entitled to during updateaggregatebalance completion criteria added by devs where n is the fee rate as a percentage n means fee is when a user mints n of the settlement tokens they minted with is stored in the user s commitment data when a user mints then claims after update interval the pool should still be in ownership of n of the original commitment amount of settlement tokens when a user mints n of the settlement tokens they minted with should be added to the balance of the side to which they are minting e g if they mint to the long side with a fee directly after minting should be added to leveragedpool longbalance when a user burns n of the settlement tokens they would normally be entitled to are in the user s aggregate balance after the update interval passes when a user burns upkeep happens then they update their balance usually through claiming n of the settlement tokens they gained should be kept in the pool and that amount added to the correct side of the pool either longbalance or shortbalance the same as above but with n of the settlement tokens deposited to the user s wallet after claiming if n then the user should get the exact amount they would be entitled to before the changes of this pr were implemented | 1 |
794,712 | 28,045,852,771 | IssuesEvent | 2023-03-28 22:48:33 | fleek-network/ursa | https://api.github.com/repos/fleek-network/ursa | opened | chore: send advertisement on `put` | high-priority | # Description
<!-- Please include a summary of the issue, including motivation, context, and the behavior that should be expected -->
Edge nodes are not advertising to the indexer after they pull content from other nodes (data replication).
# Checklist
- [ ] ~I have ensured that my version is up-to-date~
- [ ] ~I have ensured that my issue is reproducible~
- [ ] I have ensured that my issue is not a duplicate
| 1.0 | chore: send advertisement on `put` - # Description
<!-- Please include a summary of the issue, including motivation, context, and the behavior that should be expected -->
Edge nodes are not advertising to the indexer after they pull content from other nodes (data replication).
# Checklist
- [ ] ~I have ensured that my version is up-to-date~
- [ ] ~I have ensured that my issue is reproducible~
- [ ] I have ensured that my issue is not a duplicate
| priority | chore send advertisement on put description edge nodes are not advertising to the indexer after they pull content from other nodes data replication checklist i have ensured that my version is up to date i have ensured that my issue is reproducible i have ensured that my issue is not a duplicate | 1 |
536,158 | 15,705,159,411 | IssuesEvent | 2021-03-26 15:49:32 | nf-core/nf-co.re | https://api.github.com/repos/nf-core/nf-co.re | opened | Web launch tool previews JSON with quotes | bug high-priority pipeline-tools | The web launch final JSON preview wraps non-string entities in quotes. If copied into a params file for Nextflow, this breaks with the new schema validation code as that checks variable types and complains that everything is a string.
Steps to reproduce:
* Clicking _Launch_ on a pipeline such as nf-core/eager
* Fill in the form, setting some boolean and numeric values
* Check the _Launch parameters saved_ page after submission - preview JSON has quotes around booleans and numeric fields:
For example:
```json
{
"run_bam_filtering": "true",
"bamutils_clip_half_udg_left": "2"
}
```
Should be:
```json
{
"run_bam_filtering": true,
"bamutils_clip_half_udg_left": 2
}
```
If running `nf-core launch --id xxx` to pull these results, they go via the command line tools and the correct types seem to be used in the output JSON (though that has its [own problems](https://github.com/nf-core/tools/issues/976) still!). | 1.0 | Web launch tool previews JSON with quotes - The web launch final JSON preview wraps non-string entities in quotes. If copied into a params file for Nextflow, this breaks with the new schema validation code as that checks variable types and complains that everything is a string.
Steps to reproduce:
* Clicking _Launch_ on a pipeline such as nf-core/eager
* Fill in the form, setting some boolean and numeric values
* Check the _Launch parameters saved_ page after submission - preview JSON has quotes around booleans and numeric fields:
For example:
```json
{
"run_bam_filtering": "true",
"bamutils_clip_half_udg_left": "2"
}
```
Should be:
```json
{
"run_bam_filtering": true,
"bamutils_clip_half_udg_left": 2
}
```
If running `nf-core launch --id xxx` to pull these results, they go via the command line tools and the correct types seem to be used in the output JSON (though that has its [own problems](https://github.com/nf-core/tools/issues/976) still!). | priority | web launch tool previews json with quotes the web launch final json preview wraps non string entities in quotes if copied into a params file for nextflow this breaks with the new schema validation code as that checks variable types and complains that everything is a string steps to reproduce clicking launch on a pipeline such as nf core eager fill in the form setting some boolean and numeric values check the launch parameters saved page after submission preview json has quotes around booleans and numeric fields for example json run bam filtering true bamutils clip half udg left should be json run bam filtering true bamutils clip half udg left if running nf core launch id xxx to pull these results they go via the command line tools and the correct types seem to be used in the output json though that has its still | 1 |
639,969 | 20,770,231,107 | IssuesEvent | 2022-03-16 03:15:47 | NCC-CNC/whattodo | https://api.github.com/repos/NCC-CNC/whattodo | closed | Change "weights" text to "relative importance" | enhancement high priority | * change "weights" label in sidebar
* remove weights from text header in the excel spreadsheet code | 1.0 | Change "weights" text to "relative importance" - * change "weights" label in sidebar
* remove weights from text header in the excel spreadsheet code | priority | change weights text to relative importance change weights label in sidebar remove weights from text header in the excel spreadsheet code | 1 |
203,208 | 7,058,651,583 | IssuesEvent | 2018-01-04 21:17:44 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | opened | Make sure that none of the MergeOperations implement MutatingOperation | Priority: High Team: Core | David, please make sure that none of the MergeOperations in the codebase implement `MutatingOperation` after you finish the MergePolicies PRD. Thx a lot! | 1.0 | Make sure that none of the MergeOperations implement MutatingOperation - David, please make sure that none of the MergeOperations in the codebase implement `MutatingOperation` after you finish the MergePolicies PRD. Thx a lot! | priority | make sure that none of the mergeoperations implement mutatingoperation david please make sure that none of the mergeoperations in the codebase implement mutatingoperation after you finish the mergepolicies prd thx a lot | 1 |
704,361 | 24,194,142,647 | IssuesEvent | 2022-09-23 21:04:10 | returntocorp/semgrep | https://api.github.com/repos/returntocorp/semgrep | closed | CLI hangs after scan | priority:high | **Describe the bug**
I run semgrep as part of a bash script, and for some reason it does not hand control back to the parent script.
**To Reproduce**
Steps to reproduce the behavior, ideally a link to https://semgrep.dev:
~/.local/bin/semgrep --config auto --json -o ../atd_results/$repo_name.findings.json ./
I'm targeting a clone of ["Small Test Repo"](https://github.com/rtyley/small-test-repo)
**Expected behavior**
I get the report on STDERR, I see the json file, but the program just hangs, never terminating.
**Screenshots**
```
./test-agent.sh
mkdir: cannot create directory ‘atd_results’: File exists
mkdir: cannot create directory ‘temp_repo’: File exists
Initialized empty Git repository in /home/jason/atd/temp_repo/.git/
url: git@github.com:rtyley/small-test-repo.git
Fetching origin
Switched to branch 'master'
Gathering Code Metrics
Static Code Analysis
Semgrep rule registry URL is https://semgrep.dev/registry.
Scanning 2 files with 52 <multilang> rules.
100%|███████████████████████████████████████████████████████████████████████████████████████████████████████|2/2 tasks
Some files were skipped or only partially analyzed.
Scan was limited to files tracked by git.
(need more rules? `semgrep login` for additional free Semgrep Registry rules)
Ran 1036 rules on 1 file: 0 findings.
If Semgrep missed a finding, please send us feedback to let us know!
$ semgrep shouldafound --help
^C
```
**What is the priority of the bug to you?**
- [x] P0: blocking your adoption of Semgrep or workflow
- [ ] P1: important to fix or quite annoying
- [ ] P2: regular bug that should get fixed
**Environment**
brew installed semgrep on Ubuntu 22.04 in WSL2.0 Windows 11.
**Use case**
I'm trying to get semgrep reports alongside pymetrics into our dashboard.
| 1.0 | CLI hangs after scan - **Describe the bug**
I run semgrep as part of a bash script, and for some reason it does not hand control back to the parent script.
**To Reproduce**
Steps to reproduce the behavior, ideally a link to https://semgrep.dev:
~/.local/bin/semgrep --config auto --json -o ../atd_results/$repo_name.findings.json ./
I'm targeting a clone of ["Small Test Repo"](https://github.com/rtyley/small-test-repo)
**Expected behavior**
I get the report on STDERR, I see the json file, but the program just hangs, never terminating.
**Screenshots**
```
./test-agent.sh
mkdir: cannot create directory ‘atd_results’: File exists
mkdir: cannot create directory ‘temp_repo’: File exists
Initialized empty Git repository in /home/jason/atd/temp_repo/.git/
url: git@github.com:rtyley/small-test-repo.git
Fetching origin
Switched to branch 'master'
Gathering Code Metrics
Static Code Analysis
Semgrep rule registry URL is https://semgrep.dev/registry.
Scanning 2 files with 52 <multilang> rules.
100%|███████████████████████████████████████████████████████████████████████████████████████████████████████|2/2 tasks
Some files were skipped or only partially analyzed.
Scan was limited to files tracked by git.
(need more rules? `semgrep login` for additional free Semgrep Registry rules)
Ran 1036 rules on 1 file: 0 findings.
If Semgrep missed a finding, please send us feedback to let us know!
$ semgrep shouldafound --help
^C
```
**What is the priority of the bug to you?**
- [x] P0: blocking your adoption of Semgrep or workflow
- [ ] P1: important to fix or quite annoying
- [ ] P2: regular bug that should get fixed
**Environment**
brew installed semgrep on Ubuntu 22.04 in WSL2.0 Windows 11.
**Use case**
I'm trying to get semgrep reports alongside pymetrics into our dashboard.
| priority | cli hangs after scan describe the bug i run semgrep as part of a bash script and for some reason it does not hand control back to the parent script to reproduce steps to reproduce the behavior ideally a link to local bin semgrep config auto json o atd results repo name findings json i m targeting a clone of expected behavior i get the report on stderr i see the json file but the program just hangs never terminating screenshots test agent sh mkdir cannot create directory ‘atd results’ file exists mkdir cannot create directory ‘temp repo’ file exists initialized empty git repository in home jason atd temp repo git url git github com rtyley small test repo git fetching origin switched to branch master gathering code metrics static code analysis semgrep rule registry url is scanning files with rules ███████████████████████████████████████████████████████████████████████████████████████████████████████ tasks some files were skipped or only partially analyzed scan was limited to files tracked by git need more rules semgrep login for additional free semgrep registry rules ran rules on file findings if semgrep missed a finding please send us feedback to let us know semgrep shouldafound help c what is the priority of the bug to you blocking your adoption of semgrep or workflow important to fix or quite annoying regular bug that should get fixed environment brew installed semgrep on ubuntu in windows use case i m trying to get semgrep reports alongside pymetrics into our dashboard | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.