Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
310,955
9,526,388,025
IssuesEvent
2019-04-28 19:36:48
zanderbowen/AAE560-SP2019
https://api.github.com/repos/zanderbowen/AAE560-SP2019
closed
General Work Order Class Items
high priority stochastic
- [x] Add a column to obj.routing.Edges for TotalActual in the WorkOrder constructor method - [x] Write a method to sum the actual delivery time and actual hours worked time. This will need to run after each cycle to update accordingly.
1.0
General Work Order Class Items - - [x] Add a column to obj.routing.Edges for TotalActual in the WorkOrder constructor method - [x] Write a method to sum the actual delivery time and actual hours worked time. This will need to run after each cycle to update accordingly.
priority
general work order class items add a column to obj routing edges for totalactual in the workorder constructor method write a method to sum the actual delivery time and actual hours worked time this will need to run after each cycle to update accordingly
1
606,587
18,765,545,118
IssuesEvent
2021-11-05 23:10:16
gpbl/react-day-picker
https://api.github.com/repos/gpbl/react-day-picker
closed
v8. Markup is broken with "showOutsideDays" and "fixedWeeks" options in December 2022
Type: Bug Priority: High
**Describe the bug** Layout is broken with `"react-day-picker": "8.0.0-beta.29"` under certain conditions. **To Reproduce** reproduced here: https://codesandbox.io/s/daypicker-basicscustomization-weeknumber-example-forked-7ej5b?file=/example.tsx using next month button scroll to December 2022. While other months are rendered just fine, this one brakes layout **Screenshots** ![image](https://user-images.githubusercontent.com/83644/113855575-dc39af80-97a8-11eb-8fd1-3c79a9a8d34f.png) **Additional context** used beta 29 (as 30 doesn't seemed to be published to npm at the moment of writing)
1.0
v8. Markup is broken with "showOutsideDays" and "fixedWeeks" options in December 2022 - **Describe the bug** Layout is broken with `"react-day-picker": "8.0.0-beta.29"` under certain conditions. **To Reproduce** reproduced here: https://codesandbox.io/s/daypicker-basicscustomization-weeknumber-example-forked-7ej5b?file=/example.tsx using next month button scroll to December 2022. While other months are rendered just fine, this one brakes layout **Screenshots** ![image](https://user-images.githubusercontent.com/83644/113855575-dc39af80-97a8-11eb-8fd1-3c79a9a8d34f.png) **Additional context** used beta 29 (as 30 doesn't seemed to be published to npm at the moment of writing)
priority
markup is broken with showoutsidedays and fixedweeks options in december describe the bug layout is broken with react day picker beta under certain conditions to reproduce reproduced here using next month button scroll to december while other months are rendered just fine this one brakes layout screenshots additional context used beta as doesn t seemed to be published to npm at the moment of writing
1
512,646
14,906,265,190
IssuesEvent
2021-01-22 00:10:03
netlify/next-on-netlify
https://api.github.com/repos/netlify/next-on-netlify
closed
_redirects sorted wrong when using catch all route and dynamic routes
priority: high type: bug
The change [#145](https://github.com/netlify/next-on-netlify/pull/145) "revert route/redirect sorting logic to static then dynamic" causes the `_redirects` file to be sorted wrong when using an optional catch all route and dynamic paths. ## How to reproduce This problem happens when we have the following routes: ``` pages/_app.js pages/[[...any]].js pages/[bar]/test.js pages/test.js ``` Clone https://github.com/amuttsch/non-routes Execute the following command ```bash yarn install yarn netlify ``` ### Problem The resulting `_redirects` file looks like this on version `^2.8.3` ``` # Next-on-Netlify Redirects /_next/data/tveFS5JTA64mmnqiHIWfA/test.json /.netlify/functions/next_test 200 /_next/data/tveFS5JTA64mmnqiHIWfA/index.json /.netlify/functions/next_any 200 /_next/data/tveFS5JTA64mmnqiHIWfA/* /.netlify/functions/next_any 200 <------ catches ":bar/test.json" /api/hello /.netlify/functions/next_api_hello 200 /test /.netlify/functions/next_test 200 /_next/data/tveFS5JTA64mmnqiHIWfA/:bar/test.json /.netlify/functions/next_bar_test 200 /_next/image* url=:url w=:width q=:quality /.netlify/functions/next_image?url=:url&w=:width&q=:quality 200 /:bar/test /.netlify/functions/next_bar_test 200 / /.netlify/functions/next_any 200 /_next/* /_next/:splat 200 /* /.netlify/functions/next_any 200 ``` The data routes `/_next/data` are split into two groups. `/_next/data/tveFS5JTA64mmnqiHIWfA/* /.netlify/functions/next_any 200` catches any data route, but `/_next/data/tveFS5JTA64mmnqiHIWfA/:bar/test.json /.netlify/functions/next_bar_test 200` is below this redirect. This causes our application to fail to load data when the user does a client side transition. ### Expected Previous to `2.8.3` the file looks like this: ``` # Next-on-Netlify Redirects /_next/data/xcFf_YKEf4dQzabGKlRTV/test.json /.netlify/functions/next_test 200 /_next/data/xcFf_YKEf4dQzabGKlRTV/:bar/test.json /.netlify/functions/next_bar_test 200 /_next/data/xcFf_YKEf4dQzabGKlRTV/index.json /.netlify/functions/next_any 200 /_next/data/xcFf_YKEf4dQzabGKlRTV/* /.netlify/functions/next_any 200 /_next/image* url=:url w=:width q=:quality /.netlify/functions/next_image?url=:url&w=:width&q=:quality 200 /api/hello /.netlify/functions/next_api_hello 200 /test /.netlify/functions/next_test 200 /:bar/test /.netlify/functions/next_bar_test 200 / /.netlify/functions/next_any 200 /_next/* /_next/:splat 200 /* /.netlify/functions/next_any 200 ``` Workaround: Downgrade `next-on-netlify` to `2.8.2`.
1.0
_redirects sorted wrong when using catch all route and dynamic routes - The change [#145](https://github.com/netlify/next-on-netlify/pull/145) "revert route/redirect sorting logic to static then dynamic" causes the `_redirects` file to be sorted wrong when using an optional catch all route and dynamic paths. ## How to reproduce This problem happens when we have the following routes: ``` pages/_app.js pages/[[...any]].js pages/[bar]/test.js pages/test.js ``` Clone https://github.com/amuttsch/non-routes Execute the following command ```bash yarn install yarn netlify ``` ### Problem The resulting `_redirects` file looks like this on version `^2.8.3` ``` # Next-on-Netlify Redirects /_next/data/tveFS5JTA64mmnqiHIWfA/test.json /.netlify/functions/next_test 200 /_next/data/tveFS5JTA64mmnqiHIWfA/index.json /.netlify/functions/next_any 200 /_next/data/tveFS5JTA64mmnqiHIWfA/* /.netlify/functions/next_any 200 <------ catches ":bar/test.json" /api/hello /.netlify/functions/next_api_hello 200 /test /.netlify/functions/next_test 200 /_next/data/tveFS5JTA64mmnqiHIWfA/:bar/test.json /.netlify/functions/next_bar_test 200 /_next/image* url=:url w=:width q=:quality /.netlify/functions/next_image?url=:url&w=:width&q=:quality 200 /:bar/test /.netlify/functions/next_bar_test 200 / /.netlify/functions/next_any 200 /_next/* /_next/:splat 200 /* /.netlify/functions/next_any 200 ``` The data routes `/_next/data` are split into two groups. `/_next/data/tveFS5JTA64mmnqiHIWfA/* /.netlify/functions/next_any 200` catches any data route, but `/_next/data/tveFS5JTA64mmnqiHIWfA/:bar/test.json /.netlify/functions/next_bar_test 200` is below this redirect. This causes our application to fail to load data when the user does a client side transition. ### Expected Previous to `2.8.3` the file looks like this: ``` # Next-on-Netlify Redirects /_next/data/xcFf_YKEf4dQzabGKlRTV/test.json /.netlify/functions/next_test 200 /_next/data/xcFf_YKEf4dQzabGKlRTV/:bar/test.json /.netlify/functions/next_bar_test 200 /_next/data/xcFf_YKEf4dQzabGKlRTV/index.json /.netlify/functions/next_any 200 /_next/data/xcFf_YKEf4dQzabGKlRTV/* /.netlify/functions/next_any 200 /_next/image* url=:url w=:width q=:quality /.netlify/functions/next_image?url=:url&w=:width&q=:quality 200 /api/hello /.netlify/functions/next_api_hello 200 /test /.netlify/functions/next_test 200 /:bar/test /.netlify/functions/next_bar_test 200 / /.netlify/functions/next_any 200 /_next/* /_next/:splat 200 /* /.netlify/functions/next_any 200 ``` Workaround: Downgrade `next-on-netlify` to `2.8.2`.
priority
redirects sorted wrong when using catch all route and dynamic routes the change revert route redirect sorting logic to static then dynamic causes the redirects file to be sorted wrong when using an optional catch all route and dynamic paths how to reproduce this problem happens when we have the following routes pages app js pages js pages test js pages test js clone execute the following command bash yarn install yarn netlify problem the resulting redirects file looks like this on version next on netlify redirects next data test json netlify functions next test next data index json netlify functions next any next data netlify functions next any catches bar test json api hello netlify functions next api hello test netlify functions next test next data bar test json netlify functions next bar test next image url url w width q quality netlify functions next image url url w width q quality bar test netlify functions next bar test netlify functions next any next next splat netlify functions next any the data routes next data are split into two groups next data netlify functions next any catches any data route but next data bar test json netlify functions next bar test is below this redirect this causes our application to fail to load data when the user does a client side transition expected previous to the file looks like this next on netlify redirects next data xcff test json netlify functions next test next data xcff bar test json netlify functions next bar test next data xcff index json netlify functions next any next data xcff netlify functions next any next image url url w width q quality netlify functions next image url url w width q quality api hello netlify functions next api hello test netlify functions next test bar test netlify functions next bar test netlify functions next any next next splat netlify functions next any workaround downgrade next on netlify to
1
636,378
20,598,670,187
IssuesEvent
2022-03-05 23:02:01
localstack/localstack
https://api.github.com/repos/localstack/localstack
closed
bug: SSL certificate issue since v0.12.18
bug priority-high needs-triaging infra-startup networking
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior LocalStack is taking a long time to start up since updating to v0.12.18 due to timing out when attempting to pull down the local test SSL certificate. This works fine on v0.12.17. ``` localstack_1 | Starting edge router (https port 4566)... localstack_1 | 2021-11-11T09:33:23:INFO:bootstrap.py: Execution of "load_plugin_from_path" took 2430.59ms localstack_1 | 2021-11-11T09:33:23:INFO:bootstrap.py: Execution of "load_plugins" took 2431.15ms localstack_1 | Waiting for all LocalStack services to be ready localstack_1 | ... localstack_1 | Waiting for all LocalStack services to be ready localstack_1 | 2021-11-11T09:46:28:INFO:localstack_ext.bootstrap.install: Unable to download local test SSL certificate from https://cdn.jsdelivr.net/gh/localstack/localstack-artifacts@master/local-certs/server.key to /tmp/localstack/server.test.pem: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out')) Traceback (most recent call last): localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 175, in _new_conn localstack_1 | (self._dns_host, self.port), self.timeout, **extra_kw localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/connection.py", line 96, in create_connection localstack_1 | raise err localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/connection.py", line 86, in create_connection localstack_1 | sock.connect(sa) localstack_1 | TimeoutError: [Errno 110] Operation timed out localstack_1 | localstack_1 | During handling of the above exception, another exception occurred: localstack_1 | localstack_1 | Traceback (most recent call last): localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 706, in urlopen localstack_1 | chunked=chunked, localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 382, in _make_request localstack_1 | self._validate_conn(conn) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 1010, in _validate_conn localstack_1 | conn.connect() localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 358, in connect localstack_1 | conn = self._new_conn() localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 187, in _new_conn localstack_1 | self, "Failed to establish a new connection: %s" % e localstack_1 | urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out localstack_1 | localstack_1 | During handling of the above exception, another exception occurred: localstack_1 | localstack_1 | Traceback (most recent call last): localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/adapters.py", line 449, in send localstack_1 | timeout=timeout localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 756, in urlopen localstack_1 | method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2] localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/retry.py", line 574, in increment localstack_1 | raise MaxRetryError(_pool, url, error or ResponseError(cause)) localstack_1 | urllib3.exceptions.MaxRetryError: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out')) localstack_1 | localstack_1 | During handling of the above exception, another exception occurred: localstack_1 | localstack_1 | Traceback (most recent call last): localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/localstack_ext/bootstrap/install.py", line 87, in do_download localstack_1 | download(url,target_file) localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 1072, in download localstack_1 | r = s.get(url, stream=True, verify=os.getenv("REQUESTS_CA_BUNDLE", verify_ssl)) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 555, in get localstack_1 | return self.request('GET', url, **kwargs) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 542, in request localstack_1 | resp = self.send(prep, **send_kwargs) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 655, in send localstack_1 | r = adapter.send(request, **kwargs) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/adapters.py", line 516, in send localstack_1 | raise ConnectionError(e, request=request) localstack_1 | requests.exceptions.ConnectionError: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out')) ``` I expect this has something to do with this line in the release notes (but not 100%) > add startup logic to install prebuilt SSL cert if available ### Expected Behavior LocalStack to start up in a normal amount of time ### How are you starting LocalStack? With a docker-compose file ### Steps To Reproduce #### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`) Using a custom docker-compose that starts localstack up as part of a suite of services #### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands) N/A ### Environment ```markdown - OS: - LocalStack: v0.12.18 to latest ``` ### Anything else? The localstack image works fine on my local machine, but does not work when running on the CI server. The CI server is more locked down and lacks external connectivity without a proxy, so this may be causing the issue.
1.0
bug: SSL certificate issue since v0.12.18 - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior LocalStack is taking a long time to start up since updating to v0.12.18 due to timing out when attempting to pull down the local test SSL certificate. This works fine on v0.12.17. ``` localstack_1 | Starting edge router (https port 4566)... localstack_1 | 2021-11-11T09:33:23:INFO:bootstrap.py: Execution of "load_plugin_from_path" took 2430.59ms localstack_1 | 2021-11-11T09:33:23:INFO:bootstrap.py: Execution of "load_plugins" took 2431.15ms localstack_1 | Waiting for all LocalStack services to be ready localstack_1 | ... localstack_1 | Waiting for all LocalStack services to be ready localstack_1 | 2021-11-11T09:46:28:INFO:localstack_ext.bootstrap.install: Unable to download local test SSL certificate from https://cdn.jsdelivr.net/gh/localstack/localstack-artifacts@master/local-certs/server.key to /tmp/localstack/server.test.pem: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out')) Traceback (most recent call last): localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 175, in _new_conn localstack_1 | (self._dns_host, self.port), self.timeout, **extra_kw localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/connection.py", line 96, in create_connection localstack_1 | raise err localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/connection.py", line 86, in create_connection localstack_1 | sock.connect(sa) localstack_1 | TimeoutError: [Errno 110] Operation timed out localstack_1 | localstack_1 | During handling of the above exception, another exception occurred: localstack_1 | localstack_1 | Traceback (most recent call last): localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 706, in urlopen localstack_1 | chunked=chunked, localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 382, in _make_request localstack_1 | self._validate_conn(conn) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 1010, in _validate_conn localstack_1 | conn.connect() localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 358, in connect localstack_1 | conn = self._new_conn() localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connection.py", line 187, in _new_conn localstack_1 | self, "Failed to establish a new connection: %s" % e localstack_1 | urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out localstack_1 | localstack_1 | During handling of the above exception, another exception occurred: localstack_1 | localstack_1 | Traceback (most recent call last): localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/adapters.py", line 449, in send localstack_1 | timeout=timeout localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/connectionpool.py", line 756, in urlopen localstack_1 | method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2] localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/urllib3/util/retry.py", line 574, in increment localstack_1 | raise MaxRetryError(_pool, url, error or ResponseError(cause)) localstack_1 | urllib3.exceptions.MaxRetryError: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out')) localstack_1 | localstack_1 | During handling of the above exception, another exception occurred: localstack_1 | localstack_1 | Traceback (most recent call last): localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/localstack_ext/bootstrap/install.py", line 87, in do_download localstack_1 | download(url,target_file) localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 1072, in download localstack_1 | r = s.get(url, stream=True, verify=os.getenv("REQUESTS_CA_BUNDLE", verify_ssl)) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 555, in get localstack_1 | return self.request('GET', url, **kwargs) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 542, in request localstack_1 | resp = self.send(prep, **send_kwargs) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/sessions.py", line 655, in send localstack_1 | r = adapter.send(request, **kwargs) localstack_1 | File "/opt/code/localstack/.venv/lib/python3.7/site-packages/requests/adapters.py", line 516, in send localstack_1 | raise ConnectionError(e, request=request) localstack_1 | requests.exceptions.ConnectionError: MyHTTPSConnectionPool(host='cdn.jsdelivr.net', port=443): Max retries exceeded with url: /gh/localstack/localstack-artifacts@master/local-certs/server.key (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f5ad9f41cd0>: Failed to establish a new connection: [Errno 110] Operation timed out')) ``` I expect this has something to do with this line in the release notes (but not 100%) > add startup logic to install prebuilt SSL cert if available ### Expected Behavior LocalStack to start up in a normal amount of time ### How are you starting LocalStack? With a docker-compose file ### Steps To Reproduce #### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`) Using a custom docker-compose that starts localstack up as part of a suite of services #### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands) N/A ### Environment ```markdown - OS: - LocalStack: v0.12.18 to latest ``` ### Anything else? The localstack image works fine on my local machine, but does not work when running on the CI server. The CI server is more locked down and lacks external connectivity without a proxy, so this may be causing the issue.
priority
bug ssl certificate issue since is there an existing issue for this i have searched the existing issues current behavior localstack is taking a long time to start up since updating to due to timing out when attempting to pull down the local test ssl certificate this works fine on localstack starting edge router https port localstack info bootstrap py execution of load plugin from path took localstack info bootstrap py execution of load plugins took localstack waiting for all localstack services to be ready localstack localstack waiting for all localstack services to be ready localstack info localstack ext bootstrap install unable to download local test ssl certificate from to tmp localstack server test pem myhttpsconnectionpool host cdn jsdelivr net port max retries exceeded with url gh localstack localstack artifacts master local certs server key caused by newconnectionerror failed to establish a new connection operation timed out traceback most recent call last localstack file opt code localstack venv lib site packages connection py line in new conn localstack self dns host self port self timeout extra kw localstack file opt code localstack venv lib site packages util connection py line in create connection localstack raise err localstack file opt code localstack venv lib site packages util connection py line in create connection localstack sock connect sa localstack timeouterror operation timed out localstack localstack during handling of the above exception another exception occurred localstack localstack traceback most recent call last localstack file opt code localstack venv lib site packages connectionpool py line in urlopen localstack chunked chunked localstack file opt code localstack venv lib site packages connectionpool py line in make request localstack self validate conn conn localstack file opt code localstack venv lib site packages connectionpool py line in validate conn localstack conn connect localstack file opt code localstack venv lib site packages connection py line in connect localstack conn self new conn localstack file opt code localstack venv lib site packages connection py line in new conn localstack self failed to establish a new connection s e localstack exceptions newconnectionerror failed to establish a new connection operation timed out localstack localstack during handling of the above exception another exception occurred localstack localstack traceback most recent call last localstack file opt code localstack venv lib site packages requests adapters py line in send localstack timeout timeout localstack file opt code localstack venv lib site packages connectionpool py line in urlopen localstack method url error e pool self stacktrace sys exc info localstack file opt code localstack venv lib site packages util retry py line in increment localstack raise maxretryerror pool url error or responseerror cause localstack exceptions maxretryerror myhttpsconnectionpool host cdn jsdelivr net port max retries exceeded with url gh localstack localstack artifacts master local certs server key caused by newconnectionerror failed to establish a new connection operation timed out localstack localstack during handling of the above exception another exception occurred localstack localstack traceback most recent call last localstack file opt code localstack venv lib site packages localstack ext bootstrap install py line in do download localstack download url target file localstack file opt code localstack localstack utils common py line in download localstack r s get url stream true verify os getenv requests ca bundle verify ssl localstack file opt code localstack venv lib site packages requests sessions py line in get localstack return self request get url kwargs localstack file opt code localstack venv lib site packages requests sessions py line in request localstack resp self send prep send kwargs localstack file opt code localstack venv lib site packages requests sessions py line in send localstack r adapter send request kwargs localstack file opt code localstack venv lib site packages requests adapters py line in send localstack raise connectionerror e request request localstack requests exceptions connectionerror myhttpsconnectionpool host cdn jsdelivr net port max retries exceeded with url gh localstack localstack artifacts master local certs server key caused by newconnectionerror failed to establish a new connection operation timed out i expect this has something to do with this line in the release notes but not add startup logic to install prebuilt ssl cert if available expected behavior localstack to start up in a normal amount of time how are you starting localstack with a docker compose file steps to reproduce how are you starting localstack e g bin localstack command arguments or docker compose yml using a custom docker compose that starts localstack up as part of a suite of services client commands e g aws sdk code snippet or sequence of awslocal commands n a environment markdown os localstack to latest anything else the localstack image works fine on my local machine but does not work when running on the ci server the ci server is more locked down and lacks external connectivity without a proxy so this may be causing the issue
1
451,235
13,031,722,226
IssuesEvent
2020-07-28 02:06:56
scprogramming/Open-Source-Scan
https://api.github.com/repos/scprogramming/Open-Source-Scan
opened
Add a way to clear scans and all non cpe/cve data
Analysis High Priority
I should have a way to delete scans, clear all scans, projects, etc to preserve cpe/cve data, but allow for clearning unneeded data.
1.0
Add a way to clear scans and all non cpe/cve data - I should have a way to delete scans, clear all scans, projects, etc to preserve cpe/cve data, but allow for clearning unneeded data.
priority
add a way to clear scans and all non cpe cve data i should have a way to delete scans clear all scans projects etc to preserve cpe cve data but allow for clearning unneeded data
1
236,909
7,753,586,792
IssuesEvent
2018-05-31 01:33:49
Gloirin/m2gTest
https://api.github.com/repos/Gloirin/m2gTest
closed
0006634: custom fields missing in XLS export
Addressbook bug high priority
**Reported by cweiss on 18 Jun 2012 07:49** **Version:** Milan (2012-03-3) custom fields missing in XLS export
1.0
0006634: custom fields missing in XLS export - **Reported by cweiss on 18 Jun 2012 07:49** **Version:** Milan (2012-03-3) custom fields missing in XLS export
priority
custom fields missing in xls export reported by cweiss on jun version milan custom fields missing in xls export
1
718,537
24,721,619,448
IssuesEvent
2022-10-20 11:08:43
harvester/harvester
https://api.github.com/repos/harvester/harvester
closed
[FEATURE]Dedicated storage network
kind/enhancement area/ui area/installer area/network priority/0 area/storage highlight blocker area/longhorn-related require-ui/small
We would like to have a dedicated storage network interface that can be specified (either as single nic or bonded NICs) to separate the storage traffic from the overlay traffic. Depends on https://github.com/longhorn/longhorn/issues/2285 and #1048
1.0
[FEATURE]Dedicated storage network - We would like to have a dedicated storage network interface that can be specified (either as single nic or bonded NICs) to separate the storage traffic from the overlay traffic. Depends on https://github.com/longhorn/longhorn/issues/2285 and #1048
priority
dedicated storage network we would like to have a dedicated storage network interface that can be specified either as single nic or bonded nics to separate the storage traffic from the overlay traffic depends on and
1
212,122
7,228,799,288
IssuesEvent
2018-02-11 13:41:13
allure-framework/allure2
https://api.github.com/repos/allure-framework/allure2
closed
Packages tab is not shown in demo report due to invalid initializer
priority:high theme:ui type:bug work:review
[//]: # ( . Note: for support questions, please use Stackoverflow or Gitter**. . This repository's issues are reserved for feature requests and bug reports. . . In case of any problems with Allure Jenkins plugin** please use the following repository . to create an issue: https://github.com/jenkinsci/allure-plugin/issues . . Make sure you have a clear name for your issue. The name should start with a capital . letter and no dot is required in the end of the sentence. An example of good issue names: . . - The report is broken in IE11 . - Add an ability to disable default plugins . - Support emoji in test descriptions ) #### I'm submitting a ... - [x] bug report - [ ] feature request - [ ] support request => Please do not submit support request here, see note at the top of this template. #### What is the current behavior? ![2018-02-05_13-26-42](https://user-images.githubusercontent.com/9446141/35799926-641e9864-0a78-11e8-83ce-fb1ae16cf254.png) #### If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem Screenshot is attached. #### What is the expected behavior? Correct identifier, Packages tab is shown #### What is the motivation / use case for changing the behavior? #### Please tell us about your environment: | Allure version | 2.2.0 | | --- | --- | | Test framework | testng@6.8 | | Allure adaptor | allure-testng@2.0-BETA11 | | Generate report using | allure-maven@2.18 | #### Other information [//]: # ( . e.g. detailed explanation, stacktraces, related issues, suggestions . how to fix, links for us to have more context, eg. Stackoverflow, Gitter etc ) <!-- Love allure-report? Please consider supporting our collective: 👉 https://opencollective.com/allure-report/donate -->
1.0
Packages tab is not shown in demo report due to invalid initializer - [//]: # ( . Note: for support questions, please use Stackoverflow or Gitter**. . This repository's issues are reserved for feature requests and bug reports. . . In case of any problems with Allure Jenkins plugin** please use the following repository . to create an issue: https://github.com/jenkinsci/allure-plugin/issues . . Make sure you have a clear name for your issue. The name should start with a capital . letter and no dot is required in the end of the sentence. An example of good issue names: . . - The report is broken in IE11 . - Add an ability to disable default plugins . - Support emoji in test descriptions ) #### I'm submitting a ... - [x] bug report - [ ] feature request - [ ] support request => Please do not submit support request here, see note at the top of this template. #### What is the current behavior? ![2018-02-05_13-26-42](https://user-images.githubusercontent.com/9446141/35799926-641e9864-0a78-11e8-83ce-fb1ae16cf254.png) #### If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem Screenshot is attached. #### What is the expected behavior? Correct identifier, Packages tab is shown #### What is the motivation / use case for changing the behavior? #### Please tell us about your environment: | Allure version | 2.2.0 | | --- | --- | | Test framework | testng@6.8 | | Allure adaptor | allure-testng@2.0-BETA11 | | Generate report using | allure-maven@2.18 | #### Other information [//]: # ( . e.g. detailed explanation, stacktraces, related issues, suggestions . how to fix, links for us to have more context, eg. Stackoverflow, Gitter etc ) <!-- Love allure-report? Please consider supporting our collective: 👉 https://opencollective.com/allure-report/donate -->
priority
packages tab is not shown in demo report due to invalid initializer note for support questions please use stackoverflow or gitter this repository s issues are reserved for feature requests and bug reports in case of any problems with allure jenkins plugin please use the following repository to create an issue make sure you have a clear name for your issue the name should start with a capital letter and no dot is required in the end of the sentence an example of good issue names the report is broken in add an ability to disable default plugins support emoji in test descriptions i m submitting a bug report feature request support request please do not submit support request here see note at the top of this template what is the current behavior if the current behavior is a bug please provide the steps to reproduce and if possible a minimal demo of the problem screenshot is attached what is the expected behavior correct identifier packages tab is shown what is the motivation use case for changing the behavior please tell us about your environment allure version test framework testng allure adaptor allure testng generate report using allure maven other information e g detailed explanation stacktraces related issues suggestions how to fix links for us to have more context eg stackoverflow gitter etc love allure report please consider supporting our collective 👉
1
628,798
20,014,459,241
IssuesEvent
2022-02-01 10:35:21
kubermatic/kubeone
https://api.github.com/repos/kubermatic/kubeone
opened
Test Cilium in kube-proxy mode -- Test Release 1.4
priority/high sig/cluster-management
Instructions: * Download the latest KubeOne 1.4.0 release candidate * Follow the [Create a Kubernetes cluster tutorial](https://docs.kubermatic.com/kubeone/master/tutorials/creating_clusters/) to create your cluster * Make sure to have Cilium enabled: ```yaml clusterNetwork: cni: cilium: enableHubble: true ``` * Wait for machine-controller-managed nodes to join the cluster * Ensure all pods are Running * Create a LoadBalancer Service and point to some test pod (e.g. Nginx pod). The LB should be reachable and serve the content as expected * Ensure Hubble is running and reachable (you might need to port-forward to it) This test can be done on a single cloud provider, on a single operating system, with a single Kubernetes version (e.g. 1.23.3 on Ubuntu on AWS).
1.0
Test Cilium in kube-proxy mode -- Test Release 1.4 - Instructions: * Download the latest KubeOne 1.4.0 release candidate * Follow the [Create a Kubernetes cluster tutorial](https://docs.kubermatic.com/kubeone/master/tutorials/creating_clusters/) to create your cluster * Make sure to have Cilium enabled: ```yaml clusterNetwork: cni: cilium: enableHubble: true ``` * Wait for machine-controller-managed nodes to join the cluster * Ensure all pods are Running * Create a LoadBalancer Service and point to some test pod (e.g. Nginx pod). The LB should be reachable and serve the content as expected * Ensure Hubble is running and reachable (you might need to port-forward to it) This test can be done on a single cloud provider, on a single operating system, with a single Kubernetes version (e.g. 1.23.3 on Ubuntu on AWS).
priority
test cilium in kube proxy mode test release instructions download the latest kubeone release candidate follow the to create your cluster make sure to have cilium enabled yaml clusternetwork cni cilium enablehubble true wait for machine controller managed nodes to join the cluster ensure all pods are running create a loadbalancer service and point to some test pod e g nginx pod the lb should be reachable and serve the content as expected ensure hubble is running and reachable you might need to port forward to it this test can be done on a single cloud provider on a single operating system with a single kubernetes version e g on ubuntu on aws
1
86,263
3,704,411,907
IssuesEvent
2016-03-01 00:04:24
Macainian/Django-Maced
https://api.github.com/repos/Macainian/Django-Maced
opened
When items are modified, they only propagate select name/value changes.
Bug High Priority
This is an extension of issue #52
1.0
When items are modified, they only propagate select name/value changes. - This is an extension of issue #52
priority
when items are modified they only propagate select name value changes this is an extension of issue
1
441,199
12,709,385,628
IssuesEvent
2020-06-23 12:16:34
AlonDiskin/Visuals
https://api.github.com/repos/AlonDiskin/Visuals
closed
Videos and recycle bin state update error
bug high priority
reproduce steps: - move some videos to recycle bin - restore all trash items from recycle bin browser screen - trash more videos from videos browser - videos browser not update items as removed - the now empty recycle bin browser do not show the new items as expected
1.0
Videos and recycle bin state update error - reproduce steps: - move some videos to recycle bin - restore all trash items from recycle bin browser screen - trash more videos from videos browser - videos browser not update items as removed - the now empty recycle bin browser do not show the new items as expected
priority
videos and recycle bin state update error reproduce steps move some videos to recycle bin restore all trash items from recycle bin browser screen trash more videos from videos browser videos browser not update items as removed the now empty recycle bin browser do not show the new items as expected
1
9,932
2,608,937,862
IssuesEvent
2015-02-26 11:05:57
CSSE1001/MyPyTutor
https://api.github.com/repos/CSSE1001/MyPyTutor
closed
Verify installer behaviour for directories with spaces on Windows
bug priority: high
The installer (new and old) need to be checked on Windows to ensure that they correctly handle directories with spaces. I've had trouble with this issue in the past, especially with the use of `os.execv`; Windows sometimes seems to like screwing up the components of the vector, even though they're given as a list.
1.0
Verify installer behaviour for directories with spaces on Windows - The installer (new and old) need to be checked on Windows to ensure that they correctly handle directories with spaces. I've had trouble with this issue in the past, especially with the use of `os.execv`; Windows sometimes seems to like screwing up the components of the vector, even though they're given as a list.
priority
verify installer behaviour for directories with spaces on windows the installer new and old need to be checked on windows to ensure that they correctly handle directories with spaces i ve had trouble with this issue in the past especially with the use of os execv windows sometimes seems to like screwing up the components of the vector even though they re given as a list
1
144,859
5,547,067,858
IssuesEvent
2017-03-23 03:45:42
CS2103JAN2017-T16-B4/main
https://api.github.com/repos/CS2103JAN2017-T16-B4/main
closed
As a user I want to edit the deadline, name and schedule of any task
priority.high type.story
Update any task as required, as well as marking it as done and deleting it
1.0
As a user I want to edit the deadline, name and schedule of any task - Update any task as required, as well as marking it as done and deleting it
priority
as a user i want to edit the deadline name and schedule of any task update any task as required as well as marking it as done and deleting it
1
266,999
8,378,159,735
IssuesEvent
2018-10-06 11:03:58
rit-sse/OneRepoToRuleThemAll
https://api.github.com/repos/rit-sse/OneRepoToRuleThemAll
closed
Pin Down Node Version
High Priority
In our `Dockerfile` we say `FROM node` which will install the latest node version. It appears that this is installing node v10. We should pin down the version (eg. `FROM node:carbon`) to the LTS release so that we're running on a consistent node version, not running on odd numbered Node releases, and can have a consistent upgrade process for our dependencies (not having them break under us). Node v10 will become LTS this month (October 2018), so we can wait until it does to pin down the version. We can also delete the `Dockerfile.edge` file as it's out-of-date and no longer needed.
1.0
Pin Down Node Version - In our `Dockerfile` we say `FROM node` which will install the latest node version. It appears that this is installing node v10. We should pin down the version (eg. `FROM node:carbon`) to the LTS release so that we're running on a consistent node version, not running on odd numbered Node releases, and can have a consistent upgrade process for our dependencies (not having them break under us). Node v10 will become LTS this month (October 2018), so we can wait until it does to pin down the version. We can also delete the `Dockerfile.edge` file as it's out-of-date and no longer needed.
priority
pin down node version in our dockerfile we say from node which will install the latest node version it appears that this is installing node we should pin down the version eg from node carbon to the lts release so that we re running on a consistent node version not running on odd numbered node releases and can have a consistent upgrade process for our dependencies not having them break under us node will become lts this month october so we can wait until it does to pin down the version we can also delete the dockerfile edge file as it s out of date and no longer needed
1
732,954
25,282,044,689
IssuesEvent
2022-11-16 16:25:50
leka/LekaOS
https://api.github.com/repos/leka/LekaOS
closed
[Story] - LekaOS BLE v1.0.0
01 - type: story 90 - priority: high
# Introduction BLE communication with the iPad is one of the most important feature of LekaOS. Without it the robot is useless and its presence allow us to: - control the robot - debug/log what's going on - tests new functionalities (FOTA, scheduling, etc.) ## Roadmap 1. Setup a basic working example using mbed 1. Test the integration with timed tasks and scheduling 1. Implement [Leka Communication Specifications](https://github.com/leka/LKAlphaComSpecs) ## 1. Setup a basic working example using mbed This first step is really easy: use the examples from mbed-os to create a simple working example of BLE for Leka to pave the way for future developments This must include: - [x] read, write and notifications services/characteristics - [x] the use of predefined services (battery) and proprietary services (temperature, firmware version, basic command) ## 2. Test integration with timed tasks and scheduling Through out our development we will need to start tasks and stop them. Gaining a deeper understanding about how this will work between BLE and mbed os is paramount as all future development will rely on that. More information in #1 - https://github.com/leka/LekaOS/issues/1#issuecomment-557007922 This must include: - [x] starting a long running thread - [x] stoping a long running thread anytime - [x] starting again from the beginning - [x] starting again where we left off (if possible) ## 3. Implement Leka Communication Specifications We'll review, update and implement the Leka Communication Specifications in parallel with the iOS app. The main goal is to have the two perfectly synchronised. This must include: - [x] implement all the services/characteristics - [x] implement the command analyser - [x] a lot of testing ## 4. Mandatory commands - [x] implement reset command and test in case of infinite loop
1.0
[Story] - LekaOS BLE v1.0.0 - # Introduction BLE communication with the iPad is one of the most important feature of LekaOS. Without it the robot is useless and its presence allow us to: - control the robot - debug/log what's going on - tests new functionalities (FOTA, scheduling, etc.) ## Roadmap 1. Setup a basic working example using mbed 1. Test the integration with timed tasks and scheduling 1. Implement [Leka Communication Specifications](https://github.com/leka/LKAlphaComSpecs) ## 1. Setup a basic working example using mbed This first step is really easy: use the examples from mbed-os to create a simple working example of BLE for Leka to pave the way for future developments This must include: - [x] read, write and notifications services/characteristics - [x] the use of predefined services (battery) and proprietary services (temperature, firmware version, basic command) ## 2. Test integration with timed tasks and scheduling Through out our development we will need to start tasks and stop them. Gaining a deeper understanding about how this will work between BLE and mbed os is paramount as all future development will rely on that. More information in #1 - https://github.com/leka/LekaOS/issues/1#issuecomment-557007922 This must include: - [x] starting a long running thread - [x] stoping a long running thread anytime - [x] starting again from the beginning - [x] starting again where we left off (if possible) ## 3. Implement Leka Communication Specifications We'll review, update and implement the Leka Communication Specifications in parallel with the iOS app. The main goal is to have the two perfectly synchronised. This must include: - [x] implement all the services/characteristics - [x] implement the command analyser - [x] a lot of testing ## 4. Mandatory commands - [x] implement reset command and test in case of infinite loop
priority
lekaos ble introduction ble communication with the ipad is one of the most important feature of lekaos without it the robot is useless and its presence allow us to control the robot debug log what s going on tests new functionalities fota scheduling etc roadmap setup a basic working example using mbed test the integration with timed tasks and scheduling implement setup a basic working example using mbed this first step is really easy use the examples from mbed os to create a simple working example of ble for leka to pave the way for future developments this must include read write and notifications services characteristics the use of predefined services battery and proprietary services temperature firmware version basic command test integration with timed tasks and scheduling through out our development we will need to start tasks and stop them gaining a deeper understanding about how this will work between ble and mbed os is paramount as all future development will rely on that more information in this must include starting a long running thread stoping a long running thread anytime starting again from the beginning starting again where we left off if possible implement leka communication specifications we ll review update and implement the leka communication specifications in parallel with the ios app the main goal is to have the two perfectly synchronised this must include implement all the services characteristics implement the command analyser a lot of testing mandatory commands implement reset command and test in case of infinite loop
1
788,885
27,771,807,270
IssuesEvent
2023-03-16 14:53:35
janus-idp/software-templates
https://api.github.com/repos/janus-idp/software-templates
opened
Various updates in the 6 Janus GPTs
kind/enhancement priority/high
1. Rename the GPTs titles. ie From: `.NET Frontend Golden Path Template` To: `Create a .NET Frontend application with CI/CD` 2. Show the CI step before the CD step. The CI step will become step 3 and Argo will be step 4. 3. Update the current description in the CI section. From: `This action will create a simple CI based on chosen method` To: `This action will create a CI pipeline for your application based on chosen method` 4. In the ArgoCD step: - Select Quay registry as default - Move the image url before the namespace 5. Make sure the ports default are correct. It seems that most templates are using 5000 as default which is incorrect.
1.0
Various updates in the 6 Janus GPTs - 1. Rename the GPTs titles. ie From: `.NET Frontend Golden Path Template` To: `Create a .NET Frontend application with CI/CD` 2. Show the CI step before the CD step. The CI step will become step 3 and Argo will be step 4. 3. Update the current description in the CI section. From: `This action will create a simple CI based on chosen method` To: `This action will create a CI pipeline for your application based on chosen method` 4. In the ArgoCD step: - Select Quay registry as default - Move the image url before the namespace 5. Make sure the ports default are correct. It seems that most templates are using 5000 as default which is incorrect.
priority
various updates in the janus gpts rename the gpts titles ie from net frontend golden path template to create a net frontend application with ci cd show the ci step before the cd step the ci step will become step and argo will be step update the current description in the ci section from this action will create a simple ci based on chosen method to this action will create a ci pipeline for your application based on chosen method in the argocd step select quay registry as default move the image url before the namespace make sure the ports default are correct it seems that most templates are using as default which is incorrect
1
197,744
6,963,453,986
IssuesEvent
2017-12-08 17:26:59
python/mypy
https://api.github.com/repos/python/mypy
closed
Slow incremental run when single module has errors
bug priority-0-high topic-daemon topic-incremental
`mypy -i` is slower than expected in the following scenario: * Check out fresh mypy repository * Create tiny module `mypy.x` that generates an error: `echo '1 + ""' > mypy/x.py` * Run `mypy -i mypy` * This takes a while, generates an error for `mypy/x.py` as expected * Run `mypy -i mypy` * This is as slow as the previous run, but caching should make this much faster
1.0
Slow incremental run when single module has errors - `mypy -i` is slower than expected in the following scenario: * Check out fresh mypy repository * Create tiny module `mypy.x` that generates an error: `echo '1 + ""' > mypy/x.py` * Run `mypy -i mypy` * This takes a while, generates an error for `mypy/x.py` as expected * Run `mypy -i mypy` * This is as slow as the previous run, but caching should make this much faster
priority
slow incremental run when single module has errors mypy i is slower than expected in the following scenario check out fresh mypy repository create tiny module mypy x that generates an error echo mypy x py run mypy i mypy this takes a while generates an error for mypy x py as expected run mypy i mypy this is as slow as the previous run but caching should make this much faster
1
529,381
15,387,560,930
IssuesEvent
2021-03-03 09:41:22
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Coverity - High Impact Outstanding issues in MDEvent files
Framework High Priority Stale
This issue was originally [TRAC 9938](http://trac.mantidproject.org/mantid/ticket/9938) There are 17 Coverity high impact outstanding issued in the `MDEvent` module. Distribute as you see fit.
1.0
Coverity - High Impact Outstanding issues in MDEvent files - This issue was originally [TRAC 9938](http://trac.mantidproject.org/mantid/ticket/9938) There are 17 Coverity high impact outstanding issued in the `MDEvent` module. Distribute as you see fit.
priority
coverity high impact outstanding issues in mdevent files this issue was originally there are coverity high impact outstanding issued in the mdevent module distribute as you see fit
1
509,616
14,740,532,189
IssuesEvent
2021-01-07 09:14:23
canonical-web-and-design/vanilla-framework
https://api.github.com/repos/canonical-web-and-design/vanilla-framework
closed
Fieldset: Transparent border on fieldset results in border colour spilling outside element
Bug 🐛 Priority: High
**Describe the bug** The transparent border currently applied to fieldsets causes a strange bug in chrome where the border color fills the window. Replacing the fieldset with a div and copyign all fieldset styles onto the div doesn't seem to trigger the bug, so it seems highly specific to the fieldset or an inherent property of it: ![image](https://user-images.githubusercontent.com/2741678/99959134-60ebda00-2d82-11eb-81e1-5a5ae4cd1c2b.png) **To Reproduce** Paste this [markup](https://pastebin.canonical.com/p/2pnRKXPPkR/) in any site using vanilla. - Device: Mac - OS: Latest - Browser Version 87.0.4280.67 (Official Build) (x86_64)
1.0
Fieldset: Transparent border on fieldset results in border colour spilling outside element - **Describe the bug** The transparent border currently applied to fieldsets causes a strange bug in chrome where the border color fills the window. Replacing the fieldset with a div and copyign all fieldset styles onto the div doesn't seem to trigger the bug, so it seems highly specific to the fieldset or an inherent property of it: ![image](https://user-images.githubusercontent.com/2741678/99959134-60ebda00-2d82-11eb-81e1-5a5ae4cd1c2b.png) **To Reproduce** Paste this [markup](https://pastebin.canonical.com/p/2pnRKXPPkR/) in any site using vanilla. - Device: Mac - OS: Latest - Browser Version 87.0.4280.67 (Official Build) (x86_64)
priority
fieldset transparent border on fieldset results in border colour spilling outside element describe the bug the transparent border currently applied to fieldsets causes a strange bug in chrome where the border color fills the window replacing the fieldset with a div and copyign all fieldset styles onto the div doesn t seem to trigger the bug so it seems highly specific to the fieldset or an inherent property of it to reproduce paste this in any site using vanilla device mac os latest browser version official build
1
155,746
5,960,139,006
IssuesEvent
2017-05-29 13:16:15
Aurorastation/Aurora.3
https://api.github.com/repos/Aurorastation/Aurora.3
opened
SMC: Mapped in power connections do not appear to be updated properly
bug:confirmed flag:development flag:high priority
Okay. Long story short. We've been running into issues in new map meme where connections that are valid, and exist since round start, are not flagged properly until an admin invokes `make-powernets`. Example cases of this: * Tesla setup. This is all mapped in to be 100% functional (as far as wiring goes). However, after starting the engine, power does not flow until `make-powernets` or a local powernet update (I presume, haven't tried) is invoked. * #2345 . Where mapped in z-wires do not move power until `make-powernets` is invoked. In this case, not even remaking the wires helps reliably, thus it's possible that local updates fail at times. Whether or not this is specific to z-wires should be confirmed. * Possibly #2520 . Needs confirmation. @Lohikar since this is something you've worked on, could you investigate during the week?
1.0
SMC: Mapped in power connections do not appear to be updated properly - Okay. Long story short. We've been running into issues in new map meme where connections that are valid, and exist since round start, are not flagged properly until an admin invokes `make-powernets`. Example cases of this: * Tesla setup. This is all mapped in to be 100% functional (as far as wiring goes). However, after starting the engine, power does not flow until `make-powernets` or a local powernet update (I presume, haven't tried) is invoked. * #2345 . Where mapped in z-wires do not move power until `make-powernets` is invoked. In this case, not even remaking the wires helps reliably, thus it's possible that local updates fail at times. Whether or not this is specific to z-wires should be confirmed. * Possibly #2520 . Needs confirmation. @Lohikar since this is something you've worked on, could you investigate during the week?
priority
smc mapped in power connections do not appear to be updated properly okay long story short we ve been running into issues in new map meme where connections that are valid and exist since round start are not flagged properly until an admin invokes make powernets example cases of this tesla setup this is all mapped in to be functional as far as wiring goes however after starting the engine power does not flow until make powernets or a local powernet update i presume haven t tried is invoked where mapped in z wires do not move power until make powernets is invoked in this case not even remaking the wires helps reliably thus it s possible that local updates fail at times whether or not this is specific to z wires should be confirmed possibly needs confirmation lohikar since this is something you ve worked on could you investigate during the week
1
272,410
8,508,152,747
IssuesEvent
2018-10-30 21:06:11
isawnyu/isaw.web
https://api.github.com/repos/isawnyu/isaw.web
opened
Unable to edit images
bug high priority
I'm getting a blank dialog box this afternoon when I attempt to edit an image. It looks like this: <img width="985" alt="screen shot 2018-10-30 at 4 57 29 pm" src="https://user-images.githubusercontent.com/7882896/47750546-00067c00-dc66-11e8-8be3-ac5886561082.png"> I am using Firefox (Mac).
1.0
Unable to edit images - I'm getting a blank dialog box this afternoon when I attempt to edit an image. It looks like this: <img width="985" alt="screen shot 2018-10-30 at 4 57 29 pm" src="https://user-images.githubusercontent.com/7882896/47750546-00067c00-dc66-11e8-8be3-ac5886561082.png"> I am using Firefox (Mac).
priority
unable to edit images i m getting a blank dialog box this afternoon when i attempt to edit an image it looks like this img width alt screen shot at pm src i am using firefox mac
1
739,995
25,731,498,609
IssuesEvent
2022-12-07 20:41:38
fecgov/fec-cms
https://api.github.com/repos/fecgov/fec-cms
closed
Update meeting page template to change default open meeting and executive session time
Work: Front-end High priority
### Summary **What we're after:** _OCS has indicated that meeting times will shift to start meetings at 10:30. We need to update the default time in the meeting page template so that users interested in Commission meetings have accurate information._ ### Completion criteria - [ ] Default time is changed on the meeting page template to 10:30.
1.0
Update meeting page template to change default open meeting and executive session time - ### Summary **What we're after:** _OCS has indicated that meeting times will shift to start meetings at 10:30. We need to update the default time in the meeting page template so that users interested in Commission meetings have accurate information._ ### Completion criteria - [ ] Default time is changed on the meeting page template to 10:30.
priority
update meeting page template to change default open meeting and executive session time summary what we re after ocs has indicated that meeting times will shift to start meetings at we need to update the default time in the meeting page template so that users interested in commission meetings have accurate information completion criteria default time is changed on the meeting page template to
1
740,879
25,771,981,009
IssuesEvent
2022-12-09 08:48:15
sorrowcode/taesch-
https://api.github.com/repos/sorrowcode/taesch-
closed
feature - Verbindung NearShopsScreen und MapScreen
USP - 5 feature 0 - highest priority
Als Nutzer möchte ich, dass ich beim Aufrufen eines Ladens eine Möglichkeit habe, dass ich zur Karte geführt werde, damit ich sehen kann, wo genau der Laden in meiner Nähe sich befindet. Hier soll eine Strategie ausgearbeitet und implementiert werden, wodurch die Möglichkeit besteht, dynamisch zwischen NearShopsScreen und MapScreen zu navigieren und möglicherweise einzelne Läden anzeigen zu lassen.
1.0
feature - Verbindung NearShopsScreen und MapScreen - Als Nutzer möchte ich, dass ich beim Aufrufen eines Ladens eine Möglichkeit habe, dass ich zur Karte geführt werde, damit ich sehen kann, wo genau der Laden in meiner Nähe sich befindet. Hier soll eine Strategie ausgearbeitet und implementiert werden, wodurch die Möglichkeit besteht, dynamisch zwischen NearShopsScreen und MapScreen zu navigieren und möglicherweise einzelne Läden anzeigen zu lassen.
priority
feature verbindung nearshopsscreen und mapscreen als nutzer möchte ich dass ich beim aufrufen eines ladens eine möglichkeit habe dass ich zur karte geführt werde damit ich sehen kann wo genau der laden in meiner nähe sich befindet hier soll eine strategie ausgearbeitet und implementiert werden wodurch die möglichkeit besteht dynamisch zwischen nearshopsscreen und mapscreen zu navigieren und möglicherweise einzelne läden anzeigen zu lassen
1
586,988
17,601,238,335
IssuesEvent
2021-08-17 12:08:16
gitpod-io/gitpod
https://api.github.com/repos/gitpod-io/gitpod
opened
Agent Smith signature check fails because of permission issues
type: bug component: agent-smith priority: highest (user impact)
### Bug description Agent Smith cannot check signatures of running processes at all because of permission issues. All violations that are found are found because if blacklisted commands, not exec signatures. Log [link](https://cloudlogging.app.goo.gl/KcAc7153HjjXA6uE8) Example log output: ``` {"@type":"type.googleapis.com/google.devtools.clouderrorreporting.v1beta1.ReportedErrorEvent","error":"open /proc/560401/exe: permission denied","level":"warning","message":"cannot open executable to check signatures","path":"id","serviceContext":{"service":"agent-smith","version":""},"severity":"WARNING","time":"2021-08-17T10:59:18Z"} ``` ### Steps to reproduce - Create a preview environment - start a workspace - check log of (relevant) agent smith daemon ### Expected behavior Agent smith should be able to check signatures of all processes running inside a workspace. ### Example repository _No response_ ### Anything else? _No response_
1.0
Agent Smith signature check fails because of permission issues - ### Bug description Agent Smith cannot check signatures of running processes at all because of permission issues. All violations that are found are found because if blacklisted commands, not exec signatures. Log [link](https://cloudlogging.app.goo.gl/KcAc7153HjjXA6uE8) Example log output: ``` {"@type":"type.googleapis.com/google.devtools.clouderrorreporting.v1beta1.ReportedErrorEvent","error":"open /proc/560401/exe: permission denied","level":"warning","message":"cannot open executable to check signatures","path":"id","serviceContext":{"service":"agent-smith","version":""},"severity":"WARNING","time":"2021-08-17T10:59:18Z"} ``` ### Steps to reproduce - Create a preview environment - start a workspace - check log of (relevant) agent smith daemon ### Expected behavior Agent smith should be able to check signatures of all processes running inside a workspace. ### Example repository _No response_ ### Anything else? _No response_
priority
agent smith signature check fails because of permission issues bug description agent smith cannot check signatures of running processes at all because of permission issues all violations that are found are found because if blacklisted commands not exec signatures log example log output type type googleapis com google devtools clouderrorreporting reportederrorevent error open proc exe permission denied level warning message cannot open executable to check signatures path id servicecontext service agent smith version severity warning time steps to reproduce create a preview environment start a workspace check log of relevant agent smith daemon expected behavior agent smith should be able to check signatures of all processes running inside a workspace example repository no response anything else no response
1
800,379
28,363,711,686
IssuesEvent
2023-04-12 12:36:37
flowforge/flowforge
https://api.github.com/repos/flowforge/flowforge
closed
Cancel button application/settings edit page doesn't work
bug area:frontend priority:high
### Current Behavior ``` ntime-core.esm-bundler.js:169 [Vue warn]: Property "cancelEditName" was accessed during render but is not defined on instance. at <ProjectSettings application= Object instances= Array(1) is-visiting-admin=false ... > at <RouterView application= Object instances= Array(1) is-visiting-admin=false ... > at <ProjectPage onVnodeUnmounted=fn<onVnodeUnmounted> ref=Ref< Proxy(Object) > > at <RouterView> at <FfLayoutPlatform key=0 > at <App> w ``` ### Expected Behavior _No response_ ### Steps To Reproduce _No response_ ### Environment - FlowForge version: - Node.js version: - npm version: - Platform/OS: - Browser:
1.0
Cancel button application/settings edit page doesn't work - ### Current Behavior ``` ntime-core.esm-bundler.js:169 [Vue warn]: Property "cancelEditName" was accessed during render but is not defined on instance. at <ProjectSettings application= Object instances= Array(1) is-visiting-admin=false ... > at <RouterView application= Object instances= Array(1) is-visiting-admin=false ... > at <ProjectPage onVnodeUnmounted=fn<onVnodeUnmounted> ref=Ref< Proxy(Object) > > at <RouterView> at <FfLayoutPlatform key=0 > at <App> w ``` ### Expected Behavior _No response_ ### Steps To Reproduce _No response_ ### Environment - FlowForge version: - Node.js version: - npm version: - Platform/OS: - Browser:
priority
cancel button application settings edit page doesn t work current behavior ntime core esm bundler js property canceleditname was accessed during render but is not defined on instance at at at ref ref at at at w expected behavior no response steps to reproduce no response environment flowforge version node js version npm version platform os browser
1
772,523
27,125,714,699
IssuesEvent
2023-02-16 05:01:02
FastcampusMini/mini-project
https://api.github.com/repos/FastcampusMini/mini-project
closed
Basket API
For: API Priority: High Status: In Progress Type: Feature
## Title 장바구니 기능 ## Description 장바구니 ## Tasks - [x] 장바구니 repo, service, dto 구현
1.0
Basket API - ## Title 장바구니 기능 ## Description 장바구니 ## Tasks - [x] 장바구니 repo, service, dto 구현
priority
basket api title 장바구니 기능 description 장바구니 tasks 장바구니 repo service dto 구현
1
777,656
27,289,606,690
IssuesEvent
2023-02-23 15:43:58
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
PFacct issue in cluster
Type: Bug Priority: High
PacketFence 12.2. Switching from standalone to cluster when creating a cluster, the load balancer does not want to start because PF acct is already listening.
1.0
PFacct issue in cluster - PacketFence 12.2. Switching from standalone to cluster when creating a cluster, the load balancer does not want to start because PF acct is already listening.
priority
pfacct issue in cluster packetfence switching from standalone to cluster when creating a cluster the load balancer does not want to start because pf acct is already listening
1
329,670
10,022,953,619
IssuesEvent
2019-07-16 17:58:54
ArtskydJ/comicsrss.com
https://api.github.com/repos/ArtskydJ/comicsrss.com
closed
Some gocomics strips are not showing up
high-priority
From @megadr01d, originally posted in #86 Hope you don't mind... I removed your other comment since I see these as two separate issues. ----------------------------------- Why are some GoComics comics are not available? Like these two, which I'd love to see in the future: https://www.gocomics.com/how-to-cat https://www.gocomics.com/redmeat
1.0
Some gocomics strips are not showing up - From @megadr01d, originally posted in #86 Hope you don't mind... I removed your other comment since I see these as two separate issues. ----------------------------------- Why are some GoComics comics are not available? Like these two, which I'd love to see in the future: https://www.gocomics.com/how-to-cat https://www.gocomics.com/redmeat
priority
some gocomics strips are not showing up from originally posted in hope you don t mind i removed your other comment since i see these as two separate issues why are some gocomics comics are not available like these two which i d love to see in the future
1
530,309
15,420,784,113
IssuesEvent
2021-03-05 12:05:53
apluslms/a-plus
https://api.github.com/repos/apluslms/a-plus
closed
Latest course instance redirection must pick the latest visible instance
O1 needs area: UX student effort: hours experience: good first issue priority: high requester: Aalto teacher type: bug
Pull request #772 added the redirection view to the latest course instance: https://plus.cs.aalto.fi/o1/ However, if the latest instance is hidden from students, then this redirection becomes useless and confusing to students. It is common that the latest instance is hidden before it has begun because it is still work in progress. The redirection should always pick the latest VISIBLE instance.
1.0
Latest course instance redirection must pick the latest visible instance - Pull request #772 added the redirection view to the latest course instance: https://plus.cs.aalto.fi/o1/ However, if the latest instance is hidden from students, then this redirection becomes useless and confusing to students. It is common that the latest instance is hidden before it has begun because it is still work in progress. The redirection should always pick the latest VISIBLE instance.
priority
latest course instance redirection must pick the latest visible instance pull request added the redirection view to the latest course instance however if the latest instance is hidden from students then this redirection becomes useless and confusing to students it is common that the latest instance is hidden before it has begun because it is still work in progress the redirection should always pick the latest visible instance
1
566,102
16,796,056,019
IssuesEvent
2021-06-16 03:49:18
parallel-finance/parallel
https://api.github.com/repos/parallel-finance/parallel
closed
tokio-runtime-worker runtime::timestamp: `pallet_timestamp::UnixTime::now` is called at genesis, invalid value returned: 0
high priority
<img width="1297" alt="image" src="https://user-images.githubusercontent.com/33961674/121112081-81711280-c842-11eb-9f7d-3391524261f4.png">
1.0
tokio-runtime-worker runtime::timestamp: `pallet_timestamp::UnixTime::now` is called at genesis, invalid value returned: 0 - <img width="1297" alt="image" src="https://user-images.githubusercontent.com/33961674/121112081-81711280-c842-11eb-9f7d-3391524261f4.png">
priority
tokio runtime worker runtime timestamp pallet timestamp unixtime now is called at genesis invalid value returned img width alt image src
1
138,583
5,344,628,530
IssuesEvent
2017-02-17 15:01:22
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Feature Request: torch 'module' object has no attribute '__version__'
enhancement high priority
could you please add `__version__` to the `torch` module! all the cool modules have it! ;)
1.0
Feature Request: torch 'module' object has no attribute '__version__' - could you please add `__version__` to the `torch` module! all the cool modules have it! ;)
priority
feature request torch module object has no attribute version could you please add version to the torch module all the cool modules have it
1
761,872
26,700,578,959
IssuesEvent
2023-01-27 14:05:46
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Error when Renaming Workspaces with Plots
High Priority Bug
**Describe the bug** Workbench throws errors when a workspace is renamed with a plot of its data open. **To Reproduce** 1. Load data 2. Plot a spectrum 3. Rename the workspace **Expected behaviour** Title of the plot is changed and no errors are thrown. **Platform/Version (please complete the following information):** - OS: All - Mantid Version: 6.6 Nightly (2023-01-24) **Additional context** ``` Error occurred in handler: Traceback (most recent call last): File "/home/conor/repos/mantid/qt/applications/workbench/workbench/plotting/figuremanager.py", line 62, in wrapper func(*args, **kwargs) File "/home/conor/repos/mantid/qt/applications/workbench/workbench/plotting/figuremanager.py", line 162, in renameHandle self.canvas.manager.set_window_title(_replace_workspace_name_in_string(oldName, newName, self.canvas.get_window_title())) AttributeError: 'MantidFigureCanvas' object has no attribute 'get_window_title' ```
1.0
Error when Renaming Workspaces with Plots - **Describe the bug** Workbench throws errors when a workspace is renamed with a plot of its data open. **To Reproduce** 1. Load data 2. Plot a spectrum 3. Rename the workspace **Expected behaviour** Title of the plot is changed and no errors are thrown. **Platform/Version (please complete the following information):** - OS: All - Mantid Version: 6.6 Nightly (2023-01-24) **Additional context** ``` Error occurred in handler: Traceback (most recent call last): File "/home/conor/repos/mantid/qt/applications/workbench/workbench/plotting/figuremanager.py", line 62, in wrapper func(*args, **kwargs) File "/home/conor/repos/mantid/qt/applications/workbench/workbench/plotting/figuremanager.py", line 162, in renameHandle self.canvas.manager.set_window_title(_replace_workspace_name_in_string(oldName, newName, self.canvas.get_window_title())) AttributeError: 'MantidFigureCanvas' object has no attribute 'get_window_title' ```
priority
error when renaming workspaces with plots describe the bug workbench throws errors when a workspace is renamed with a plot of its data open to reproduce load data plot a spectrum rename the workspace expected behaviour title of the plot is changed and no errors are thrown platform version please complete the following information os all mantid version nightly additional context error occurred in handler traceback most recent call last file home conor repos mantid qt applications workbench workbench plotting figuremanager py line in wrapper func args kwargs file home conor repos mantid qt applications workbench workbench plotting figuremanager py line in renamehandle self canvas manager set window title replace workspace name in string oldname newname self canvas get window title attributeerror mantidfigurecanvas object has no attribute get window title
1
829,179
31,857,819,731
IssuesEvent
2023-09-15 08:45:36
rcpch/rcpch-audit-engine
https://api.github.com/repos/rcpch/rcpch-audit-engine
closed
Current code commit/version should be linked in topmost nav (left)
before-launch feature request priority: high
For ease of administrator/developer debugging, it would be a nice-to-have if the current commit hash of the running version of the software was visible in the top left nav. * [ ] Should show **short** commit hash of the current version * [ ] Should only be shown to Admins and Developers * [ ] Should click through to the commit on Github for quickly viewing the relevant code * [ ] Version number +/- Git tag +/- Github Release should also be displayed once we have semver established Sketch of what this would look like ![Image](https://user-images.githubusercontent.com/2348385/231704602-3108b7e8-7705-4874-a149-814e92bfa592.png)
1.0
Current code commit/version should be linked in topmost nav (left) - For ease of administrator/developer debugging, it would be a nice-to-have if the current commit hash of the running version of the software was visible in the top left nav. * [ ] Should show **short** commit hash of the current version * [ ] Should only be shown to Admins and Developers * [ ] Should click through to the commit on Github for quickly viewing the relevant code * [ ] Version number +/- Git tag +/- Github Release should also be displayed once we have semver established Sketch of what this would look like ![Image](https://user-images.githubusercontent.com/2348385/231704602-3108b7e8-7705-4874-a149-814e92bfa592.png)
priority
current code commit version should be linked in topmost nav left for ease of administrator developer debugging it would be a nice to have if the current commit hash of the running version of the software was visible in the top left nav should show short commit hash of the current version should only be shown to admins and developers should click through to the commit on github for quickly viewing the relevant code version number git tag github release should also be displayed once we have semver established sketch of what this would look like
1
472,092
13,616,165,547
IssuesEvent
2020-09-23 15:17:51
OpenLightingProject/ola
https://api.github.com/repos/OpenLightingProject/ola
closed
New warnings with GCC9
Component-Plugin Difficulty-Easy Language-C++ Maintainability OpSys-Linux Priority-High bug
I received [Debian bug 925793](https://bugs.debian.org/925793), which produces a few new warnings of this type: ``` libs/acn/HeaderSetTest.cpp: In member function 'void HeaderSetTest::testTransportHeader()': libs/acn/HeaderSetTest.cpp:78:29: error: implicitly-declared 'ola::acn::TransportHeader::TransportHeader(const ola::acn::TransportHeader&)' is deprecated [-Werror=deprecated-copy] 78 | TransportHeader header2 = header; | ^~~~~~ In file included from ./libs/acn/HeaderSet.h:30, from libs/acn/HeaderSetTest.cpp:28: ../libs/acn/TransportHeader.h:58:8: note: because 'ola::acn::TransportHeader' has user-provided 'void ola::acn::TransportHeader::operator=(const ola::acn::TransportHeader&)' 58 | void operator=(const TransportHeader &other) { | ^~~~~~~~ ``` Basically, what happens here is that the compiler complains about a class which does have an assignment operator, but no copy constructor. You should probably add them.
1.0
New warnings with GCC9 - I received [Debian bug 925793](https://bugs.debian.org/925793), which produces a few new warnings of this type: ``` libs/acn/HeaderSetTest.cpp: In member function 'void HeaderSetTest::testTransportHeader()': libs/acn/HeaderSetTest.cpp:78:29: error: implicitly-declared 'ola::acn::TransportHeader::TransportHeader(const ola::acn::TransportHeader&)' is deprecated [-Werror=deprecated-copy] 78 | TransportHeader header2 = header; | ^~~~~~ In file included from ./libs/acn/HeaderSet.h:30, from libs/acn/HeaderSetTest.cpp:28: ../libs/acn/TransportHeader.h:58:8: note: because 'ola::acn::TransportHeader' has user-provided 'void ola::acn::TransportHeader::operator=(const ola::acn::TransportHeader&)' 58 | void operator=(const TransportHeader &other) { | ^~~~~~~~ ``` Basically, what happens here is that the compiler complains about a class which does have an assignment operator, but no copy constructor. You should probably add them.
priority
new warnings with i received which produces a few new warnings of this type libs acn headersettest cpp in member function void headersettest testtransportheader libs acn headersettest cpp error implicitly declared ola acn transportheader transportheader const ola acn transportheader is deprecated transportheader header in file included from libs acn headerset h from libs acn headersettest cpp libs acn transportheader h note because ola acn transportheader has user provided void ola acn transportheader operator const ola acn transportheader void operator const transportheader other basically what happens here is that the compiler complains about a class which does have an assignment operator but no copy constructor you should probably add them
1
286,709
8,792,000,176
IssuesEvent
2018-12-21 14:43:02
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
opened
support symbolic random functions (Vector<Expression>)
priority: high team: automotive type: feature request
Unlike the existing Vector<Expression>, where each element can always be evaluated separately, I think in the random case I need to call `Evaluate(vector<Expression>, generator)` and have the same random value used for all of the outputs? in other words... if I have ``` v = Variable(RANDOM_UNIFORM) y1 = 5*v + 1 y2 = 3*v + 3 y = Eigen::Vector2<Expression>(y1, y2) ``` i would like to be able to call `Evaluate(y1, generator)` and `Evaluate(y2, generator)` separately, which would result in independent draws from the distribution of `v`. but I would also like to be able to call `Evaluate(y, generator)`, and have `v` drawn only once, but used for both `y1` and `y2`.
1.0
support symbolic random functions (Vector<Expression>) - Unlike the existing Vector<Expression>, where each element can always be evaluated separately, I think in the random case I need to call `Evaluate(vector<Expression>, generator)` and have the same random value used for all of the outputs? in other words... if I have ``` v = Variable(RANDOM_UNIFORM) y1 = 5*v + 1 y2 = 3*v + 3 y = Eigen::Vector2<Expression>(y1, y2) ``` i would like to be able to call `Evaluate(y1, generator)` and `Evaluate(y2, generator)` separately, which would result in independent draws from the distribution of `v`. but I would also like to be able to call `Evaluate(y, generator)`, and have `v` drawn only once, but used for both `y1` and `y2`.
priority
support symbolic random functions vector unlike the existing vector where each element can always be evaluated separately i think in the random case i need to call evaluate vector generator and have the same random value used for all of the outputs in other words if i have v variable random uniform v v y eigen i would like to be able to call evaluate generator and evaluate generator separately which would result in independent draws from the distribution of v but i would also like to be able to call evaluate y generator and have v drawn only once but used for both and
1
423,427
12,296,028,115
IssuesEvent
2020-05-11 05:57:13
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
color.firefox.com - see bug description
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal type-webrender-enabled
<!-- @browser: Firefox 78.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; rv:78.0) Gecko/20100101 Firefox/78.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52670 --> <!-- @extra_labels: type-webrender-enabled --> **URL**: https://color.firefox.com/ **Browser / Version**: Firefox 78.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Something else **Description**: Background seen below the toolbar **Steps to Reproduce**: Just noticed that there's a line (maybe one pixel high) below my toolbar (on every website) when setting a background image with a different color than my toolbar is. <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/5/435ee96a-d92d-456d-b553-c6a057259ed9.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: true</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200509091423</li><li>channel: nightly</li><li>GPUs: <ul> <li>active: true</li><li>description: Intel(R) Iris(R) Graphics 540</li><li>deviceID: 0x1926</li><li>vendorID: 0x8086</li><li>driverVersion: 26.20.100.7529</li> </ul></li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/5/b66ac394-49ab-48be-b5f9-fe79cfddd30c) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
color.firefox.com - see bug description - <!-- @browser: Firefox 78.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; rv:78.0) Gecko/20100101 Firefox/78.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52670 --> <!-- @extra_labels: type-webrender-enabled --> **URL**: https://color.firefox.com/ **Browser / Version**: Firefox 78.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Something else **Description**: Background seen below the toolbar **Steps to Reproduce**: Just noticed that there's a line (maybe one pixel high) below my toolbar (on every website) when setting a background image with a different color than my toolbar is. <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/5/435ee96a-d92d-456d-b553-c6a057259ed9.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: true</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200509091423</li><li>channel: nightly</li><li>GPUs: <ul> <li>active: true</li><li>description: Intel(R) Iris(R) Graphics 540</li><li>deviceID: 0x1926</li><li>vendorID: 0x8086</li><li>driverVersion: 26.20.100.7529</li> </ul></li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/5/b66ac394-49ab-48be-b5f9-fe79cfddd30c) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
color firefox com see bug description url browser version firefox operating system windows tested another browser no problem type something else description background seen below the toolbar steps to reproduce just noticed that there s a line maybe one pixel high below my toolbar on every website when setting a background image with a different color than my toolbar is view the screenshot img alt screenshot src browser configuration gfx webrender all true gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly gpus active true description intel r iris r graphics deviceid vendorid driverversion hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
830,331
32,002,753,395
IssuesEvent
2023-09-21 13:14:17
electric-sql/electric
https://api.github.com/repos/electric-sql/electric
closed
[VAX-1004] Supported dev OSes
bug linear High priority
Given the wide usage of Ubuntu 22.04, it would seem a reasonable dev system to work with. Are you guys all working/testing on MacOS or something? Several steps in the quickstart just seem to fail, and trying to hack around them requires installing and sourcing an `emsdk`, installing `node_modules` with `PUPPETEER_SKIP_DOWNLOAD=true npm i @electric-sql/prisma-generator`, doing `yarn install`, etc. - all things not listed at all in the quickstart. I realise things are moving fast, but a quick test in a fresh Ubuntu 22.04 VM for the quickstart seems pretty reasonable to me... The issue being, it is very much not "quick" to get started, and wouldn't be possible for someone who hasn't spent a lot of time trying to get things working. <sub>[VAX-1004](https://linear.app/electric-sql/issue/VAX-1004/supported-dev-oses)</sub>
1.0
[VAX-1004] Supported dev OSes - Given the wide usage of Ubuntu 22.04, it would seem a reasonable dev system to work with. Are you guys all working/testing on MacOS or something? Several steps in the quickstart just seem to fail, and trying to hack around them requires installing and sourcing an `emsdk`, installing `node_modules` with `PUPPETEER_SKIP_DOWNLOAD=true npm i @electric-sql/prisma-generator`, doing `yarn install`, etc. - all things not listed at all in the quickstart. I realise things are moving fast, but a quick test in a fresh Ubuntu 22.04 VM for the quickstart seems pretty reasonable to me... The issue being, it is very much not "quick" to get started, and wouldn't be possible for someone who hasn't spent a lot of time trying to get things working. <sub>[VAX-1004](https://linear.app/electric-sql/issue/VAX-1004/supported-dev-oses)</sub>
priority
supported dev oses given the wide usage of ubuntu it would seem a reasonable dev system to work with are you guys all working testing on macos or something several steps in the quickstart just seem to fail and trying to hack around them requires installing and sourcing an emsdk installing node modules with puppeteer skip download true npm i electric sql prisma generator doing yarn install etc all things not listed at all in the quickstart i realise things are moving fast but a quick test in a fresh ubuntu vm for the quickstart seems pretty reasonable to me the issue being it is very much not quick to get started and wouldn t be possible for someone who hasn t spent a lot of time trying to get things working
1
451,270
13,032,575,573
IssuesEvent
2020-07-28 04:39:42
rica441/takutaku
https://api.github.com/repos/rica441/takutaku
closed
かんたんログイン機能
feature high priority
## なにをやるのか - ユーザー情報を登録せずにゲストユーザーとしてログインする ## なぜそれをやるのか - いろんな人にtekutekuを使ってみて欲しいから - フィードバックをもらいたい ## 参考リンク - https://www.isoroot.jp/blog/2451/
1.0
かんたんログイン機能 - ## なにをやるのか - ユーザー情報を登録せずにゲストユーザーとしてログインする ## なぜそれをやるのか - いろんな人にtekutekuを使ってみて欲しいから - フィードバックをもらいたい ## 参考リンク - https://www.isoroot.jp/blog/2451/
priority
かんたんログイン機能 なにをやるのか ユーザー情報を登録せずにゲストユーザーとしてログインする なぜそれをやるのか いろんな人にtekutekuを使ってみて欲しいから フィードバックをもらいたい 参考リンク
1
552,680
16,246,837,957
IssuesEvent
2021-05-07 15:23:37
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YCQL] Index not chosen based on ordering and certain ordering allowed that shouldn't be
priority/high
Instead of optimizing for the ordering an index is chosen only based on the criteria which means it's impossible to make some queries. Additionally, queries that involve the order of the primary index are allowed despite using a secondary index. Here's an example schema: ```cql CREATE KEYSPACE test; USE test; CREATE TABLE test ( id int, scope text, key text, value text, PRIMARY KEY (id) ) WITH default_time_to_live = 0 AND transactions = {'enabled': 'true'}; CREATE INDEX test_index_1 ON test (scope, key, value) WITH CLUSTERING ORDER BY (key ASC, value ASC) AND transactions = {'enabled': 'true'}; CREATE INDEX test_index_2 ON test (scope, key, id) WITH CLUSTERING ORDER BY (key ASC, id ASC) AND transactions = {'enabled': 'true'}; INSERT into test (id, scope, key, value) VALUES (4, 'test', 'bar', 'baz'); INSERT into test (id, scope, key, value) VALUES (3, 'test', 'foo', 'baz'); INSERT into test (id, scope, key, value) VALUES (2, 'test', 'foo', 'bar'); INSERT into test (id, scope, key, value) VALUES (1, 'test', 'foo', 'bar'); ``` Now let's say you want to paginate over all of the "foo" `key` and "bar" `value`. You need to order by id because `value` is mutable and could cause the pagination to duplicate/change as your paginating. So your query is: ``` EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id LIMIT 1; InvalidRequest: Error from server: code=2200 [Invalid query] message="Invalid Arguments. Order by currently only support the ordering of columns following their declared order in the PRIMARY KEY EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id LIMIT 1; ^^^ (ql error -304)" ``` Which errors. This is because it chooses `test_index_1` instead of choosing `test_index_2`. If we add `value` to the end of the order by, however: ``` EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id, value LIMIT 1; QUERY PLAN ------------------------------------------------------ Index Only Scan using test.test_index_1 on test.test Key Conditions: (scope = 'test') Filter: (key = 'foo') AND (value = 'bar') ``` Then doesn't error but it doesn't follow the order in `test_index_1 (key, value, id)`. The `id` at the end is assumed based on the docs saying: > Any primary key column of the table not indexed explicitly in `index_columns` is added as a clustering column to the index implicitly.
1.0
[YCQL] Index not chosen based on ordering and certain ordering allowed that shouldn't be - Instead of optimizing for the ordering an index is chosen only based on the criteria which means it's impossible to make some queries. Additionally, queries that involve the order of the primary index are allowed despite using a secondary index. Here's an example schema: ```cql CREATE KEYSPACE test; USE test; CREATE TABLE test ( id int, scope text, key text, value text, PRIMARY KEY (id) ) WITH default_time_to_live = 0 AND transactions = {'enabled': 'true'}; CREATE INDEX test_index_1 ON test (scope, key, value) WITH CLUSTERING ORDER BY (key ASC, value ASC) AND transactions = {'enabled': 'true'}; CREATE INDEX test_index_2 ON test (scope, key, id) WITH CLUSTERING ORDER BY (key ASC, id ASC) AND transactions = {'enabled': 'true'}; INSERT into test (id, scope, key, value) VALUES (4, 'test', 'bar', 'baz'); INSERT into test (id, scope, key, value) VALUES (3, 'test', 'foo', 'baz'); INSERT into test (id, scope, key, value) VALUES (2, 'test', 'foo', 'bar'); INSERT into test (id, scope, key, value) VALUES (1, 'test', 'foo', 'bar'); ``` Now let's say you want to paginate over all of the "foo" `key` and "bar" `value`. You need to order by id because `value` is mutable and could cause the pagination to duplicate/change as your paginating. So your query is: ``` EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id LIMIT 1; InvalidRequest: Error from server: code=2200 [Invalid query] message="Invalid Arguments. Order by currently only support the ordering of columns following their declared order in the PRIMARY KEY EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id LIMIT 1; ^^^ (ql error -304)" ``` Which errors. This is because it chooses `test_index_1` instead of choosing `test_index_2`. If we add `value` to the end of the order by, however: ``` EXPLAIN SELECT id FROM test WHERE scope = 'test' AND key = 'foo' AND value = 'bar' ORDER BY key, id, value LIMIT 1; QUERY PLAN ------------------------------------------------------ Index Only Scan using test.test_index_1 on test.test Key Conditions: (scope = 'test') Filter: (key = 'foo') AND (value = 'bar') ``` Then doesn't error but it doesn't follow the order in `test_index_1 (key, value, id)`. The `id` at the end is assumed based on the docs saying: > Any primary key column of the table not indexed explicitly in `index_columns` is added as a clustering column to the index implicitly.
priority
index not chosen based on ordering and certain ordering allowed that shouldn t be instead of optimizing for the ordering an index is chosen only based on the criteria which means it s impossible to make some queries additionally queries that involve the order of the primary index are allowed despite using a secondary index here s an example schema cql create keyspace test use test create table test id int scope text key text value text primary key id with default time to live and transactions enabled true create index test index on test scope key value with clustering order by key asc value asc and transactions enabled true create index test index on test scope key id with clustering order by key asc id asc and transactions enabled true insert into test id scope key value values test bar baz insert into test id scope key value values test foo baz insert into test id scope key value values test foo bar insert into test id scope key value values test foo bar now let s say you want to paginate over all of the foo key and bar value you need to order by id because value is mutable and could cause the pagination to duplicate change as your paginating so your query is explain select id from test where scope test and key foo and value bar order by key id limit invalidrequest error from server code message invalid arguments order by currently only support the ordering of columns following their declared order in the primary key explain select id from test where scope test and key foo and value bar order by key id limit ql error which errors this is because it chooses test index instead of choosing test index if we add value to the end of the order by however explain select id from test where scope test and key foo and value bar order by key id value limit query plan index only scan using test test index on test test key conditions scope test filter key foo and value bar then doesn t error but it doesn t follow the order in test index key value id the id at the end is assumed based on the docs saying any primary key column of the table not indexed explicitly in index columns is added as a clustering column to the index implicitly
1
598,144
18,237,879,504
IssuesEvent
2021-10-01 09:14:46
AbsaOSS/enceladus
https://api.github.com/repos/AbsaOSS/enceladus
opened
Add the possibility to limit the output files sizes
feature Conformance under discussion Standardization priority: high
## Background Sometimes the output files are too big and other applications have troubles operating with them. ## Feature Make it possible to specify the maximum size of output parquet files. ## Proposed Solution Solution Ideas: 1. Have an optional setting for maximal file size. 2. If the limit would be overshot, repartition. 3. This should be done both for _Standardization_ and _Conformance_.
1.0
Add the possibility to limit the output files sizes - ## Background Sometimes the output files are too big and other applications have troubles operating with them. ## Feature Make it possible to specify the maximum size of output parquet files. ## Proposed Solution Solution Ideas: 1. Have an optional setting for maximal file size. 2. If the limit would be overshot, repartition. 3. This should be done both for _Standardization_ and _Conformance_.
priority
add the possibility to limit the output files sizes background sometimes the output files are too big and other applications have troubles operating with them feature make it possible to specify the maximum size of output parquet files proposed solution solution ideas have an optional setting for maximal file size if the limit would be overshot repartition this should be done both for standardization and conformance
1
439,719
12,685,663,232
IssuesEvent
2020-06-20 06:01:37
sodafoundation/dashboard
https://api.github.com/repos/sodafoundation/dashboard
closed
Dashboard still shows OpenSDS Logo and references to OpenSDS
High Priority enhancement
**Issue/Feature Description:** Change OpenSDS logo to SODA and all occurrences of OpenSDS to SODA **Why this issue to fixed / feature is needed(give scenarios or use cases):** Since the project is now renamed to SODA all occurrences of the OpenSDS logo and OpenSDS title have to be changed to SODA. The main places that the changes have to be made are: - Login Page Logo - Home Page Menu Logo - Browser Tab title - Favicon
1.0
Dashboard still shows OpenSDS Logo and references to OpenSDS - **Issue/Feature Description:** Change OpenSDS logo to SODA and all occurrences of OpenSDS to SODA **Why this issue to fixed / feature is needed(give scenarios or use cases):** Since the project is now renamed to SODA all occurrences of the OpenSDS logo and OpenSDS title have to be changed to SODA. The main places that the changes have to be made are: - Login Page Logo - Home Page Menu Logo - Browser Tab title - Favicon
priority
dashboard still shows opensds logo and references to opensds issue feature description change opensds logo to soda and all occurrences of opensds to soda why this issue to fixed feature is needed give scenarios or use cases since the project is now renamed to soda all occurrences of the opensds logo and opensds title have to be changed to soda the main places that the changes have to be made are login page logo home page menu logo browser tab title favicon
1
262,284
8,263,191,459
IssuesEvent
2018-09-14 00:50:23
ampproject/amphtml
https://api.github.com/repos/ampproject/amphtml
closed
Cleanup amp-image-slider loading issue
Category: Presentation P1: High Priority Type: Feature Request
Currently, there is no code that handles loading for `amp-image-slider`. By default, the slider should construct its structure and load the images normally, which shows the placeholder and the loading 3 dots of `amp-img`. This is actually one of the proposal: __Proposal A__: leave as is. (LEFT is the placeholder for left image. The following are under "Fast 3G" so should be much faster on actual devices) ![loading-nochange](https://user-images.githubusercontent.com/15007517/43985835-74bc009c-9cbf-11e8-86ab-3b69323646bf.gif) However, due to ownership related issues, the 3 dots would only show when `layer` experiment is enabled. Otherwise, the loading process of images is never displayed. Another proposal (__Proposal B__) to the loading state is to whitelist the slider itself for loading and display 3 dots of the slider, which would mask on top of the loading images until `LOAD_END` is signaled. This is attempted in #17420 ![loading-proposal](https://user-images.githubusercontent.com/15007517/43985811-333f57cc-9cbf-11e8-9ce7-fb6657f39575.gif)
1.0
Cleanup amp-image-slider loading issue - Currently, there is no code that handles loading for `amp-image-slider`. By default, the slider should construct its structure and load the images normally, which shows the placeholder and the loading 3 dots of `amp-img`. This is actually one of the proposal: __Proposal A__: leave as is. (LEFT is the placeholder for left image. The following are under "Fast 3G" so should be much faster on actual devices) ![loading-nochange](https://user-images.githubusercontent.com/15007517/43985835-74bc009c-9cbf-11e8-86ab-3b69323646bf.gif) However, due to ownership related issues, the 3 dots would only show when `layer` experiment is enabled. Otherwise, the loading process of images is never displayed. Another proposal (__Proposal B__) to the loading state is to whitelist the slider itself for loading and display 3 dots of the slider, which would mask on top of the loading images until `LOAD_END` is signaled. This is attempted in #17420 ![loading-proposal](https://user-images.githubusercontent.com/15007517/43985811-333f57cc-9cbf-11e8-9ce7-fb6657f39575.gif)
priority
cleanup amp image slider loading issue currently there is no code that handles loading for amp image slider by default the slider should construct its structure and load the images normally which shows the placeholder and the loading dots of amp img this is actually one of the proposal proposal a leave as is left is the placeholder for left image the following are under fast so should be much faster on actual devices however due to ownership related issues the dots would only show when layer experiment is enabled otherwise the loading process of images is never displayed another proposal proposal b to the loading state is to whitelist the slider itself for loading and display dots of the slider which would mask on top of the loading images until load end is signaled this is attempted in
1
671,530
22,764,866,088
IssuesEvent
2022-07-08 02:33:02
sacloud/terraform-provider-sakuracloud
https://api.github.com/repos/sacloud/terraform-provider-sakuracloud
closed
sakuracloud_archiveのos_typeパラメータがスコープに対応していない
bug v2 priority/high
os_typeを指定すると対応するパブリックアーカイブを検索するが、その際にscopeが考慮されないためパブリックアーカイブと同一のタグを持つアーカイブがあると意図したアーカイブを参照できない。
1.0
sakuracloud_archiveのos_typeパラメータがスコープに対応していない - os_typeを指定すると対応するパブリックアーカイブを検索するが、その際にscopeが考慮されないためパブリックアーカイブと同一のタグを持つアーカイブがあると意図したアーカイブを参照できない。
priority
sakuracloud archiveのos typeパラメータがスコープに対応していない os typeを指定すると対応するパブリックアーカイブを検索するが、その際にscopeが考慮されないためパブリックアーカイブと同一のタグを持つアーカイブがあると意図したアーカイブを参照できない。
1
230,957
7,621,839,427
IssuesEvent
2018-05-03 09:56:33
HBHWoolacotts/RPii
https://api.github.com/repos/HBHWoolacotts/RPii
closed
Unable to view uploaded Files in Service Jobs (Forbidden)
Label: General RP Bugs and Support Priority - High
Since the server upgrade, we cannot access uploaded files in service jobs. When we try to view them by pressing the little paperclip, it says Forbidden: ![image](https://user-images.githubusercontent.com/10868496/38940847-17bdda78-4323-11e8-944f-3ad01221fd20.png) ![image](https://user-images.githubusercontent.com/10868496/38940853-1b10f840-4323-11e8-839c-e241b56ba919.png)
1.0
Unable to view uploaded Files in Service Jobs (Forbidden) - Since the server upgrade, we cannot access uploaded files in service jobs. When we try to view them by pressing the little paperclip, it says Forbidden: ![image](https://user-images.githubusercontent.com/10868496/38940847-17bdda78-4323-11e8-944f-3ad01221fd20.png) ![image](https://user-images.githubusercontent.com/10868496/38940853-1b10f840-4323-11e8-839c-e241b56ba919.png)
priority
unable to view uploaded files in service jobs forbidden since the server upgrade we cannot access uploaded files in service jobs when we try to view them by pressing the little paperclip it says forbidden
1
366,443
10,820,907,083
IssuesEvent
2019-11-08 17:23:39
Automattic/simplenote-macos
https://api.github.com/repos/Automattic/simplenote-macos
opened
Crash: Emojis Again!
[priority] high bug
### Details: 1. Add a new note 2. Enter the following emojis: `☺️🖖🏿` 3. Insert the following emoji in between: `😃` As a result the app will crash. We'll end up with a malformed unicode sequence, which effectively breaks Apple's JSON Parser.
1.0
Crash: Emojis Again! - ### Details: 1. Add a new note 2. Enter the following emojis: `☺️🖖🏿` 3. Insert the following emoji in between: `😃` As a result the app will crash. We'll end up with a malformed unicode sequence, which effectively breaks Apple's JSON Parser.
priority
crash emojis again details add a new note enter the following emojis ☺️🖖🏿 insert the following emoji in between 😃 as a result the app will crash we ll end up with a malformed unicode sequence which effectively breaks apple s json parser
1
486,499
14,010,133,872
IssuesEvent
2020-10-29 04:15:18
CIA-Homebrew/BJCP-Scoresheet
https://api.github.com/repos/CIA-Homebrew/BJCP-Scoresheet
closed
Multiple scoresheet PDFs should be downloaded as a zip
HIGH PRIORITY back end enhancement
Currently, downloading multiple scoresheets will start multiple individual downloads. When more than one download is requested, there should be a zip handler that puts all pdfs in a single zip Consider using this lib: https://github.com/archiverjs/node-archiver
1.0
Multiple scoresheet PDFs should be downloaded as a zip - Currently, downloading multiple scoresheets will start multiple individual downloads. When more than one download is requested, there should be a zip handler that puts all pdfs in a single zip Consider using this lib: https://github.com/archiverjs/node-archiver
priority
multiple scoresheet pdfs should be downloaded as a zip currently downloading multiple scoresheets will start multiple individual downloads when more than one download is requested there should be a zip handler that puts all pdfs in a single zip consider using this lib
1
504,337
14,616,802,865
IssuesEvent
2020-12-22 13:50:19
SAP/ownid-webapp
https://api.github.com/repos/SAP/ownid-webapp
opened
IOS 14 cookies
Priority: High Type: Bug
If the user disabled FIDO, what will happen after 8 days? Login flow ends with "Account recovery needed" message. Tested on PROD (demo.ownid.com) with iPhone without FIDO
1.0
IOS 14 cookies - If the user disabled FIDO, what will happen after 8 days? Login flow ends with "Account recovery needed" message. Tested on PROD (demo.ownid.com) with iPhone without FIDO
priority
ios cookies if the user disabled fido what will happen after days login flow ends with account recovery needed message tested on prod demo ownid com with iphone without fido
1
359,885
10,682,144,325
IssuesEvent
2019-10-22 03:59:45
ele-l10n-cjk/wingpanel-indicator-inputmethod
https://api.github.com/repos/ele-l10n-cjk/wingpanel-indicator-inputmethod
opened
Human readable engine names
Priority: High
"mozc-jp" should be "Mozc", "libpynin" should be "Intelligent Pynin", and so on
1.0
Human readable engine names - "mozc-jp" should be "Mozc", "libpynin" should be "Intelligent Pynin", and so on
priority
human readable engine names mozc jp should be mozc libpynin should be intelligent pynin and so on
1
373,032
11,032,092,490
IssuesEvent
2019-12-06 19:20:10
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
Giving Reputation to a User with an ASCII username from steam (displayed as ????? ingame) will crash the game of the person giving the reputation
High Priority
Log: [crash (1).txt](https://github.com/StrangeLoopGames/EcoIssues/files/3933667/crash.1.txt)
1.0
Giving Reputation to a User with an ASCII username from steam (displayed as ????? ingame) will crash the game of the person giving the reputation - Log: [crash (1).txt](https://github.com/StrangeLoopGames/EcoIssues/files/3933667/crash.1.txt)
priority
giving reputation to a user with an ascii username from steam displayed as ingame will crash the game of the person giving the reputation log
1
180,350
6,648,819,156
IssuesEvent
2017-09-28 10:46:53
minio/minio
https://api.github.com/repos/minio/minio
closed
Prepare first Minio Azure VM
priority: high
Create a json template to download Minio binary and configure it with a azure storage account entered by the user before VM deployment.
1.0
Prepare first Minio Azure VM - Create a json template to download Minio binary and configure it with a azure storage account entered by the user before VM deployment.
priority
prepare first minio azure vm create a json template to download minio binary and configure it with a azure storage account entered by the user before vm deployment
1
702,440
24,122,321,872
IssuesEvent
2022-09-20 19:55:28
misskey-dev/misskey
https://api.github.com/repos/misskey-dev/misskey
opened
Jestが動作しない
:question:needs more investigation ⚠️bug? 🔥high priority
`SyntaxError: The requested module '@/config.js' does not provide an export named 'Config'`とか出る
1.0
Jestが動作しない - `SyntaxError: The requested module '@/config.js' does not provide an export named 'Config'`とか出る
priority
jestが動作しない syntaxerror the requested module config js does not provide an export named config とか出る
1
216,004
7,300,015,087
IssuesEvent
2018-02-26 22:03:47
opl-/school-ebook-store
https://api.github.com/repos/opl-/school-ebook-store
closed
[Backend] System logowania
priority-high
Pozwala na łatwe przechowywanie zakupionych książek i zapamiętywanie ustawień użytkownika.
1.0
[Backend] System logowania - Pozwala na łatwe przechowywanie zakupionych książek i zapamiętywanie ustawień użytkownika.
priority
system logowania pozwala na łatwe przechowywanie zakupionych książek i zapamiętywanie ustawień użytkownika
1
439,470
12,683,117,186
IssuesEvent
2020-06-19 18:58:15
RE-SS3D/SS3D
https://api.github.com/repos/RE-SS3D/SS3D
closed
Add client-side interactions & Examine
High Priority
Allow for interactions that only run client-side, so that we can add interactions such as 'examine' ### Summary - Any given interaction should be able to 'mark' that it is client-side. - As a test implement the 'examine' interaction. Should be as extensible as possible. - It'd probably be a good idea to consult on specifically how the examine interaction is extensible
1.0
Add client-side interactions & Examine - Allow for interactions that only run client-side, so that we can add interactions such as 'examine' ### Summary - Any given interaction should be able to 'mark' that it is client-side. - As a test implement the 'examine' interaction. Should be as extensible as possible. - It'd probably be a good idea to consult on specifically how the examine interaction is extensible
priority
add client side interactions examine allow for interactions that only run client side so that we can add interactions such as examine summary any given interaction should be able to mark that it is client side as a test implement the examine interaction should be as extensible as possible it d probably be a good idea to consult on specifically how the examine interaction is extensible
1
678,127
23,189,531,641
IssuesEvent
2022-08-01 11:23:50
kubermatic/kubermatic
https://api.github.com/repos/kubermatic/kubermatic
closed
User cannot delete Cluster Template
kind/bug priority/high
### What happened? As a project user I was trying to remove one of existing cluster templates. The API request return 200 but probably due to an existing finalizer (kubermatic.k8c.io/cleanup-credentials-secrets) it stays in the terminating state forever. ### Expected behavior After deleting `clustertemplate` it should be removed from the cluster. ### How to reproduce the issue? 1. Create a project or use one of existing ones. 2. Go to `Cluster Templates` page and create a new cluster template. 3. After the new cluster template is saved go back to the cluster templates page again and try to remove it. ### How is your environment configured? - KKP version: 2.21 (dev on 29.07.2022)
1.0
User cannot delete Cluster Template - ### What happened? As a project user I was trying to remove one of existing cluster templates. The API request return 200 but probably due to an existing finalizer (kubermatic.k8c.io/cleanup-credentials-secrets) it stays in the terminating state forever. ### Expected behavior After deleting `clustertemplate` it should be removed from the cluster. ### How to reproduce the issue? 1. Create a project or use one of existing ones. 2. Go to `Cluster Templates` page and create a new cluster template. 3. After the new cluster template is saved go back to the cluster templates page again and try to remove it. ### How is your environment configured? - KKP version: 2.21 (dev on 29.07.2022)
priority
user cannot delete cluster template what happened as a project user i was trying to remove one of existing cluster templates the api request return but probably due to an existing finalizer kubermatic io cleanup credentials secrets it stays in the terminating state forever expected behavior after deleting clustertemplate it should be removed from the cluster how to reproduce the issue create a project or use one of existing ones go to cluster templates page and create a new cluster template after the new cluster template is saved go back to the cluster templates page again and try to remove it how is your environment configured kkp version dev on
1
653,398
21,581,373,702
IssuesEvent
2022-05-02 19:07:02
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
WorkQueue fails to create subscription due to ORACLE unique constraint violation
BUG WMAgent WorkQueue Highest Priority
**Impact of the bug** WorkQueue, WMBS **Describe the bug** While checking the status of few workflows upon a P&R request, I've found plenty of Oracle errors: `ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated` at the two newly deployed CERN agents. The agent version deployed is **v2.0.2.patch1**. This is an error we do not observe neither in the FNAL agents (they are using MAriaDB anyway) nor at the agents currently on drain. Taking as an example a workflow that is been distributed amongst those all agents: [cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873](https://cmsweb.cern.ch/reqmgr2/fetch?rid=cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873) **agent url:** cmsgwms-submit4.fnal.gov,cmsgwms-submit5.fnal.gov,cmsgwms-submit6.fnal.gov,cmsgwms-submit7.fnal.gov,vocms0252.cern.ch,vocms0253.cern.ch,vocms0282.cern.ch,vocms0283.cern.ch We observe the following at `vocms0282 - v2.0.2.patch1`: [1], and no errors at all in both `vocms0252`. In addition to that I see quite low number of WMBS jobs created for those two new CERN agents. Something like 866 jobs for `vocms0282` vs 43K jobs for `cmsgwms-submit7.fnal.gov`. And hence the low number of idle condor jobs for the two new CERN agents: ``` vocms0282: $ condor_q --totals Total for query: 6187 jobs; 0 completed, 0 removed, 4224 idle, 1963 running, 0 held, 0 suspended vocms0283: $ condor_q --totals Total for query: 30827 jobs; 0 completed, 0 removed, 3695 idle, 27132 running, 0 held, 0 suspended ``` **How to reproduce it** Steps to reproduce the behavior: **Expected behavior** A clear and concise description of what you expected to happen. **Additional context and error message** Add any other context about the problem here, like error message and/or traceback. You might want to use triple back ticks to properly format it. [1] ``` 2022-04-13 23:44:52,787:139862773548800:ERROR:WMBSHelper:Failed to create subscription. Error: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated [SQL: INSERT INTO dbsbuffer_dataset_subscription (id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime) SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve, :move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime FROM DUAL WHERE NOT EXISTS ( SELECT * FROM dbsbuffer_dataset_subscription WHERE dataset_id = :id AND site = :site AND custodial = :custodial AND auto_approve = :auto_approve AND move = :move AND priority = :priority AND dataset_lifetime = :dataset_lifetime ) ] [parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}] (Background on this error at: http://sqlalche.me/e/gkpj) Traceback (most recent call last): File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context self.dialect.do_execute( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute cursor.execute(statement, parameters) cx_Oracle.IntegrityError: ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 413, in createSubscriptionAndAddFiles self.createSubscription(self.topLevelTask, self.topLevelFileset) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 266, in createSubscription self._createDatasetSubscriptionsInDBSBuffer() File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 520, in _createDatasetSubscriptionsInDBSBuffer dbsDataset.addSubscription(subInfo[dataset]) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/DBSBufferDataset.py", line 140, in addSubscription action.execute(self['id'], subscriptionInformation, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/MySQL/NewSubscription.py", line 66, in execute self.dbi.processData(self.sql, binds=binds, conn=conn, transaction=transaction) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 172, in processData r = self.executebinds(s, b, connection=connection, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 64, in executebinds resultProxy = connection.execute(s, b) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 982, in execute return self._execute_text(object_, multiparams, params) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1149, in _execute_text ret = self._execute_context( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1247, in _execute_context self._handle_dbapi_exception( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1466, in _handle_dbapi_exception util.raise_from_cause(sqlalchemy_exception, exc_info) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 383, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 128, in reraise raise value.with_traceback(tb) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context self.dialect.do_execute( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute cursor.execute(statement, parameters) sqlalchemy.exc.IntegrityError: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated [SQL: INSERT INTO dbsbuffer_dataset_subscription (id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime) SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve, :move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime FROM DUAL WHERE NOT EXISTS ( SELECT * FROM dbsbuffer_dataset_subscription WHERE dataset_id = :id AND site = :site AND custodial = :custodial AND auto_approve = :auto_approve AND move = :move AND priority = :priority AND dataset_lifetime = :dataset_lifetime ) ] [parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}] (Background on this error at: http://sqlalche.me/e/gkpj) 2022-04-13 23:44:52,789:139862773548800:ERROR:WorkQueue:Failed to create subscription for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 with block name None Error: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated [SQL: INSERT INTO dbsbuffer_dataset_subscription (id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime) SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve, :move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime FROM DUAL WHERE NOT EXISTS ( SELECT * FROM dbsbuffer_dataset_subscription WHERE dataset_id = :id AND site = :site AND custodial = :custodial AND auto_approve = :auto_approve AND move = :move AND priority = :priority AND dataset_lifetime = :dataset_lifetime ) ] [parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}] (Background on this error at: http://sqlalche.me/e/gkpj) Traceback (most recent call last): File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context self.dialect.do_execute( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute cursor.execute(statement, parameters) cx_Oracle.IntegrityError: ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WorkQueue.py", line 352, in getWork match['Subscription'] = self._wmbsPreparation(match, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WorkQueue.py", line 453, in _wmbsPreparation sub, match['NumOfFilesAdded'] = wmbsHelper.createSubscriptionAndAddFiles(block=dbsBlock) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 418, in createSubscriptionAndAddFiles raise ex File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 413, in createSubscriptionAndAddFiles self.createSubscription(self.topLevelTask, self.topLevelFileset) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 266, in createSubscription self._createDatasetSubscriptionsInDBSBuffer() File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 520, in _createDatasetSubscriptionsInDBSBuffer dbsDataset.addSubscription(subInfo[dataset]) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/DBSBufferDataset.py", line 140, in addSubscription action.execute(self['id'], subscriptionInformation, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/MySQL/NewSubscription.py", line 66, in execute self.dbi.processData(self.sql, binds=binds, conn=conn, transaction=transaction) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 172, in processData r = self.executebinds(s, b, connection=connection, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 64, in executebinds resultProxy = connection.execute(s, b) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 982, in execute return self._execute_text(object_, multiparams, params) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1149, in _execute_text ret = self._execute_context( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1247, in _execute_context self._handle_dbapi_exception( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1466, in _handle_dbapi_exception util.raise_from_cause(sqlalchemy_exception, exc_info) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 383, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 128, in reraise raise value.with_traceback(tb) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context self.dialect.do_execute( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute cursor.execute(statement, parameters) sqlalchemy.exc.IntegrityError: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated [SQL: INSERT INTO dbsbuffer_dataset_subscription (id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime) SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve, :move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime FROM DUAL WHERE NOT EXISTS ( SELECT * FROM dbsbuffer_dataset_subscription WHERE dataset_id = :id AND site = :site AND custodial = :custodial AND auto_approve = :auto_approve AND move = :move AND priority = :priority AND dataset_lifetime = :dataset_lifetime ) ] [parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}] (Background on this error at: http://sqlalche.me/e/gkpj) ``` [2] ``` 2022-04-11 10:24:24,733:140665460737792:INFO:WorkQueue:Running WMBS preparation for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 with ParentQueueId 84cadfd2061398ef871a19d9bb5d7466, with common location ['T2_US_Florida'] 2022-04-11 10:24:24,744:140665460737792:INFO:Fileset:Fileset created: cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873-TSG-Run3Winter22GS-00002_0-27765ea9ee953e8b90915707e32347a3 2022-04-11 10:24:26,349:140665460737792:INFO:Rucio:WMCore Rucio initialization parameters: {'account': 'wmcore_transferor', 'rucio_host': None, 'auth_host': None, 'ca_cert': None, 'auth_type': None, 'creds': None, 'timeout': 600, 'user_agent': 'wmcore-client'} 2022-04-11 10:24:26,352:140665460737792:INFO:Rucio:Rucio client initialization parameters: {'host': 'http://cms-rucio.cern.ch', 'auth_host': 'https://cms-rucio-auth.cern.ch', 'auth_type': 'x509', 'account': 'wmcore_transferor', 'user_agent': 'wmcore-client/1.25.5', 'ca_cert': '/etc/grid-security/certificates/', 'creds': {'client_cert': '/data/certs/servicecert.pem', 'client_key': '/data/certs/servicekey.pem'}, 'timeout': 600, 'request_retries': 3} 2022-04-11 10:24:26,462:140665460737792:INFO:Rucio:Pileup container location for /MinBias_TuneCP5_13p6TeV-pythia8/Run3Winter22GS-122X_mcRun3_2021_realistic_v9-v1/GEN-SIM from single RSE locks at: ['T1_US_FNAL_Disk', 'T2_US_Florida', 'T1_ES_PIC_Disk'] 2022-04-11 10:24:28,757:140665460737792:INFO:SandboxCreator:Created sandbox cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873-Sandbox.tar.bz2 with size 3197422 2022-04-11 10:24:28,776:140665460737792:INFO:Workflow:Workflow id 47920 created for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 2022-04-11 10:24:28,788:140665460737792:INFO:WMBSHelper:Top level subscription 67000 created for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 2022-04-11 10:24:28,793:140665460737792:INFO:Fileset:Fileset created: /cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873/TSG-Run3Winter22GS-00002_0/unmerged-RAWSIMoutputGEN-SIM ``` [3] ``` ```
1.0
WorkQueue fails to create subscription due to ORACLE unique constraint violation - **Impact of the bug** WorkQueue, WMBS **Describe the bug** While checking the status of few workflows upon a P&R request, I've found plenty of Oracle errors: `ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated` at the two newly deployed CERN agents. The agent version deployed is **v2.0.2.patch1**. This is an error we do not observe neither in the FNAL agents (they are using MAriaDB anyway) nor at the agents currently on drain. Taking as an example a workflow that is been distributed amongst those all agents: [cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873](https://cmsweb.cern.ch/reqmgr2/fetch?rid=cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873) **agent url:** cmsgwms-submit4.fnal.gov,cmsgwms-submit5.fnal.gov,cmsgwms-submit6.fnal.gov,cmsgwms-submit7.fnal.gov,vocms0252.cern.ch,vocms0253.cern.ch,vocms0282.cern.ch,vocms0283.cern.ch We observe the following at `vocms0282 - v2.0.2.patch1`: [1], and no errors at all in both `vocms0252`. In addition to that I see quite low number of WMBS jobs created for those two new CERN agents. Something like 866 jobs for `vocms0282` vs 43K jobs for `cmsgwms-submit7.fnal.gov`. And hence the low number of idle condor jobs for the two new CERN agents: ``` vocms0282: $ condor_q --totals Total for query: 6187 jobs; 0 completed, 0 removed, 4224 idle, 1963 running, 0 held, 0 suspended vocms0283: $ condor_q --totals Total for query: 30827 jobs; 0 completed, 0 removed, 3695 idle, 27132 running, 0 held, 0 suspended ``` **How to reproduce it** Steps to reproduce the behavior: **Expected behavior** A clear and concise description of what you expected to happen. **Additional context and error message** Add any other context about the problem here, like error message and/or traceback. You might want to use triple back ticks to properly format it. [1] ``` 2022-04-13 23:44:52,787:139862773548800:ERROR:WMBSHelper:Failed to create subscription. Error: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated [SQL: INSERT INTO dbsbuffer_dataset_subscription (id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime) SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve, :move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime FROM DUAL WHERE NOT EXISTS ( SELECT * FROM dbsbuffer_dataset_subscription WHERE dataset_id = :id AND site = :site AND custodial = :custodial AND auto_approve = :auto_approve AND move = :move AND priority = :priority AND dataset_lifetime = :dataset_lifetime ) ] [parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}] (Background on this error at: http://sqlalche.me/e/gkpj) Traceback (most recent call last): File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context self.dialect.do_execute( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute cursor.execute(statement, parameters) cx_Oracle.IntegrityError: ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 413, in createSubscriptionAndAddFiles self.createSubscription(self.topLevelTask, self.topLevelFileset) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 266, in createSubscription self._createDatasetSubscriptionsInDBSBuffer() File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 520, in _createDatasetSubscriptionsInDBSBuffer dbsDataset.addSubscription(subInfo[dataset]) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/DBSBufferDataset.py", line 140, in addSubscription action.execute(self['id'], subscriptionInformation, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/MySQL/NewSubscription.py", line 66, in execute self.dbi.processData(self.sql, binds=binds, conn=conn, transaction=transaction) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 172, in processData r = self.executebinds(s, b, connection=connection, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 64, in executebinds resultProxy = connection.execute(s, b) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 982, in execute return self._execute_text(object_, multiparams, params) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1149, in _execute_text ret = self._execute_context( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1247, in _execute_context self._handle_dbapi_exception( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1466, in _handle_dbapi_exception util.raise_from_cause(sqlalchemy_exception, exc_info) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 383, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 128, in reraise raise value.with_traceback(tb) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context self.dialect.do_execute( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute cursor.execute(statement, parameters) sqlalchemy.exc.IntegrityError: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated [SQL: INSERT INTO dbsbuffer_dataset_subscription (id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime) SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve, :move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime FROM DUAL WHERE NOT EXISTS ( SELECT * FROM dbsbuffer_dataset_subscription WHERE dataset_id = :id AND site = :site AND custodial = :custodial AND auto_approve = :auto_approve AND move = :move AND priority = :priority AND dataset_lifetime = :dataset_lifetime ) ] [parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}] (Background on this error at: http://sqlalche.me/e/gkpj) 2022-04-13 23:44:52,789:139862773548800:ERROR:WorkQueue:Failed to create subscription for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 with block name None Error: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated [SQL: INSERT INTO dbsbuffer_dataset_subscription (id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime) SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve, :move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime FROM DUAL WHERE NOT EXISTS ( SELECT * FROM dbsbuffer_dataset_subscription WHERE dataset_id = :id AND site = :site AND custodial = :custodial AND auto_approve = :auto_approve AND move = :move AND priority = :priority AND dataset_lifetime = :dataset_lifetime ) ] [parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}] (Background on this error at: http://sqlalche.me/e/gkpj) Traceback (most recent call last): File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context self.dialect.do_execute( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute cursor.execute(statement, parameters) cx_Oracle.IntegrityError: ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WorkQueue.py", line 352, in getWork match['Subscription'] = self._wmbsPreparation(match, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WorkQueue.py", line 453, in _wmbsPreparation sub, match['NumOfFilesAdded'] = wmbsHelper.createSubscriptionAndAddFiles(block=dbsBlock) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 418, in createSubscriptionAndAddFiles raise ex File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 413, in createSubscriptionAndAddFiles self.createSubscription(self.topLevelTask, self.topLevelFileset) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 266, in createSubscription self._createDatasetSubscriptionsInDBSBuffer() File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/WorkQueue/WMBSHelper.py", line 520, in _createDatasetSubscriptionsInDBSBuffer dbsDataset.addSubscription(subInfo[dataset]) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/DBSBufferDataset.py", line 140, in addSubscription action.execute(self['id'], subscriptionInformation, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMComponent/DBS3Buffer/MySQL/NewSubscription.py", line 66, in execute self.dbi.processData(self.sql, binds=binds, conn=conn, transaction=transaction) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 172, in processData r = self.executebinds(s, b, connection=connection, File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/cms/wmagentpy3/2.0.2.patch1/lib/python3.8/site-packages/WMCore/Database/DBCore.py", line 64, in executebinds resultProxy = connection.execute(s, b) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 982, in execute return self._execute_text(object_, multiparams, params) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1149, in _execute_text ret = self._execute_context( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1247, in _execute_context self._handle_dbapi_exception( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1466, in _handle_dbapi_exception util.raise_from_cause(sqlalchemy_exception, exc_info) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 383, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 128, in reraise raise value.with_traceback(tb) File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1243, in _execute_context self.dialect.do_execute( File "/data/srv/wmagent/v2.0.2.patch1/sw/slc7_amd64_gcc630/external/py3-sqlalchemy/1.3.3-comp/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 552, in do_execute cursor.execute(statement, parameters) sqlalchemy.exc.IntegrityError: (cx_Oracle.IntegrityError) ORA-00001: unique constraint (CMS_WMBS_PROD12.UQ_DBS_DAT_SUB) violated [SQL: INSERT INTO dbsbuffer_dataset_subscription (id, dataset_id, site, custodial, auto_approve, move, priority, subscribed, phedex_group, delete_blocks, dataset_lifetime) SELECT dbsbuffer_dataset_sub_seq.nextval, :id, :site, :custodial, :auto_approve, :move, :priority, 0, :phedex_group, :delete_blocks, :dataset_lifetime FROM DUAL WHERE NOT EXISTS ( SELECT * FROM dbsbuffer_dataset_subscription WHERE dataset_id = :id AND site = :site AND custodial = :custodial AND auto_approve = :auto_approve AND move = :move AND priority = :priority AND dataset_lifetime = :dataset_lifetime ) ] [parameters: {'id': 1, 'site': 'T2_US_Florida', 'custodial': False, 'auto_approve': 1, 'move': 0, 'priority': 'Low', 'phedex_group': 'DataOps', 'delete_blocks': None, 'dataset_lifetime': None}] (Background on this error at: http://sqlalche.me/e/gkpj) ``` [2] ``` 2022-04-11 10:24:24,733:140665460737792:INFO:WorkQueue:Running WMBS preparation for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 with ParentQueueId 84cadfd2061398ef871a19d9bb5d7466, with common location ['T2_US_Florida'] 2022-04-11 10:24:24,744:140665460737792:INFO:Fileset:Fileset created: cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873-TSG-Run3Winter22GS-00002_0-27765ea9ee953e8b90915707e32347a3 2022-04-11 10:24:26,349:140665460737792:INFO:Rucio:WMCore Rucio initialization parameters: {'account': 'wmcore_transferor', 'rucio_host': None, 'auth_host': None, 'ca_cert': None, 'auth_type': None, 'creds': None, 'timeout': 600, 'user_agent': 'wmcore-client'} 2022-04-11 10:24:26,352:140665460737792:INFO:Rucio:Rucio client initialization parameters: {'host': 'http://cms-rucio.cern.ch', 'auth_host': 'https://cms-rucio-auth.cern.ch', 'auth_type': 'x509', 'account': 'wmcore_transferor', 'user_agent': 'wmcore-client/1.25.5', 'ca_cert': '/etc/grid-security/certificates/', 'creds': {'client_cert': '/data/certs/servicecert.pem', 'client_key': '/data/certs/servicekey.pem'}, 'timeout': 600, 'request_retries': 3} 2022-04-11 10:24:26,462:140665460737792:INFO:Rucio:Pileup container location for /MinBias_TuneCP5_13p6TeV-pythia8/Run3Winter22GS-122X_mcRun3_2021_realistic_v9-v1/GEN-SIM from single RSE locks at: ['T1_US_FNAL_Disk', 'T2_US_Florida', 'T1_ES_PIC_Disk'] 2022-04-11 10:24:28,757:140665460737792:INFO:SandboxCreator:Created sandbox cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873-Sandbox.tar.bz2 with size 3197422 2022-04-11 10:24:28,776:140665460737792:INFO:Workflow:Workflow id 47920 created for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 2022-04-11 10:24:28,788:140665460737792:INFO:WMBSHelper:Top level subscription 67000 created for cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873 2022-04-11 10:24:28,793:140665460737792:INFO:Fileset:Fileset created: /cmsunified_task_TSG-Run3Winter22GS-00002__v1_T_220323_185605_8873/TSG-Run3Winter22GS-00002_0/unmerged-RAWSIMoutputGEN-SIM ``` [3] ``` ```
priority
workqueue fails to create subscription due to oracle unique constraint violation impact of the bug workqueue wmbs describe the bug while checking the status of few workflows upon a p r request i ve found plenty of oracle errors ora unique constraint cms wmbs uq dbs dat sub violated at the two newly deployed cern agents the agent version deployed is this is an error we do not observe neither in the fnal agents they are using mariadb anyway nor at the agents currently on drain taking as an example a workflow that is been distributed amongst those all agents agent url cmsgwms fnal gov cmsgwms fnal gov cmsgwms fnal gov cmsgwms fnal gov cern ch cern ch cern ch cern ch we observe the following at and no errors at all in both in addition to that i see quite low number of wmbs jobs created for those two new cern agents something like jobs for vs jobs for cmsgwms fnal gov and hence the low number of idle condor jobs for the two new cern agents condor q totals total for query jobs completed removed idle running held suspended condor q totals total for query jobs completed removed idle running held suspended how to reproduce it steps to reproduce the behavior expected behavior a clear and concise description of what you expected to happen additional context and error message add any other context about the problem here like error message and or traceback you might want to use triple back ticks to properly format it error wmbshelper failed to create subscription error cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated sql insert into dbsbuffer dataset subscription id dataset id site custodial auto approve move priority subscribed phedex group delete blocks dataset lifetime select dbsbuffer dataset sub seq nextval id site custodial auto approve move priority phedex group delete blocks dataset lifetime from dual where not exists select from dbsbuffer dataset subscription where dataset id id and site site and custodial custodial and auto approve auto approve and move move and priority priority and dataset lifetime dataset lifetime background on this error at traceback most recent call last file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self dialect do execute file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated the above exception was the direct cause of the following exception traceback most recent call last file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscriptionandaddfiles self createsubscription self topleveltask self toplevelfileset file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscription self createdatasetsubscriptionsindbsbuffer file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createdatasetsubscriptionsindbsbuffer dbsdataset addsubscription subinfo file data srv wmagent sw cms lib site packages wmcomponent dbsbufferdataset py line in addsubscription action execute self subscriptioninformation file data srv wmagent sw cms lib site packages wmcomponent mysql newsubscription py line in execute self dbi processdata self sql binds binds conn conn transaction transaction file data srv wmagent sw cms lib site packages wmcore database dbcore py line in processdata r self executebinds s b connection connection file data srv wmagent sw cms lib site packages wmcore database dbcore py line in executebinds resultproxy connection execute s b file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute return self execute text object multiparams params file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute text ret self execute context file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self handle dbapi exception file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in handle dbapi exception util raise from cause sqlalchemy exception exc info file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy util compat py line in raise from cause reraise type exception exception tb exc tb cause cause file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy util compat py line in reraise raise value with traceback tb file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self dialect do execute file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters sqlalchemy exc integrityerror cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated sql insert into dbsbuffer dataset subscription id dataset id site custodial auto approve move priority subscribed phedex group delete blocks dataset lifetime select dbsbuffer dataset sub seq nextval id site custodial auto approve move priority phedex group delete blocks dataset lifetime from dual where not exists select from dbsbuffer dataset subscription where dataset id id and site site and custodial custodial and auto approve auto approve and move move and priority priority and dataset lifetime dataset lifetime background on this error at error workqueue failed to create subscription for cmsunified task tsg t with block name none error cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated sql insert into dbsbuffer dataset subscription id dataset id site custodial auto approve move priority subscribed phedex group delete blocks dataset lifetime select dbsbuffer dataset sub seq nextval id site custodial auto approve move priority phedex group delete blocks dataset lifetime from dual where not exists select from dbsbuffer dataset subscription where dataset id id and site site and custodial custodial and auto approve auto approve and move move and priority priority and dataset lifetime dataset lifetime background on this error at traceback most recent call last file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self dialect do execute file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated the above exception was the direct cause of the following exception traceback most recent call last file data srv wmagent sw cms lib site packages wmcore workqueue workqueue py line in getwork match self wmbspreparation match file data srv wmagent sw cms lib site packages wmcore workqueue workqueue py line in wmbspreparation sub match wmbshelper createsubscriptionandaddfiles block dbsblock file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscriptionandaddfiles raise ex file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscriptionandaddfiles self createsubscription self topleveltask self toplevelfileset file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createsubscription self createdatasetsubscriptionsindbsbuffer file data srv wmagent sw cms lib site packages wmcore workqueue wmbshelper py line in createdatasetsubscriptionsindbsbuffer dbsdataset addsubscription subinfo file data srv wmagent sw cms lib site packages wmcomponent dbsbufferdataset py line in addsubscription action execute self subscriptioninformation file data srv wmagent sw cms lib site packages wmcomponent mysql newsubscription py line in execute self dbi processdata self sql binds binds conn conn transaction transaction file data srv wmagent sw cms lib site packages wmcore database dbcore py line in processdata r self executebinds s b connection connection file data srv wmagent sw cms lib site packages wmcore database dbcore py line in executebinds resultproxy connection execute s b file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute return self execute text object multiparams params file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute text ret self execute context file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self handle dbapi exception file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in handle dbapi exception util raise from cause sqlalchemy exception exc info file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy util compat py line in raise from cause reraise type exception exception tb exc tb cause cause file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy util compat py line in reraise raise value with traceback tb file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine base py line in execute context self dialect do execute file data srv wmagent sw external sqlalchemy comp lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters sqlalchemy exc integrityerror cx oracle integrityerror ora unique constraint cms wmbs uq dbs dat sub violated sql insert into dbsbuffer dataset subscription id dataset id site custodial auto approve move priority subscribed phedex group delete blocks dataset lifetime select dbsbuffer dataset sub seq nextval id site custodial auto approve move priority phedex group delete blocks dataset lifetime from dual where not exists select from dbsbuffer dataset subscription where dataset id id and site site and custodial custodial and auto approve auto approve and move move and priority priority and dataset lifetime dataset lifetime background on this error at info workqueue running wmbs preparation for cmsunified task tsg t with parentqueueid with common location info fileset fileset created cmsunified task tsg t tsg info rucio wmcore rucio initialization parameters account wmcore transferor rucio host none auth host none ca cert none auth type none creds none timeout user agent wmcore client info rucio rucio client initialization parameters host auth host auth type account wmcore transferor user agent wmcore client ca cert etc grid security certificates creds client cert data certs servicecert pem client key data certs servicekey pem timeout request retries info rucio pileup container location for minbias realistic gen sim from single rse locks at info sandboxcreator created sandbox cmsunified task tsg t sandbox tar with size info workflow workflow id created for cmsunified task tsg t info wmbshelper top level subscription created for cmsunified task tsg t info fileset fileset created cmsunified task tsg t tsg unmerged rawsimoutputgen sim
1
315,145
9,606,963,978
IssuesEvent
2019-05-11 15:00:31
thechutrain/rc-coffee-chats
https://api.github.com/repos/thechutrain/rc-coffee-chats
closed
Change the fallback user to someone else during NGW
high priority
Facilitators will be busy during NGW organizing events, so it would be best to have the fallback user be a current or a Recurser who would be fine without a coffee chat.
1.0
Change the fallback user to someone else during NGW - Facilitators will be busy during NGW organizing events, so it would be best to have the fallback user be a current or a Recurser who would be fine without a coffee chat.
priority
change the fallback user to someone else during ngw facilitators will be busy during ngw organizing events so it would be best to have the fallback user be a current or a recurser who would be fine without a coffee chat
1
650,764
21,416,721,590
IssuesEvent
2022-04-22 11:37:24
lucypoulton/acebook
https://api.github.com/repos/lucypoulton/acebook
closed
Users can comment on posts
size: large type: feature priority: high
- [ ] Users can interact with a post and leave a comment. - [ ] User comments can be seen underneath the original post - [ ] Posted comments have the Commenters/Users name next to them
1.0
Users can comment on posts - - [ ] Users can interact with a post and leave a comment. - [ ] User comments can be seen underneath the original post - [ ] Posted comments have the Commenters/Users name next to them
priority
users can comment on posts users can interact with a post and leave a comment user comments can be seen underneath the original post posted comments have the commenters users name next to them
1
98,455
4,021,535,905
IssuesEvent
2016-05-16 22:22:03
theBoardlist/website
https://api.github.com/repos/theBoardlist/website
opened
Unable to add email address to account with no email
bug High Priority
Steps to reproduce: 1. Go to a profile page for an account with no email address associated with it; e.g., Amy Banse 2. Click edit > add email address 3. Enter an email that isn't already associated with an account and click SAVE Expected result: Email added to account, profile updated Actual result: No email added, profile not updated. We get the screen that we normally see if we're trying to add an email address that's already in the system. ![screen shot 2016-05-16 at 3 21 06 pm](https://cloud.githubusercontent.com/assets/16053877/15305749/efcbd4ac-1b79-11e6-8164-bb5ba232d61a.png)
1.0
Unable to add email address to account with no email - Steps to reproduce: 1. Go to a profile page for an account with no email address associated with it; e.g., Amy Banse 2. Click edit > add email address 3. Enter an email that isn't already associated with an account and click SAVE Expected result: Email added to account, profile updated Actual result: No email added, profile not updated. We get the screen that we normally see if we're trying to add an email address that's already in the system. ![screen shot 2016-05-16 at 3 21 06 pm](https://cloud.githubusercontent.com/assets/16053877/15305749/efcbd4ac-1b79-11e6-8164-bb5ba232d61a.png)
priority
unable to add email address to account with no email steps to reproduce go to a profile page for an account with no email address associated with it e g amy banse click edit add email address enter an email that isn t already associated with an account and click save expected result email added to account profile updated actual result no email added profile not updated we get the screen that we normally see if we re trying to add an email address that s already in the system
1
605,614
18,737,748,197
IssuesEvent
2021-11-04 09:52:32
betagouv/service-national-universel
https://api.github.com/repos/betagouv/service-national-universel
closed
fix(inscription): préparer dashboard pour inscription 2022
enhancement priority-HIGH inscription
### Fonctionnalité liée à un problème ? _No response_ ### Fonctionnalité il faut que le dashboard, onglet inscirpiton. faire reference a la cohorte 2022. ### Commentaires _No response_
1.0
fix(inscription): préparer dashboard pour inscription 2022 - ### Fonctionnalité liée à un problème ? _No response_ ### Fonctionnalité il faut que le dashboard, onglet inscirpiton. faire reference a la cohorte 2022. ### Commentaires _No response_
priority
fix inscription préparer dashboard pour inscription fonctionnalité liée à un problème no response fonctionnalité il faut que le dashboard onglet inscirpiton faire reference a la cohorte commentaires no response
1
400,490
11,776,002,985
IssuesEvent
2020-03-16 12:27:09
HE-Arc/CSRuby
https://api.github.com/repos/HE-Arc/CSRuby
closed
Mise en place de l'inscription
core high priority
Inscription - email - profilename - password - password confirmation
1.0
Mise en place de l'inscription - Inscription - email - profilename - password - password confirmation
priority
mise en place de l inscription inscription email profilename password password confirmation
1
534,175
15,611,414,397
IssuesEvent
2021-03-19 14:17:53
MaibornWolff/codecharta
https://api.github.com/repos/MaibornWolff/codecharta
opened
Indicate total nodes and excluded / flattened nodes in file explorer
UX / UI difficulty:low feature pr-visualization priority:high
# Feature request ## Description As a user, I want to see the total number of nodes, excluded nodes and flattened nodes so that the numbers correspond to the matching nodes. ![image](https://user-images.githubusercontent.com/8822573/111791604-a8631c80-88c3-11eb-825f-ec7a496f3e65.png) ![image](https://user-images.githubusercontent.com/8822573/111791698-bdd84680-88c3-11eb-97a7-4ebf3f88dd18.png) ## Acceptance criteria - If the search bar is empty, show the total number of nodes, excluded nodes and flattened nodes - As soon as the search is triggered show the number of matching nodes and the total nodes together as indicated in the screenshot (`matchingNodes/totalNodes`)
1.0
Indicate total nodes and excluded / flattened nodes in file explorer - # Feature request ## Description As a user, I want to see the total number of nodes, excluded nodes and flattened nodes so that the numbers correspond to the matching nodes. ![image](https://user-images.githubusercontent.com/8822573/111791604-a8631c80-88c3-11eb-825f-ec7a496f3e65.png) ![image](https://user-images.githubusercontent.com/8822573/111791698-bdd84680-88c3-11eb-97a7-4ebf3f88dd18.png) ## Acceptance criteria - If the search bar is empty, show the total number of nodes, excluded nodes and flattened nodes - As soon as the search is triggered show the number of matching nodes and the total nodes together as indicated in the screenshot (`matchingNodes/totalNodes`)
priority
indicate total nodes and excluded flattened nodes in file explorer feature request description as a user i want to see the total number of nodes excluded nodes and flattened nodes so that the numbers correspond to the matching nodes acceptance criteria if the search bar is empty show the total number of nodes excluded nodes and flattened nodes as soon as the search is triggered show the number of matching nodes and the total nodes together as indicated in the screenshot matchingnodes totalnodes
1
496,626
14,350,593,228
IssuesEvent
2020-11-29 21:33:29
swharden/ScottPlot
https://api.github.com/repos/swharden/ScottPlot
closed
Road to ScottPlot 4.1
HIGH PRIORITY
### Update on November 17, 2020 **⚠️ I am reducing effort on issues and PRs while I work on this daily.** * See [changelog.md](https://github.com/swharden/ScottPlot/blob/master/dev/changelog.md) for a summary of major changes. * See [roadmap.md](https://github.com/swharden/ScottPlot/blob/master/dev/roadmap.md) for a big-picture discussion of plans and goals * I merged #605 which accomplishes most of the goals described on this page * The master branch is now ScottPlot `4.1-beta`. Old source code has its own [4.0-stable](https://github.com/swharden/ScottPlot/tree/4.0-stable) branch. ### Plottables to Refactor - [x] Candlestick / OHLC - [PlottableOHLC.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableOHLC.cs) - [x] Annotation - [PlottableText.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableText.cs) - [x] Bar - [PlottableBar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableBar.cs) - [x] ErrorBar - [PlottableErrorBars.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableErrorBars.cs) - [x] Function - [PlottableFunction.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableFunction.cs) - [x] HLine - [PlottableAxisLine.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisLine.cs) - [x] HSpan - [PlottableAxisSpan.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisSpan.cs) - [x] VLine - [PlottableAxisLine.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisLine.cs) - [x] VSpan - [PlottableAxisSpan.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisSpan.cs) - [x] Heatmap - [PlottableHeatmap.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableHeatmap.cs) - [x] Image - [PlottableImage.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableImage.cs) - [x] Pie - [PlottablePie](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePie.cs) - [x] Polygon - [PlottablePolygon.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePolygon.cs) - [x] Polygons - [PlottablePolygons.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePolygons.cs) - [x] Populations - [PlottablePopulations.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePopulations.cs) - [x] Radar - [PlottableRadar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableRadar.cs) - [x] ScaleBar - [PlottableScaleBar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScaleBar.cs) - [x] Scatter / ScatterHighlight - [PlottableScatter.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScatter.cs) [PlottableScatterHighlight.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScatterHighlight.cs) - [x] Signal / SignalConst / SignalXY (#585) - [x] Text - [PlottableText.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableText.cs) - [x] VectorField - [PlottableVectorField.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableVectorField.cs) ### Renderables to Refactor - [x] Axis labels - [x] Axis tick marks - [x] Axis tick labels - [x] Grid lines - [x] Figure and Data background - [x] Error message - [x] Benchmark message ### Refactor / Large Features - [x] Delete the `Plottable` class and replace it with `IPlottable` - [x] Change plottable namespaces (`PlottableScatter` becomes `Plottable.Scatter`) - [x] support multiple X and Y axes ✨ ### Remaining Tasks - [x] Make GetAxisLimits() return AxisLimits object - [x] Create a control to test a timer-based render method - [x] Create new plottables specifically for growing data (mark public data fields readonly and provide setters to replace data arrays) - [x] Improve data validation for plottables (create separate pre-render and deep validation methods in an interface) ### Extra release of 4.0 - [x] Add `RenderLock()` and `RenderUnlock()` fields to help with multi-threaded plot manipulation - [x] Add NuGet message noting transition to 4.1 and to check preleases - [x] Review all the items on #412 since many have now been completed or made obsolete
1.0
Road to ScottPlot 4.1 - ### Update on November 17, 2020 **⚠️ I am reducing effort on issues and PRs while I work on this daily.** * See [changelog.md](https://github.com/swharden/ScottPlot/blob/master/dev/changelog.md) for a summary of major changes. * See [roadmap.md](https://github.com/swharden/ScottPlot/blob/master/dev/roadmap.md) for a big-picture discussion of plans and goals * I merged #605 which accomplishes most of the goals described on this page * The master branch is now ScottPlot `4.1-beta`. Old source code has its own [4.0-stable](https://github.com/swharden/ScottPlot/tree/4.0-stable) branch. ### Plottables to Refactor - [x] Candlestick / OHLC - [PlottableOHLC.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableOHLC.cs) - [x] Annotation - [PlottableText.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableText.cs) - [x] Bar - [PlottableBar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableBar.cs) - [x] ErrorBar - [PlottableErrorBars.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableErrorBars.cs) - [x] Function - [PlottableFunction.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableFunction.cs) - [x] HLine - [PlottableAxisLine.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisLine.cs) - [x] HSpan - [PlottableAxisSpan.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisSpan.cs) - [x] VLine - [PlottableAxisLine.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisLine.cs) - [x] VSpan - [PlottableAxisSpan.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableAxisSpan.cs) - [x] Heatmap - [PlottableHeatmap.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableHeatmap.cs) - [x] Image - [PlottableImage.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableImage.cs) - [x] Pie - [PlottablePie](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePie.cs) - [x] Polygon - [PlottablePolygon.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePolygon.cs) - [x] Polygons - [PlottablePolygons.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePolygons.cs) - [x] Populations - [PlottablePopulations.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottablePopulations.cs) - [x] Radar - [PlottableRadar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableRadar.cs) - [x] ScaleBar - [PlottableScaleBar.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScaleBar.cs) - [x] Scatter / ScatterHighlight - [PlottableScatter.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScatter.cs) [PlottableScatterHighlight.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableScatterHighlight.cs) - [x] Signal / SignalConst / SignalXY (#585) - [x] Text - [PlottableText.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableText.cs) - [x] VectorField - [PlottableVectorField.cs](https://github.com/swharden/ScottPlot/blob/master/src/ScottPlot/plottables/PlottableVectorField.cs) ### Renderables to Refactor - [x] Axis labels - [x] Axis tick marks - [x] Axis tick labels - [x] Grid lines - [x] Figure and Data background - [x] Error message - [x] Benchmark message ### Refactor / Large Features - [x] Delete the `Plottable` class and replace it with `IPlottable` - [x] Change plottable namespaces (`PlottableScatter` becomes `Plottable.Scatter`) - [x] support multiple X and Y axes ✨ ### Remaining Tasks - [x] Make GetAxisLimits() return AxisLimits object - [x] Create a control to test a timer-based render method - [x] Create new plottables specifically for growing data (mark public data fields readonly and provide setters to replace data arrays) - [x] Improve data validation for plottables (create separate pre-render and deep validation methods in an interface) ### Extra release of 4.0 - [x] Add `RenderLock()` and `RenderUnlock()` fields to help with multi-threaded plot manipulation - [x] Add NuGet message noting transition to 4.1 and to check preleases - [x] Review all the items on #412 since many have now been completed or made obsolete
priority
road to scottplot update on november ⚠️ i am reducing effort on issues and prs while i work on this daily see for a summary of major changes see for a big picture discussion of plans and goals i merged which accomplishes most of the goals described on this page the master branch is now scottplot beta old source code has its own branch plottables to refactor candlestick ohlc annotation bar errorbar function hline hspan vline vspan heatmap image pie polygon polygons populations radar scalebar scatter scatterhighlight signal signalconst signalxy text vectorfield renderables to refactor axis labels axis tick marks axis tick labels grid lines figure and data background error message benchmark message refactor large features delete the plottable class and replace it with iplottable change plottable namespaces plottablescatter becomes plottable scatter support multiple x and y axes ✨ remaining tasks make getaxislimits return axislimits object create a control to test a timer based render method create new plottables specifically for growing data mark public data fields readonly and provide setters to replace data arrays improve data validation for plottables create separate pre render and deep validation methods in an interface extra release of add renderlock and renderunlock fields to help with multi threaded plot manipulation add nuget message noting transition to and to check preleases review all the items on since many have now been completed or made obsolete
1
522,217
15,158,189,119
IssuesEvent
2021-02-12 00:32:28
NOAA-GSL/MATS
https://api.github.com/repos/NOAA-GSL/MATS
closed
METexpress surface app does not plot 2 curves correctly
Priority: High Project: MATS Status: Closed Type: Bug
--- Author Name: **bonny.strong** (@bonnystrong) Original Redmine Issue: 64018, https://vlab.ncep.noaa.gov/redmine/issues/64018 Original Date: 2019-05-15 Original Assignee: molly.b.smith --- I tried plotting 2 different curves from different MET databases. I'm attaching screenshots where I plotted each curve separately, then when I tried to plot both together. They have data over different time intervals, but when I did the plot together I changed the time interval to a range that included both. But the plot changed the time to only that of the second curve, then reported no data for the first curve. Also, in producing this issue, I've noticed that there is nothing on the graph itself that identifies the database that was used. I think we need to have this, or the titles are not definitive. --- - [METe-srf-plot1-defaults.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22717/METe-srf-plot1-defaults.JPG) (bonny.strong) - [METe-src-plot2.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22718/METe-src-plot2.JPG) (bonny.strong) - [METe-src-plot1and2.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22719/METe-src-plot1and2.JPG) (bonny.strong) - [Screen Shot 2019-05-21 at 12.00.42 PM.png](https://vlab.ncep.noaa.gov/redmine/attachments/download/22810/Screen%20Shot%202019-05-21%20at%2012.00.42%20PM.png) (molly.b.smith)
1.0
METexpress surface app does not plot 2 curves correctly - --- Author Name: **bonny.strong** (@bonnystrong) Original Redmine Issue: 64018, https://vlab.ncep.noaa.gov/redmine/issues/64018 Original Date: 2019-05-15 Original Assignee: molly.b.smith --- I tried plotting 2 different curves from different MET databases. I'm attaching screenshots where I plotted each curve separately, then when I tried to plot both together. They have data over different time intervals, but when I did the plot together I changed the time interval to a range that included both. But the plot changed the time to only that of the second curve, then reported no data for the first curve. Also, in producing this issue, I've noticed that there is nothing on the graph itself that identifies the database that was used. I think we need to have this, or the titles are not definitive. --- - [METe-srf-plot1-defaults.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22717/METe-srf-plot1-defaults.JPG) (bonny.strong) - [METe-src-plot2.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22718/METe-src-plot2.JPG) (bonny.strong) - [METe-src-plot1and2.JPG](https://vlab.ncep.noaa.gov/redmine/attachments/download/22719/METe-src-plot1and2.JPG) (bonny.strong) - [Screen Shot 2019-05-21 at 12.00.42 PM.png](https://vlab.ncep.noaa.gov/redmine/attachments/download/22810/Screen%20Shot%202019-05-21%20at%2012.00.42%20PM.png) (molly.b.smith)
priority
metexpress surface app does not plot curves correctly author name bonny strong bonnystrong original redmine issue original date original assignee molly b smith i tried plotting different curves from different met databases i m attaching screenshots where i plotted each curve separately then when i tried to plot both together they have data over different time intervals but when i did the plot together i changed the time interval to a range that included both but the plot changed the time to only that of the second curve then reported no data for the first curve also in producing this issue i ve noticed that there is nothing on the graph itself that identifies the database that was used i think we need to have this or the titles are not definitive bonny strong bonny strong bonny strong molly b smith
1
73,000
3,398,554,695
IssuesEvent
2015-12-02 04:53:45
R4stl1n/allianceauth
https://api.github.com/repos/R4stl1n/allianceauth
closed
Empty EveCorperationInfo query result breaks the HR app
bug High Priority
Probably need to put a try/catch in there. http://i.imgur.com/Z6K5eSS.png
1.0
Empty EveCorperationInfo query result breaks the HR app - Probably need to put a try/catch in there. http://i.imgur.com/Z6K5eSS.png
priority
empty evecorperationinfo query result breaks the hr app probably need to put a try catch in there
1
237,779
7,764,155,657
IssuesEvent
2018-06-01 19:12:00
fedora-infra/bodhi
https://api.github.com/repos/fedora-infra/bodhi
closed
Bugs should only be modified if they are Fedora/EPEL-only
High priority RFE bugzilla
Bodhi occasionally tries to update bugs it should not, such as this one: https://bugzilla.redhat.com/show_bug.cgi?id=1383657#c6 This is a top-level CVE bug and its state should not be modified by Bodhi since it tracks the issue in products other than Fedora and EPEL. Bodhi could simply check that the component is not 'vulnerability' and product 'Security Response' and skip making any state changes if it is.
1.0
Bugs should only be modified if they are Fedora/EPEL-only - Bodhi occasionally tries to update bugs it should not, such as this one: https://bugzilla.redhat.com/show_bug.cgi?id=1383657#c6 This is a top-level CVE bug and its state should not be modified by Bodhi since it tracks the issue in products other than Fedora and EPEL. Bodhi could simply check that the component is not 'vulnerability' and product 'Security Response' and skip making any state changes if it is.
priority
bugs should only be modified if they are fedora epel only bodhi occasionally tries to update bugs it should not such as this one this is a top level cve bug and its state should not be modified by bodhi since it tracks the issue in products other than fedora and epel bodhi could simply check that the component is not vulnerability and product security response and skip making any state changes if it is
1
358,897
10,651,603,088
IssuesEvent
2019-10-17 10:45:06
AY1920S1-CS2113T-F11-1/main
https://api.github.com/repos/AY1920S1-CS2113T-F11-1/main
closed
Enhance Reminders for Duke to fit requirements for SpongeBob
priority.High type.Task
Do things like adding reminding user of events and deadlines in the next one week (can modify this depending on your requirement)
1.0
Enhance Reminders for Duke to fit requirements for SpongeBob - Do things like adding reminding user of events and deadlines in the next one week (can modify this depending on your requirement)
priority
enhance reminders for duke to fit requirements for spongebob do things like adding reminding user of events and deadlines in the next one week can modify this depending on your requirement
1
690,454
23,660,592,958
IssuesEvent
2022-08-26 15:12:25
python/mypy
https://api.github.com/repos/python/mypy
closed
Regression on `master`: mypy unconditionally crashes when used with `--enable-error-code` for multiple error codes
crash priority-0-high
**Crash Report** On the `master` branch, mypy unconditionally crashes if used with the command-line options `--enable-error-code ignore-without-code`. I've bisected the regression to: - #13502 **Traceback** ``` Traceback (most recent call last): File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 86, in _run_code exec(code, run_globals) File "C:\Users\alexw\coding\mypy\venv\Scripts\mypy.exe\__main__.py", line 7, in <module> File "C:\Users\alexw\coding\mypy\mypy\__main__.py", line 15, in console_entry main() File "C:\Users\alexw\coding\mypy\mypy\main.py", line 95, in main res, messages, blockers = run_build(sources, options, fscache, t0, stdout, stderr) File "C:\Users\alexw\coding\mypy\mypy\main.py", line 174, in run_build res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 186, in build result = _build( File "C:\Users\alexw\coding\mypy\mypy\build.py", line 269, in _build graph = dispatch(sources, manager, stdout) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 2873, in dispatch process_graph(graph, manager) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 3257, in process_graph process_stale_scc(graph, scc, manager) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 3378, in process_stale_scc graph[id].write_cache() File "C:\Users\alexw\coding\mypy\mypy\build.py", line 2447, in write_cache new_interface_hash, self.meta = write_cache( File "C:\Users\alexw\coding\mypy\mypy\build.py", line 1639, in write_cache meta_str = json_dumps(meta, manager.options.debug_cache) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 1510, in json_dumps return json.dumps(obj, sort_keys=True) File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\__init__.py", line 238, in dumps **kw).encode(obj) File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 199, in encode chunks = self.iterencode(o, _one_shot=True) File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 257, in iterencode return _iterencode(o, 0) File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 179, in default raise TypeError(f'Object of type {o.__class__.__name__} ' TypeError: Object of type ErrorCode is not JSON serializable ``` **To Reproduce** - `cd` into a local clone of mypy, activate a local environment with an editable install of mypy. - Checkout the `master` branch. - Run `mypy --enable-error-code ignore-without-code` on any repository. I've reproduced it using the selfcheck (`python -m mypy --config-file mypy_self_check.ini -p mypy --enable-error-code ignore-without-code`) and also on my local clone of `flake8-pyi`, which has the `ignore-without-code` error code enabled as part of its default mypy configuration.
1.0
Regression on `master`: mypy unconditionally crashes when used with `--enable-error-code` for multiple error codes - **Crash Report** On the `master` branch, mypy unconditionally crashes if used with the command-line options `--enable-error-code ignore-without-code`. I've bisected the regression to: - #13502 **Traceback** ``` Traceback (most recent call last): File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 86, in _run_code exec(code, run_globals) File "C:\Users\alexw\coding\mypy\venv\Scripts\mypy.exe\__main__.py", line 7, in <module> File "C:\Users\alexw\coding\mypy\mypy\__main__.py", line 15, in console_entry main() File "C:\Users\alexw\coding\mypy\mypy\main.py", line 95, in main res, messages, blockers = run_build(sources, options, fscache, t0, stdout, stderr) File "C:\Users\alexw\coding\mypy\mypy\main.py", line 174, in run_build res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 186, in build result = _build( File "C:\Users\alexw\coding\mypy\mypy\build.py", line 269, in _build graph = dispatch(sources, manager, stdout) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 2873, in dispatch process_graph(graph, manager) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 3257, in process_graph process_stale_scc(graph, scc, manager) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 3378, in process_stale_scc graph[id].write_cache() File "C:\Users\alexw\coding\mypy\mypy\build.py", line 2447, in write_cache new_interface_hash, self.meta = write_cache( File "C:\Users\alexw\coding\mypy\mypy\build.py", line 1639, in write_cache meta_str = json_dumps(meta, manager.options.debug_cache) File "C:\Users\alexw\coding\mypy\mypy\build.py", line 1510, in json_dumps return json.dumps(obj, sort_keys=True) File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\__init__.py", line 238, in dumps **kw).encode(obj) File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 199, in encode chunks = self.iterencode(o, _one_shot=True) File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 257, in iterencode return _iterencode(o, 0) File "C:\Users\alexw\AppData\Local\Programs\Python\Python310\lib\json\encoder.py", line 179, in default raise TypeError(f'Object of type {o.__class__.__name__} ' TypeError: Object of type ErrorCode is not JSON serializable ``` **To Reproduce** - `cd` into a local clone of mypy, activate a local environment with an editable install of mypy. - Checkout the `master` branch. - Run `mypy --enable-error-code ignore-without-code` on any repository. I've reproduced it using the selfcheck (`python -m mypy --config-file mypy_self_check.ini -p mypy --enable-error-code ignore-without-code`) and also on my local clone of `flake8-pyi`, which has the `ignore-without-code` error code enabled as part of its default mypy configuration.
priority
regression on master mypy unconditionally crashes when used with enable error code for multiple error codes crash report on the master branch mypy unconditionally crashes if used with the command line options enable error code ignore without code i ve bisected the regression to traceback traceback most recent call last file c users alexw appdata local programs python lib runpy py line in run module as main return run code code main globals none file c users alexw appdata local programs python lib runpy py line in run code exec code run globals file c users alexw coding mypy venv scripts mypy exe main py line in file c users alexw coding mypy mypy main py line in console entry main file c users alexw coding mypy mypy main py line in main res messages blockers run build sources options fscache stdout stderr file c users alexw coding mypy mypy main py line in run build res build build sources options none flush errors fscache stdout stderr file c users alexw coding mypy mypy build py line in build result build file c users alexw coding mypy mypy build py line in build graph dispatch sources manager stdout file c users alexw coding mypy mypy build py line in dispatch process graph graph manager file c users alexw coding mypy mypy build py line in process graph process stale scc graph scc manager file c users alexw coding mypy mypy build py line in process stale scc graph write cache file c users alexw coding mypy mypy build py line in write cache new interface hash self meta write cache file c users alexw coding mypy mypy build py line in write cache meta str json dumps meta manager options debug cache file c users alexw coding mypy mypy build py line in json dumps return json dumps obj sort keys true file c users alexw appdata local programs python lib json init py line in dumps kw encode obj file c users alexw appdata local programs python lib json encoder py line in encode chunks self iterencode o one shot true file c users alexw appdata local programs python lib json encoder py line in iterencode return iterencode o file c users alexw appdata local programs python lib json encoder py line in default raise typeerror f object of type o class name typeerror object of type errorcode is not json serializable to reproduce cd into a local clone of mypy activate a local environment with an editable install of mypy checkout the master branch run mypy enable error code ignore without code on any repository i ve reproduced it using the selfcheck python m mypy config file mypy self check ini p mypy enable error code ignore without code and also on my local clone of pyi which has the ignore without code error code enabled as part of its default mypy configuration
1
239,807
7,800,063,868
IssuesEvent
2018-06-09 04:14:23
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
closed
0006880: it should be possible to save drafts/templates without subject/recipients
Felamimail Mantis high priority
**Reported by pschuele on 3 Aug 2012 20:08** **Version:** Milan (2012.03.5) it should be possible to save drafts/templates without subject/recipients
1.0
0006880: it should be possible to save drafts/templates without subject/recipients - **Reported by pschuele on 3 Aug 2012 20:08** **Version:** Milan (2012.03.5) it should be possible to save drafts/templates without subject/recipients
priority
it should be possible to save drafts templates without subject recipients reported by pschuele on aug version milan it should be possible to save drafts templates without subject recipients
1
328,254
9,991,754,586
IssuesEvent
2019-07-11 11:55:43
telstra/open-kilda
https://api.github.com/repos/telstra/open-kilda
closed
[Stats] As a user I can see table stats in OpenTSDB
area/stats feature priority/2-high
In [statsresponse](http://flowgrammable.org/sdn/openflow/message-layer/statsresponse/#ofp_1_3_0) Table message wee need to collect `active-count`, `lookup-count` and `matched-count` per table in a same cadense we do flow stats. Also, we should record a diff: `lookup-count` minus `matched-count`. Matric name should be `missed` metric details: - name: `switch.table.*` without `-count` (e.g. `switch.table.active`, etc.) - `switch.table.active` - `switch.table.lookup` - `switch.table.matched` - `switch.table.missed` - tags: switchid, tableid
1.0
[Stats] As a user I can see table stats in OpenTSDB - In [statsresponse](http://flowgrammable.org/sdn/openflow/message-layer/statsresponse/#ofp_1_3_0) Table message wee need to collect `active-count`, `lookup-count` and `matched-count` per table in a same cadense we do flow stats. Also, we should record a diff: `lookup-count` minus `matched-count`. Matric name should be `missed` metric details: - name: `switch.table.*` without `-count` (e.g. `switch.table.active`, etc.) - `switch.table.active` - `switch.table.lookup` - `switch.table.matched` - `switch.table.missed` - tags: switchid, tableid
priority
as a user i can see table stats in opentsdb in table message wee need to collect active count lookup count and matched count per table in a same cadense we do flow stats also we should record a diff lookup count minus matched count matric name should be missed metric details name switch table without count e g switch table active etc switch table active switch table lookup switch table matched switch table missed tags switchid tableid
1
109,777
4,413,816,943
IssuesEvent
2016-08-13 02:32:14
thalida/thalida.com
https://api.github.com/repos/thalida/thalida.com
closed
Selected highlighted text doesn't show selected state.
platform: all priority: high serverity: bug status: ready
<img width="503" alt="screen shot 2016-08-12 at 3 26 12 pm" src="https://cloud.githubusercontent.com/assets/3401715/17634539/22b0d216-60a1-11e6-8553-d9b535b2fd54.png">
1.0
Selected highlighted text doesn't show selected state. - <img width="503" alt="screen shot 2016-08-12 at 3 26 12 pm" src="https://cloud.githubusercontent.com/assets/3401715/17634539/22b0d216-60a1-11e6-8553-d9b535b2fd54.png">
priority
selected highlighted text doesn t show selected state img width alt screen shot at pm src
1
61,994
3,164,000,989
IssuesEvent
2015-09-20 20:12:59
audreyr/cookiecutter
https://api.github.com/repos/audreyr/cookiecutter
closed
Out of date Click dependency
high-priority
Click is up to version 5.1. Should cookiecutter support the newer release of Click?
1.0
Out of date Click dependency - Click is up to version 5.1. Should cookiecutter support the newer release of Click?
priority
out of date click dependency click is up to version should cookiecutter support the newer release of click
1
341,471
10,297,895,466
IssuesEvent
2019-08-28 09:59:04
UniversityOfHelsinkiCS/fuksilaiterekisteri
https://api.github.com/repos/UniversityOfHelsinkiCS/fuksilaiterekisteri
closed
prepare registration for oodi downtime exceptions
enhancement high priority
when oodi is down for maintenance, registration is 'down for maintenance' next downtime is (presumably) on 1st of Sep
1.0
prepare registration for oodi downtime exceptions - when oodi is down for maintenance, registration is 'down for maintenance' next downtime is (presumably) on 1st of Sep
priority
prepare registration for oodi downtime exceptions when oodi is down for maintenance registration is down for maintenance next downtime is presumably on of sep
1
716,126
24,622,255,598
IssuesEvent
2022-10-16 03:41:26
HaDuve/TravelCostNative
https://api.github.com/repos/HaDuve/TravelCostNative
closed
Fix the bug about the expense context time frame
Bug 1 - High Priority
It should be today, not last 24 hours. Same with week month etc.
1.0
Fix the bug about the expense context time frame - It should be today, not last 24 hours. Same with week month etc.
priority
fix the bug about the expense context time frame it should be today not last hours same with week month etc
1
602,419
18,468,743,286
IssuesEvent
2021-10-17 11:11:30
AY2122S1-CS2103T-F13-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-F13-2/tp
opened
Modify Help Command
priority.High type.Story
As a nurse, I can view all commands available to me, so that I can better take advantage of the full functionality of the application.
1.0
Modify Help Command - As a nurse, I can view all commands available to me, so that I can better take advantage of the full functionality of the application.
priority
modify help command as a nurse i can view all commands available to me so that i can better take advantage of the full functionality of the application
1
241,834
7,834,916,458
IssuesEvent
2018-06-16 20:19:10
alakajam-team/alakajam
https://api.github.com/repos/alakajam-team/alakajam
closed
"Save comment & ratings" only saves the ratings if there is a comment
bug high priority
* `npm run watch` * Go to http://localhost:8000/2nd-alakajam/11/game-9/ * Put 10 stars on Overall * Click "Save comment & ratings" Expected: ratings are saved and we see 10 stars. Actual: ratings are not saved, we still see empty stars. If a nonempty comment is supplied, both are saved correctly.
1.0
"Save comment & ratings" only saves the ratings if there is a comment - * `npm run watch` * Go to http://localhost:8000/2nd-alakajam/11/game-9/ * Put 10 stars on Overall * Click "Save comment & ratings" Expected: ratings are saved and we see 10 stars. Actual: ratings are not saved, we still see empty stars. If a nonempty comment is supplied, both are saved correctly.
priority
save comment ratings only saves the ratings if there is a comment npm run watch go to put stars on overall click save comment ratings expected ratings are saved and we see stars actual ratings are not saved we still see empty stars if a nonempty comment is supplied both are saved correctly
1
83,995
3,645,586,391
IssuesEvent
2016-02-15 15:16:08
onyxfish/csvkit
https://api.github.com/repos/onyxfish/csvkit
closed
Unicode issues causing tests to fail in Python 2
High Priority
I changed a passing test to read from a file containing a UTF-8 character, and now tests are failing. Not sure how to correct. See https://travis-ci.org/onyxfish/csvkit/jobs/108149327 /cc @onyxfish
1.0
Unicode issues causing tests to fail in Python 2 - I changed a passing test to read from a file containing a UTF-8 character, and now tests are failing. Not sure how to correct. See https://travis-ci.org/onyxfish/csvkit/jobs/108149327 /cc @onyxfish
priority
unicode issues causing tests to fail in python i changed a passing test to read from a file containing a utf character and now tests are failing not sure how to correct see cc onyxfish
1
559,498
16,564,721,620
IssuesEvent
2021-05-29 06:45:17
hatnote/montage
https://api.github.com/repos/hatnote/montage
closed
Unable to run Montage in Firefox
priority high type bug
On clicking the link 'Log in using Wikimedia account', Firefox throws the error message: **Internal server error** _<ExceptionInfo [AttributeError: 'JSONCookie' object has no attribute 'set_expires'] (27 frames, last=Callpoint('complete_login', 162, 'montage.public_endpoints', './montage/public_endpoints.py', 121, ' cookie.set_expires()'))> Error type: http://docs.python.org/2/library/exceptions.html#exceptions.AttributeError_ Same result on several machines. Firefox 62.0.3 for Mac.
1.0
Unable to run Montage in Firefox - On clicking the link 'Log in using Wikimedia account', Firefox throws the error message: **Internal server error** _<ExceptionInfo [AttributeError: 'JSONCookie' object has no attribute 'set_expires'] (27 frames, last=Callpoint('complete_login', 162, 'montage.public_endpoints', './montage/public_endpoints.py', 121, ' cookie.set_expires()'))> Error type: http://docs.python.org/2/library/exceptions.html#exceptions.AttributeError_ Same result on several machines. Firefox 62.0.3 for Mac.
priority
unable to run montage in firefox on clicking the link log in using wikimedia account firefox throws the error message internal server error error type same result on several machines firefox for mac
1
109,080
4,369,800,252
IssuesEvent
2016-08-04 02:07:45
jsdsa/jsdsa
https://api.github.com/repos/jsdsa/jsdsa
closed
List of sorting algorithms
Algorithms High priority
### Simple Sorts - [x] Insertion sort - [x] Selection sort ### Efficient Sorts - [x] Heap sort - [x] Merge sort - [x] Quick sort ### Bubble sort and variants - [x] Bubble sort - [x] Comb sort - [x] Shell sort ### Distribution sort - [x] Bucket sort - [x] Counting sort - [x] Radix sort Optimizations, alternatives, variants to be added later.
1.0
List of sorting algorithms - ### Simple Sorts - [x] Insertion sort - [x] Selection sort ### Efficient Sorts - [x] Heap sort - [x] Merge sort - [x] Quick sort ### Bubble sort and variants - [x] Bubble sort - [x] Comb sort - [x] Shell sort ### Distribution sort - [x] Bucket sort - [x] Counting sort - [x] Radix sort Optimizations, alternatives, variants to be added later.
priority
list of sorting algorithms simple sorts insertion sort selection sort efficient sorts heap sort merge sort quick sort bubble sort and variants bubble sort comb sort shell sort distribution sort bucket sort counting sort radix sort optimizations alternatives variants to be added later
1
335,735
10,165,696,369
IssuesEvent
2019-08-07 14:23:44
OpenLiveWriter/OpenLiveWriter
https://api.github.com/repos/OpenLiveWriter/OpenLiveWriter
closed
CI builds continuously spawn new instances
bug high-priority
Builds from AppVeyor continuously spawn new instances of themselves. Renaming Update.exe and starting OpenLiveWriter.exe directly seems workaround the issue, however this is a considerable roadblock for many users that wish to try out dev builds. OpenLiveWriter.ApplicationMain.LaunchAdditionalInstance seems to be called each time an extra instance is spawned. Below is a stack trace of one of these calls; ``` at OpenLiveWriter.ApplicationMain.LaunchAdditionalInstance(String[] args) at OpenLiveWriter.ApplicationMain.LaunchAction(String[] args, Boolean isFirstInstance) at OpenLiveWriter.CoreServices.SingleInstanceApplicationManager.LaunchActionThreadWithState.ThreadProc() at System.Threading.ThreadHelper.ThreadStart_Context(Object state) at System.Threading.ExecutionContext.RunInternal(ExecutionContext executionContext, ContextCallback callback, Object state, Boolean preserveSyncCtx) at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state, Boolean preserveSyncCtx) at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state) at System.Threading.ThreadHelper.ThreadStart() ``` ## Workarounds Numerous workarounds have been outlined in issue #786 for specifically installing the #810 fix. Installing a recent dev build would involve applying these instructions with a more recent build artifact rather than the build for #810. - http://kgbthatsme.blogspot.com/2019/05/open-live-writer-works-again-with.html - http://www.thanislim.com/2019/03/how-to-fix-open-live-writer-error-400.html In summary: 1. Rename `%localappdata%\OpenLiveWriter\Update.exe` to another file name. 2. Change all Open Live Writer shortcuts to point towards `%localappdata%\OpenLiveWriter\app-x.y.z\OpenLiveWriter.exe` where x.y.z is the installed app version. Please keep in mind that these workarounds break auto-updating.
1.0
CI builds continuously spawn new instances - Builds from AppVeyor continuously spawn new instances of themselves. Renaming Update.exe and starting OpenLiveWriter.exe directly seems workaround the issue, however this is a considerable roadblock for many users that wish to try out dev builds. OpenLiveWriter.ApplicationMain.LaunchAdditionalInstance seems to be called each time an extra instance is spawned. Below is a stack trace of one of these calls; ``` at OpenLiveWriter.ApplicationMain.LaunchAdditionalInstance(String[] args) at OpenLiveWriter.ApplicationMain.LaunchAction(String[] args, Boolean isFirstInstance) at OpenLiveWriter.CoreServices.SingleInstanceApplicationManager.LaunchActionThreadWithState.ThreadProc() at System.Threading.ThreadHelper.ThreadStart_Context(Object state) at System.Threading.ExecutionContext.RunInternal(ExecutionContext executionContext, ContextCallback callback, Object state, Boolean preserveSyncCtx) at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state, Boolean preserveSyncCtx) at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state) at System.Threading.ThreadHelper.ThreadStart() ``` ## Workarounds Numerous workarounds have been outlined in issue #786 for specifically installing the #810 fix. Installing a recent dev build would involve applying these instructions with a more recent build artifact rather than the build for #810. - http://kgbthatsme.blogspot.com/2019/05/open-live-writer-works-again-with.html - http://www.thanislim.com/2019/03/how-to-fix-open-live-writer-error-400.html In summary: 1. Rename `%localappdata%\OpenLiveWriter\Update.exe` to another file name. 2. Change all Open Live Writer shortcuts to point towards `%localappdata%\OpenLiveWriter\app-x.y.z\OpenLiveWriter.exe` where x.y.z is the installed app version. Please keep in mind that these workarounds break auto-updating.
priority
ci builds continuously spawn new instances builds from appveyor continuously spawn new instances of themselves renaming update exe and starting openlivewriter exe directly seems workaround the issue however this is a considerable roadblock for many users that wish to try out dev builds openlivewriter applicationmain launchadditionalinstance seems to be called each time an extra instance is spawned below is a stack trace of one of these calls at openlivewriter applicationmain launchadditionalinstance string args at openlivewriter applicationmain launchaction string args boolean isfirstinstance at openlivewriter coreservices singleinstanceapplicationmanager launchactionthreadwithstate threadproc at system threading threadhelper threadstart context object state at system threading executioncontext runinternal executioncontext executioncontext contextcallback callback object state boolean preservesyncctx at system threading executioncontext run executioncontext executioncontext contextcallback callback object state boolean preservesyncctx at system threading executioncontext run executioncontext executioncontext contextcallback callback object state at system threading threadhelper threadstart workarounds numerous workarounds have been outlined in issue for specifically installing the fix installing a recent dev build would involve applying these instructions with a more recent build artifact rather than the build for in summary rename localappdata openlivewriter update exe to another file name change all open live writer shortcuts to point towards localappdata openlivewriter app x y z openlivewriter exe where x y z is the installed app version please keep in mind that these workarounds break auto updating
1
397,338
11,727,101,959
IssuesEvent
2020-03-10 15:27:43
AY1920S2-CS2103T-W16-2/main
https://api.github.com/repos/AY1920S2-CS2103T-W16-2/main
opened
As a exerciser I want to see a history of all my past exercises
core priority.High type.Epic
so I can see how regularly I have been exercising
1.0
As a exerciser I want to see a history of all my past exercises - so I can see how regularly I have been exercising
priority
as a exerciser i want to see a history of all my past exercises so i can see how regularly i have been exercising
1
375,200
11,101,396,505
IssuesEvent
2019-12-16 21:20:53
carbon-design-system/design-language-website
https://api.github.com/repos/carbon-design-system/design-language-website
closed
Color picker down. The tabs do not change color choices
Severity 1 🚨 priority: high type: bug 🐛
the screen goes black when selecting PMS, RGB or CMYK. https://www.ibm.com/design/language/color/#specifications
1.0
Color picker down. The tabs do not change color choices - the screen goes black when selecting PMS, RGB or CMYK. https://www.ibm.com/design/language/color/#specifications
priority
color picker down the tabs do not change color choices the screen goes black when selecting pms rgb or cmyk
1
97,509
3,994,729,999
IssuesEvent
2016-05-10 13:27:46
GoogleCloudPlatform/gcloud-eclipse-tools
https://api.github.com/repos/GoogleCloudPlatform/gcloud-eclipse-tools
opened
call setErrorMessage when project creation fails in StandardProjectWizard
enhancement high priority
find // todo if fail, call setErrorMessage() note that this happens a lot while testing new code, adn we need it fo that if nothing else.
1.0
call setErrorMessage when project creation fails in StandardProjectWizard - find // todo if fail, call setErrorMessage() note that this happens a lot while testing new code, adn we need it fo that if nothing else.
priority
call seterrormessage when project creation fails in standardprojectwizard find todo if fail call seterrormessage note that this happens a lot while testing new code adn we need it fo that if nothing else
1
505,795
14,645,854,324
IssuesEvent
2020-12-26 10:34:14
GrottoCenter/Grottocenter3
https://api.github.com/repos/GrottoCenter/Grottocenter3
closed
Magazine
Priority: High Type: Bug
Vincent a créé un type Magazine alors que dans le modèle nous avons Collection qui correspond au type prévu dans DCMI. Il faudrait affecter l'id de Collection à tout ce qui est déclaré comme Magazine. C'est relativement urgent car les utilisateurs sont en train de recréer les Collections.
1.0
Magazine - Vincent a créé un type Magazine alors que dans le modèle nous avons Collection qui correspond au type prévu dans DCMI. Il faudrait affecter l'id de Collection à tout ce qui est déclaré comme Magazine. C'est relativement urgent car les utilisateurs sont en train de recréer les Collections.
priority
magazine vincent a créé un type magazine alors que dans le modèle nous avons collection qui correspond au type prévu dans dcmi il faudrait affecter l id de collection à tout ce qui est déclaré comme magazine c est relativement urgent car les utilisateurs sont en train de recréer les collections
1
525,130
15,238,489,435
IssuesEvent
2021-02-19 02:03:08
TerriaJS/nationalmap
https://api.github.com/repos/TerriaJS/nationalmap
closed
Add ~globalDisclaimer~ welcome message to v8
High priority
This is about welcome message - not globalDisclaimer ~This is what we have for DEA, so follow the format and change the message:~ ```json "globalDisclaimer": { "buttonTitle": "I agree", "confirmationRequired": true, "devHostRegex": "\\b(staging|preview|test|dev)\\.", "enableOnLocalhost": true, "message": "The information displayed on the DEA Maps (the \"Service\") is for general informational purposes only, and is not intended to provide any commercial, financial, or legal advice.\n\nThe Service provides access to a range of data sets and tools, some of which are in development (beta). All data and tools are provided on an \"as is\" and \"with all faults\" basis without any warranty whatsoever. Geoscience Australia and Data61 do not warrant that these data sets and tools shall meet any requirements or expectations, or that the will be fit for any intended purposes.\n\nGeoscience Australia and Data61 assumes no responsibility for errors or omissions in the contents of the Service and reserves the right to make additions, deletions, or modification to the contents on the Service at any time without prior notice.\n\nGeoscience Australia and Data61 does not guarantee the accuracy, relevance, timeliness, or completeness of any information or data available through the Service or on linked external websites.\n\n[See full terms and conditions here.](./about#terms-and-conditions)", "prodHostRegex": "prod\\.saas\\.terria\\.io$", "title": "Disclaimer" }, ```
1.0
Add ~globalDisclaimer~ welcome message to v8 - This is about welcome message - not globalDisclaimer ~This is what we have for DEA, so follow the format and change the message:~ ```json "globalDisclaimer": { "buttonTitle": "I agree", "confirmationRequired": true, "devHostRegex": "\\b(staging|preview|test|dev)\\.", "enableOnLocalhost": true, "message": "The information displayed on the DEA Maps (the \"Service\") is for general informational purposes only, and is not intended to provide any commercial, financial, or legal advice.\n\nThe Service provides access to a range of data sets and tools, some of which are in development (beta). All data and tools are provided on an \"as is\" and \"with all faults\" basis without any warranty whatsoever. Geoscience Australia and Data61 do not warrant that these data sets and tools shall meet any requirements or expectations, or that the will be fit for any intended purposes.\n\nGeoscience Australia and Data61 assumes no responsibility for errors or omissions in the contents of the Service and reserves the right to make additions, deletions, or modification to the contents on the Service at any time without prior notice.\n\nGeoscience Australia and Data61 does not guarantee the accuracy, relevance, timeliness, or completeness of any information or data available through the Service or on linked external websites.\n\n[See full terms and conditions here.](./about#terms-and-conditions)", "prodHostRegex": "prod\\.saas\\.terria\\.io$", "title": "Disclaimer" }, ```
priority
add globaldisclaimer welcome message to this is about welcome message not globaldisclaimer this is what we have for dea so follow the format and change the message json globaldisclaimer buttontitle i agree confirmationrequired true devhostregex b staging preview test dev enableonlocalhost true message the information displayed on the dea maps the service is for general informational purposes only and is not intended to provide any commercial financial or legal advice n nthe service provides access to a range of data sets and tools some of which are in development beta all data and tools are provided on an as is and with all faults basis without any warranty whatsoever geoscience australia and do not warrant that these data sets and tools shall meet any requirements or expectations or that the will be fit for any intended purposes n ngeoscience australia and assumes no responsibility for errors or omissions in the contents of the service and reserves the right to make additions deletions or modification to the contents on the service at any time without prior notice n ngeoscience australia and does not guarantee the accuracy relevance timeliness or completeness of any information or data available through the service or on linked external websites n n about terms and conditions prodhostregex prod saas terria io title disclaimer
1
339,397
10,253,739,983
IssuesEvent
2019-08-21 12:03:30
DimensionDev/Maskbook
https://api.github.com/repos/DimensionDev/Maskbook
opened
ux: new url permission scheme does not work with profile import
Priority: P2 (Most users) Severity: High Type: Bug Type: UI
importing an old profile does not trigger the webpage access permission request.
1.0
ux: new url permission scheme does not work with profile import - importing an old profile does not trigger the webpage access permission request.
priority
ux new url permission scheme does not work with profile import importing an old profile does not trigger the webpage access permission request
1
377,488
11,171,737,926
IssuesEvent
2019-12-28 22:28:37
Thorium-Sim/thorium
https://api.github.com/repos/Thorium-Sim/thorium
opened
Dilithium Stress
priority/high type/bug
### Requested By: Your Friends from the Odyssey ### Priority: High ### Version: 2.1.0 The dilithium stress still doesn't match between core and the operations station. If you adjust it on our end it doesn't change it on theirs and vice versa. It also looks super whack and jank. (Refer to image) ### Steps to Reproduce Put a dilithium screen in your configuration. Change the stress from your side. Observe as it does nothing (whether the flight is paused or not)
1.0
Dilithium Stress - ### Requested By: Your Friends from the Odyssey ### Priority: High ### Version: 2.1.0 The dilithium stress still doesn't match between core and the operations station. If you adjust it on our end it doesn't change it on theirs and vice versa. It also looks super whack and jank. (Refer to image) ### Steps to Reproduce Put a dilithium screen in your configuration. Change the stress from your side. Observe as it does nothing (whether the flight is paused or not)
priority
dilithium stress requested by your friends from the odyssey priority high version the dilithium stress still doesn t match between core and the operations station if you adjust it on our end it doesn t change it on theirs and vice versa it also looks super whack and jank refer to image steps to reproduce put a dilithium screen in your configuration change the stress from your side observe as it does nothing whether the flight is paused or not
1
472,561
13,627,009,296
IssuesEvent
2020-09-24 11:57:33
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] Restoring a backup (with the users not backed up) fails
bug priority: high
## Describe the bug Restoring a backup (from a version with the users not backed up) fails ## To Reproduce Steps to reproduce the behavior: 1. Create a backup using an earlier version of Crafter. For our example, we'll use 3.1.8 2. Using the 3.1.10 snapshot version of Crafter, restore the backup created using 3.1.8 3. Notice the error message `./crafter.sh: line 1126: /Users/vita/temp/test3/craftercms/crafter-authoring/mybackups/temp/users.sql: No such file or directory` ## Expected behavior The restore should work ## Screenshots {{If applicable, add screenshots to help explain your problem.}} ## Logs ``` 2020-09-22 10:42:36.717 INFO 78765 --- [ main] ch.vorburger.mariadb4j.DB : Database startup complete. 2020-09-22 10:42:37.032 INFO 78765 --- [ main] c.v.m.s.boot.MariaDB4jApplication : Started MariaDB4jApplication in 4.989 seconds (JVM running for 5.485) ------------------------------------------------------------------------ Restoring embedded DB ------------------------------------------------------------------------ ./crafter.sh: line 1126: /Users/vita/temp/test3/craftercms/crafter-authoring/mybackups/temp/users.sql: No such file or directory Unable to continue, an error occurred or the script was forcefully stopped ``` ## Specs ### Version 3.1.10 snapshot ### OS OS X ### Browser {{What browser did you use to produce the bug.}} ## Additional context {{Add any other context about the problem here.}}
1.0
[studio] Restoring a backup (with the users not backed up) fails - ## Describe the bug Restoring a backup (from a version with the users not backed up) fails ## To Reproduce Steps to reproduce the behavior: 1. Create a backup using an earlier version of Crafter. For our example, we'll use 3.1.8 2. Using the 3.1.10 snapshot version of Crafter, restore the backup created using 3.1.8 3. Notice the error message `./crafter.sh: line 1126: /Users/vita/temp/test3/craftercms/crafter-authoring/mybackups/temp/users.sql: No such file or directory` ## Expected behavior The restore should work ## Screenshots {{If applicable, add screenshots to help explain your problem.}} ## Logs ``` 2020-09-22 10:42:36.717 INFO 78765 --- [ main] ch.vorburger.mariadb4j.DB : Database startup complete. 2020-09-22 10:42:37.032 INFO 78765 --- [ main] c.v.m.s.boot.MariaDB4jApplication : Started MariaDB4jApplication in 4.989 seconds (JVM running for 5.485) ------------------------------------------------------------------------ Restoring embedded DB ------------------------------------------------------------------------ ./crafter.sh: line 1126: /Users/vita/temp/test3/craftercms/crafter-authoring/mybackups/temp/users.sql: No such file or directory Unable to continue, an error occurred or the script was forcefully stopped ``` ## Specs ### Version 3.1.10 snapshot ### OS OS X ### Browser {{What browser did you use to produce the bug.}} ## Additional context {{Add any other context about the problem here.}}
priority
restoring a backup with the users not backed up fails describe the bug restoring a backup from a version with the users not backed up fails to reproduce steps to reproduce the behavior create a backup using an earlier version of crafter for our example we ll use using the snapshot version of crafter restore the backup created using notice the error message crafter sh line users vita temp craftercms crafter authoring mybackups temp users sql no such file or directory expected behavior the restore should work screenshots if applicable add screenshots to help explain your problem logs info ch vorburger db database startup complete info c v m s boot started in seconds jvm running for restoring embedded db crafter sh line users vita temp craftercms crafter authoring mybackups temp users sql no such file or directory unable to continue an error occurred or the script was forcefully stopped specs version snapshot os os x browser what browser did you use to produce the bug additional context add any other context about the problem here
1
194,465
6,895,025,129
IssuesEvent
2017-11-23 12:17:24
ballerinalang/composer
https://api.github.com/repos/ballerinalang/composer
opened
Default config is wrong for ftp service
0.95.1 Priority/Highest Severity/Critical
Please see the below image. The default config is wrong. Current: ``` import ballerina.net.ftp; service<fs> service1 { resource echo1 (fs:FileSystemEvent m) { } } ``` Expected: ``` import ballerina.net.ftp; service<ftp> service1 { resource echo1 (ftp:FTPServerEvent m) { } } ``` ![issue16](https://user-images.githubusercontent.com/15624590/33172475-4d7b2924-d076-11e7-9c9b-6dea5d60d4fa.png)
1.0
Default config is wrong for ftp service - Please see the below image. The default config is wrong. Current: ``` import ballerina.net.ftp; service<fs> service1 { resource echo1 (fs:FileSystemEvent m) { } } ``` Expected: ``` import ballerina.net.ftp; service<ftp> service1 { resource echo1 (ftp:FTPServerEvent m) { } } ``` ![issue16](https://user-images.githubusercontent.com/15624590/33172475-4d7b2924-d076-11e7-9c9b-6dea5d60d4fa.png)
priority
default config is wrong for ftp service please see the below image the default config is wrong current import ballerina net ftp service resource fs filesystemevent m expected import ballerina net ftp service resource ftp ftpserverevent m
1
489,012
14,100,190,562
IssuesEvent
2020-11-06 03:27:34
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Unable to list Roles via console with Oracle19c
Affected/5.11.0-Beta Priority/Highest Severity/Blocker bug identity-core
**Describe the issue:** Role list will be empty with Oracle19c, in Console > Manage > Roles **How to reproduce:** 1. Configure Oracle as Primary user store 2. Login to console with superuser credentials 3. View roles (Manage) **Expected behavior:** Added roles should be display properly **Environment information** - Product Version: 5.11 Beta 3 - OS: Windows - Database: Oracle 19c - Userstore: JDBC --- ![image](https://user-images.githubusercontent.com/39120228/97624915-c8e22580-1a4d-11eb-855e-ad31d858df04.png) ![image](https://user-images.githubusercontent.com/39120228/97625177-2fffda00-1a4e-11eb-9664-0a8bc91ad6a3.png)
1.0
Unable to list Roles via console with Oracle19c - **Describe the issue:** Role list will be empty with Oracle19c, in Console > Manage > Roles **How to reproduce:** 1. Configure Oracle as Primary user store 2. Login to console with superuser credentials 3. View roles (Manage) **Expected behavior:** Added roles should be display properly **Environment information** - Product Version: 5.11 Beta 3 - OS: Windows - Database: Oracle 19c - Userstore: JDBC --- ![image](https://user-images.githubusercontent.com/39120228/97624915-c8e22580-1a4d-11eb-855e-ad31d858df04.png) ![image](https://user-images.githubusercontent.com/39120228/97625177-2fffda00-1a4e-11eb-9664-0a8bc91ad6a3.png)
priority
unable to list roles via console with describe the issue role list will be empty with in console manage roles how to reproduce configure oracle as primary user store login to console with superuser credentials view roles manage expected behavior added roles should be display properly environment information product version beta os windows database oracle userstore jdbc
1
603,974
18,674,989,732
IssuesEvent
2021-10-31 12:05:18
AY2122S1-CS2103-T14-2/tp
https://api.github.com/repos/AY2122S1-CS2103-T14-2/tp
closed
[PE-D] Snoozed reminders re-activate after closing and re-opening app.
priority.High type.Task
Steps to re-produce: 1. Add a reminder for current day 2. Snooze reminder 3. Type exit command 4. Re-open ePoch 5. Find that occurrences has been reduced, reminder not snoozed, and CCA label is not shown properly. Before re-opening: ![snooze_beforeopening.png](https://raw.githubusercontent.com/juliussneezer04/ped/main/files/359970ff-0486-42aa-a1d0-7fefd846bbc3.png) After re-opening: ![snooze_reopening.png](https://raw.githubusercontent.com/juliussneezer04/ped/main/files/b19bce5c-9cb4-46c7-afcc-aa6de740c4bd.png) <!--session: 1635494417648-992e5fad-48d0-4008-813a-369898477680--> <!--Version: Web v3.4.1--> ------------- Labels: `type.FunctionalityBug` `severity.High` original: juliussneezer04/ped#9
1.0
[PE-D] Snoozed reminders re-activate after closing and re-opening app. - Steps to re-produce: 1. Add a reminder for current day 2. Snooze reminder 3. Type exit command 4. Re-open ePoch 5. Find that occurrences has been reduced, reminder not snoozed, and CCA label is not shown properly. Before re-opening: ![snooze_beforeopening.png](https://raw.githubusercontent.com/juliussneezer04/ped/main/files/359970ff-0486-42aa-a1d0-7fefd846bbc3.png) After re-opening: ![snooze_reopening.png](https://raw.githubusercontent.com/juliussneezer04/ped/main/files/b19bce5c-9cb4-46c7-afcc-aa6de740c4bd.png) <!--session: 1635494417648-992e5fad-48d0-4008-813a-369898477680--> <!--Version: Web v3.4.1--> ------------- Labels: `type.FunctionalityBug` `severity.High` original: juliussneezer04/ped#9
priority
snoozed reminders re activate after closing and re opening app steps to re produce add a reminder for current day snooze reminder type exit command re open epoch find that occurrences has been reduced reminder not snoozed and cca label is not shown properly before re opening after re opening labels type functionalitybug severity high original ped
1
158,074
6,021,371,865
IssuesEvent
2017-06-07 18:32:36
apollographql/apollo-android
https://api.github.com/repos/apollographql/apollo-android
closed
Trying to assign a Double value into Integer field.
Priority: High Type: Bug Type: compiler
Hey, I don't know if by my error, or by some kind of bug in the compiler an erroneus code is being generated. Following is one of the errors I'm getting: `Error:(42, 39) error: incompatible types: double cannot be converted to Integer` Generated file looks like this: ``` package type; import java.lang.Boolean; import java.lang.Integer; import java.lang.String; import javax.annotation.Generated; import javax.annotation.Nullable; @Generated("Apollo GraphQL") public final class NotificationFilters { private final @Nullable Boolean isNew; private final @Nullable Integer limit; private final @Nullable String cursor; NotificationFilters(@Nullable Boolean isNew, @Nullable Integer limit, @Nullable String cursor) { this.isNew = isNew; this.limit = limit; this.cursor = cursor; } public @Nullable Boolean isNew() { return this.isNew; } public @Nullable Integer limit() { return this.limit; } public @Nullable String cursor() { return this.cursor; } public static Builder builder() { return new Builder(); } public static final class Builder { private @Nullable Boolean isNew; private @Nullable Integer limit = 10.0; private @Nullable String cursor; Builder() { } public Builder isNew(@Nullable Boolean isNew) { this.isNew = isNew; return this; } public Builder limit(@Nullable Integer limit) { this.limit = limit; return this; } public Builder cursor(@Nullable String cursor) { this.cursor = cursor; return this; } public NotificationFilters build() { return new NotificationFilters(isNew, limit, cursor); } } } ``` When you look at the generated Builder, you can indeed see that a double value is being assigned to an Integer. These errors happen for all models that have a default value set. Following is part of the schema file: ``` { "kind": "INPUT_OBJECT", "name": "NotificationFilters", "description": "", "fields": null, "inputFields": [ { "name": "isNew", "description": "", "type": { "kind": "SCALAR", "name": "Boolean", "ofType": null }, "defaultValue": null }, { "name": "limit", "description": "", "type": { "kind": "SCALAR", "name": "Int", "ofType": null }, "defaultValue": "10" }, { "name": "cursor", "description": "", "type": { "kind": "SCALAR", "name": "String", "ofType": null }, "defaultValue": null } ], "interfaces": null, "enumValues": null, "possibleTypes": null } ``` Any ideas or tips on what might be going wrong? I'm using the 0.3.1-SNAPSHOT.
1.0
Trying to assign a Double value into Integer field. - Hey, I don't know if by my error, or by some kind of bug in the compiler an erroneus code is being generated. Following is one of the errors I'm getting: `Error:(42, 39) error: incompatible types: double cannot be converted to Integer` Generated file looks like this: ``` package type; import java.lang.Boolean; import java.lang.Integer; import java.lang.String; import javax.annotation.Generated; import javax.annotation.Nullable; @Generated("Apollo GraphQL") public final class NotificationFilters { private final @Nullable Boolean isNew; private final @Nullable Integer limit; private final @Nullable String cursor; NotificationFilters(@Nullable Boolean isNew, @Nullable Integer limit, @Nullable String cursor) { this.isNew = isNew; this.limit = limit; this.cursor = cursor; } public @Nullable Boolean isNew() { return this.isNew; } public @Nullable Integer limit() { return this.limit; } public @Nullable String cursor() { return this.cursor; } public static Builder builder() { return new Builder(); } public static final class Builder { private @Nullable Boolean isNew; private @Nullable Integer limit = 10.0; private @Nullable String cursor; Builder() { } public Builder isNew(@Nullable Boolean isNew) { this.isNew = isNew; return this; } public Builder limit(@Nullable Integer limit) { this.limit = limit; return this; } public Builder cursor(@Nullable String cursor) { this.cursor = cursor; return this; } public NotificationFilters build() { return new NotificationFilters(isNew, limit, cursor); } } } ``` When you look at the generated Builder, you can indeed see that a double value is being assigned to an Integer. These errors happen for all models that have a default value set. Following is part of the schema file: ``` { "kind": "INPUT_OBJECT", "name": "NotificationFilters", "description": "", "fields": null, "inputFields": [ { "name": "isNew", "description": "", "type": { "kind": "SCALAR", "name": "Boolean", "ofType": null }, "defaultValue": null }, { "name": "limit", "description": "", "type": { "kind": "SCALAR", "name": "Int", "ofType": null }, "defaultValue": "10" }, { "name": "cursor", "description": "", "type": { "kind": "SCALAR", "name": "String", "ofType": null }, "defaultValue": null } ], "interfaces": null, "enumValues": null, "possibleTypes": null } ``` Any ideas or tips on what might be going wrong? I'm using the 0.3.1-SNAPSHOT.
priority
trying to assign a double value into integer field hey i don t know if by my error or by some kind of bug in the compiler an erroneus code is being generated following is one of the errors i m getting error error incompatible types double cannot be converted to integer generated file looks like this package type import java lang boolean import java lang integer import java lang string import javax annotation generated import javax annotation nullable generated apollo graphql public final class notificationfilters private final nullable boolean isnew private final nullable integer limit private final nullable string cursor notificationfilters nullable boolean isnew nullable integer limit nullable string cursor this isnew isnew this limit limit this cursor cursor public nullable boolean isnew return this isnew public nullable integer limit return this limit public nullable string cursor return this cursor public static builder builder return new builder public static final class builder private nullable boolean isnew private nullable integer limit private nullable string cursor builder public builder isnew nullable boolean isnew this isnew isnew return this public builder limit nullable integer limit this limit limit return this public builder cursor nullable string cursor this cursor cursor return this public notificationfilters build return new notificationfilters isnew limit cursor when you look at the generated builder you can indeed see that a double value is being assigned to an integer these errors happen for all models that have a default value set following is part of the schema file kind input object name notificationfilters description fields null inputfields name isnew description type kind scalar name boolean oftype null defaultvalue null name limit description type kind scalar name int oftype null defaultvalue name cursor description type kind scalar name string oftype null defaultvalue null interfaces null enumvalues null possibletypes null any ideas or tips on what might be going wrong i m using the snapshot
1
593,645
18,013,058,463
IssuesEvent
2021-09-16 10:52:13
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Prioritise methods when showing completions for module
Type/Improvement Priority/High Team/LanguageServer Points/1 Area/Completion
**Description:** Check the below figure. It would be great if we optimise the completions to show the methods before the variables. <img width="820" alt="Screenshot 2021-08-30 at 13 41 40" src="https://user-images.githubusercontent.com/7523274/131852202-6ee30149-5383-44e2-b683-e98efba80f33.png">
1.0
Prioritise methods when showing completions for module - **Description:** Check the below figure. It would be great if we optimise the completions to show the methods before the variables. <img width="820" alt="Screenshot 2021-08-30 at 13 41 40" src="https://user-images.githubusercontent.com/7523274/131852202-6ee30149-5383-44e2-b683-e98efba80f33.png">
priority
prioritise methods when showing completions for module description check the below figure it would be great if we optimise the completions to show the methods before the variables img width alt screenshot at src
1
578,782
17,154,348,875
IssuesEvent
2021-07-14 03:38:03
nerPatricia/imoveis-frontend
https://api.github.com/repos/nerPatricia/imoveis-frontend
closed
Update: Fazer o html do Editar Imoveis adicionar dados no form
high priority
utilizar formControlName, [formGroup], essas coisas de form reativo <br> ![image](https://user-images.githubusercontent.com/15932439/125326922-f4294e00-e318-11eb-8604-fd7d03bc2e3c.png)
1.0
Update: Fazer o html do Editar Imoveis adicionar dados no form - utilizar formControlName, [formGroup], essas coisas de form reativo <br> ![image](https://user-images.githubusercontent.com/15932439/125326922-f4294e00-e318-11eb-8604-fd7d03bc2e3c.png)
priority
update fazer o html do editar imoveis adicionar dados no form utilizar formcontrolname essas coisas de form reativo
1
598,464
18,245,829,067
IssuesEvent
2021-10-01 18:15:11
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
closed
default displaySetupWizard to true for cloud
type/enhancement priority/high cloud-public-launch
@Jamakase noticed that this value wasn't being set properly. Currently in OSS we set it to be `true` for the first workspace created/seeded on server startup and false otherwise. For cloud we need to be able to select this option via the api call. This requires adding an optional `displaySetupWizard` to `WorkspaceCreate` in the OSS API and setting that value in the call made form the Cloud API.
1.0
default displaySetupWizard to true for cloud - @Jamakase noticed that this value wasn't being set properly. Currently in OSS we set it to be `true` for the first workspace created/seeded on server startup and false otherwise. For cloud we need to be able to select this option via the api call. This requires adding an optional `displaySetupWizard` to `WorkspaceCreate` in the OSS API and setting that value in the call made form the Cloud API.
priority
default displaysetupwizard to true for cloud jamakase noticed that this value wasn t being set properly currently in oss we set it to be true for the first workspace created seeded on server startup and false otherwise for cloud we need to be able to select this option via the api call this requires adding an optional displaysetupwizard to workspacecreate in the oss api and setting that value in the call made form the cloud api
1
612,887
19,058,461,316
IssuesEvent
2021-11-26 02:02:01
tracer-protocol/perpetual-pools-contracts
https://api.github.com/repos/tracer-protocol/perpetual-pools-contracts
closed
Minting and burning fee
enhancement priority: high review started
**Overview** - Within V2, there will need to be a fee charged both the action of minting and burning. In other words, users will be charged a fee at the time of minting tokens, and also at the time of burning tokens. - The fee will be a percentage. e.g. 1%. It should be a configurable parameter. The fee should be able to be changed by the contract owner after the time of deployment. - The fee rate for minting will not necessarily be the same as the fee rate for burning. - The proceeds of the fee will be added to the pool at which the user was minting into (or burning from). --- **Minting Fee Logic** - Hypothetically, let's say the mint fee rate is 1%. - A user, Bob, commits to minting 100 USDC. - At the time of committing to mint, 1 USDC will be taken away from the user, Bob. It is 1 USDC because it's 100 USDC * 1%. - Bob's commit is now only 99 USDC. This will proceed as normal, and will mint tokens at the fair price. Dev notes: - In the above example, The user commits 100 Usdc, all 100 are transferred to the pool, but the amount of tokens stored as part of the commit would be 99, so when they go to claim, they will only get that 99 usdc worth of tokens. --- **Burning Fee Logic** - Hypothetically, let's say the burn fee rate is 2%. - A user, Bob, has commited to burn his 200 Long Pool tokens. - The 200 long pool tokens will sit in the shadow pool until the wait time is up. - At the rebalancing event in which the commit will be executed, the SC should calculate the value of the 200 long pool tokens (e.g. if price is $2, then the overall value is $400). - From this, the Smart contract takes $8 ($400 * 2%) from the user, and adds it back into the pool from which it was taken. All the users pool tokens are destroyed. - 98% of the value is sent to the user. --- ### Extra notes - Burning fees will require storing the fee rate at in a given update interval (since fee rate can change at any time) and using that to calculate how much quote token a user is entitled to during `updateAggregateBalance` --- ### Completion criteria (added by devs) Where `n` is the fee rate as a percentage (`n == 1` means fee is 1%) - When a user mints, `100 - n%` of the settlement tokens they minted with is stored in the user's commitment data. - When a user mints then claims after update interval, the pool should still be in ownership of `n%` of the original commitment amount of settlement tokens. - When a user mints, `n%` of the settlement tokens they minted with should be added to the balance of the side to which they are minting. - e.g. if they mint $100 to the LONG side, with a 1% fee, directly after minting, $1 should be added to `LeveragedPool::longBalance` - When a user burns, `100 - n%` of the settlement tokens they would normally be entitled to are in the user's aggregate balance after the update interval passes. - When a user burns, upkeep happens, then they update their balance (usually through claiming), n% of the settlement tokens they gained should be kept in the pool and that amount added to the correct side of the pool (either `longBalance` or `shortBalance`) - The same as above, but with `100 - n%` of the settlement tokens deposited to the user's wallet after claiming. - If `n == 0`, then the user should get the exact amount they would be entitled to before the changes of this PR were implemented
1.0
Minting and burning fee - **Overview** - Within V2, there will need to be a fee charged both the action of minting and burning. In other words, users will be charged a fee at the time of minting tokens, and also at the time of burning tokens. - The fee will be a percentage. e.g. 1%. It should be a configurable parameter. The fee should be able to be changed by the contract owner after the time of deployment. - The fee rate for minting will not necessarily be the same as the fee rate for burning. - The proceeds of the fee will be added to the pool at which the user was minting into (or burning from). --- **Minting Fee Logic** - Hypothetically, let's say the mint fee rate is 1%. - A user, Bob, commits to minting 100 USDC. - At the time of committing to mint, 1 USDC will be taken away from the user, Bob. It is 1 USDC because it's 100 USDC * 1%. - Bob's commit is now only 99 USDC. This will proceed as normal, and will mint tokens at the fair price. Dev notes: - In the above example, The user commits 100 Usdc, all 100 are transferred to the pool, but the amount of tokens stored as part of the commit would be 99, so when they go to claim, they will only get that 99 usdc worth of tokens. --- **Burning Fee Logic** - Hypothetically, let's say the burn fee rate is 2%. - A user, Bob, has commited to burn his 200 Long Pool tokens. - The 200 long pool tokens will sit in the shadow pool until the wait time is up. - At the rebalancing event in which the commit will be executed, the SC should calculate the value of the 200 long pool tokens (e.g. if price is $2, then the overall value is $400). - From this, the Smart contract takes $8 ($400 * 2%) from the user, and adds it back into the pool from which it was taken. All the users pool tokens are destroyed. - 98% of the value is sent to the user. --- ### Extra notes - Burning fees will require storing the fee rate at in a given update interval (since fee rate can change at any time) and using that to calculate how much quote token a user is entitled to during `updateAggregateBalance` --- ### Completion criteria (added by devs) Where `n` is the fee rate as a percentage (`n == 1` means fee is 1%) - When a user mints, `100 - n%` of the settlement tokens they minted with is stored in the user's commitment data. - When a user mints then claims after update interval, the pool should still be in ownership of `n%` of the original commitment amount of settlement tokens. - When a user mints, `n%` of the settlement tokens they minted with should be added to the balance of the side to which they are minting. - e.g. if they mint $100 to the LONG side, with a 1% fee, directly after minting, $1 should be added to `LeveragedPool::longBalance` - When a user burns, `100 - n%` of the settlement tokens they would normally be entitled to are in the user's aggregate balance after the update interval passes. - When a user burns, upkeep happens, then they update their balance (usually through claiming), n% of the settlement tokens they gained should be kept in the pool and that amount added to the correct side of the pool (either `longBalance` or `shortBalance`) - The same as above, but with `100 - n%` of the settlement tokens deposited to the user's wallet after claiming. - If `n == 0`, then the user should get the exact amount they would be entitled to before the changes of this PR were implemented
priority
minting and burning fee overview within there will need to be a fee charged both the action of minting and burning in other words users will be charged a fee at the time of minting tokens and also at the time of burning tokens the fee will be a percentage e g it should be a configurable parameter the fee should be able to be changed by the contract owner after the time of deployment the fee rate for minting will not necessarily be the same as the fee rate for burning the proceeds of the fee will be added to the pool at which the user was minting into or burning from minting fee logic hypothetically let s say the mint fee rate is a user bob commits to minting usdc at the time of committing to mint usdc will be taken away from the user bob it is usdc because it s usdc bob s commit is now only usdc this will proceed as normal and will mint tokens at the fair price dev notes in the above example the user commits usdc all are transferred to the pool but the amount of tokens stored as part of the commit would be so when they go to claim they will only get that usdc worth of tokens burning fee logic hypothetically let s say the burn fee rate is a user bob has commited to burn his long pool tokens the long pool tokens will sit in the shadow pool until the wait time is up at the rebalancing event in which the commit will be executed the sc should calculate the value of the long pool tokens e g if price is then the overall value is from this the smart contract takes from the user and adds it back into the pool from which it was taken all the users pool tokens are destroyed of the value is sent to the user extra notes burning fees will require storing the fee rate at in a given update interval since fee rate can change at any time and using that to calculate how much quote token a user is entitled to during updateaggregatebalance completion criteria added by devs where n is the fee rate as a percentage n means fee is when a user mints n of the settlement tokens they minted with is stored in the user s commitment data when a user mints then claims after update interval the pool should still be in ownership of n of the original commitment amount of settlement tokens when a user mints n of the settlement tokens they minted with should be added to the balance of the side to which they are minting e g if they mint to the long side with a fee directly after minting should be added to leveragedpool longbalance when a user burns n of the settlement tokens they would normally be entitled to are in the user s aggregate balance after the update interval passes when a user burns upkeep happens then they update their balance usually through claiming n of the settlement tokens they gained should be kept in the pool and that amount added to the correct side of the pool either longbalance or shortbalance the same as above but with n of the settlement tokens deposited to the user s wallet after claiming if n then the user should get the exact amount they would be entitled to before the changes of this pr were implemented
1
794,712
28,045,852,771
IssuesEvent
2023-03-28 22:48:33
fleek-network/ursa
https://api.github.com/repos/fleek-network/ursa
opened
chore: send advertisement on `put`
high-priority
# Description <!-- Please include a summary of the issue, including motivation, context, and the behavior that should be expected --> Edge nodes are not advertising to the indexer after they pull content from other nodes (data replication). # Checklist - [ ] ~I have ensured that my version is up-to-date~ - [ ] ~I have ensured that my issue is reproducible~ - [ ] I have ensured that my issue is not a duplicate
1.0
chore: send advertisement on `put` - # Description <!-- Please include a summary of the issue, including motivation, context, and the behavior that should be expected --> Edge nodes are not advertising to the indexer after they pull content from other nodes (data replication). # Checklist - [ ] ~I have ensured that my version is up-to-date~ - [ ] ~I have ensured that my issue is reproducible~ - [ ] I have ensured that my issue is not a duplicate
priority
chore send advertisement on put description edge nodes are not advertising to the indexer after they pull content from other nodes data replication checklist i have ensured that my version is up to date i have ensured that my issue is reproducible i have ensured that my issue is not a duplicate
1
536,158
15,705,159,411
IssuesEvent
2021-03-26 15:49:32
nf-core/nf-co.re
https://api.github.com/repos/nf-core/nf-co.re
opened
Web launch tool previews JSON with quotes
bug high-priority pipeline-tools
The web launch final JSON preview wraps non-string entities in quotes. If copied into a params file for Nextflow, this breaks with the new schema validation code as that checks variable types and complains that everything is a string. Steps to reproduce: * Clicking _Launch_ on a pipeline such as nf-core/eager * Fill in the form, setting some boolean and numeric values * Check the _Launch parameters saved_ page after submission - preview JSON has quotes around booleans and numeric fields: For example: ```json { "run_bam_filtering": "true", "bamutils_clip_half_udg_left": "2" } ``` Should be: ```json { "run_bam_filtering": true, "bamutils_clip_half_udg_left": 2 } ``` If running `nf-core launch --id xxx` to pull these results, they go via the command line tools and the correct types seem to be used in the output JSON (though that has its [own problems](https://github.com/nf-core/tools/issues/976) still!).
1.0
Web launch tool previews JSON with quotes - The web launch final JSON preview wraps non-string entities in quotes. If copied into a params file for Nextflow, this breaks with the new schema validation code as that checks variable types and complains that everything is a string. Steps to reproduce: * Clicking _Launch_ on a pipeline such as nf-core/eager * Fill in the form, setting some boolean and numeric values * Check the _Launch parameters saved_ page after submission - preview JSON has quotes around booleans and numeric fields: For example: ```json { "run_bam_filtering": "true", "bamutils_clip_half_udg_left": "2" } ``` Should be: ```json { "run_bam_filtering": true, "bamutils_clip_half_udg_left": 2 } ``` If running `nf-core launch --id xxx` to pull these results, they go via the command line tools and the correct types seem to be used in the output JSON (though that has its [own problems](https://github.com/nf-core/tools/issues/976) still!).
priority
web launch tool previews json with quotes the web launch final json preview wraps non string entities in quotes if copied into a params file for nextflow this breaks with the new schema validation code as that checks variable types and complains that everything is a string steps to reproduce clicking launch on a pipeline such as nf core eager fill in the form setting some boolean and numeric values check the launch parameters saved page after submission preview json has quotes around booleans and numeric fields for example json run bam filtering true bamutils clip half udg left should be json run bam filtering true bamutils clip half udg left if running nf core launch id xxx to pull these results they go via the command line tools and the correct types seem to be used in the output json though that has its still
1
639,969
20,770,231,107
IssuesEvent
2022-03-16 03:15:47
NCC-CNC/whattodo
https://api.github.com/repos/NCC-CNC/whattodo
closed
Change "weights" text to "relative importance"
enhancement high priority
* change "weights" label in sidebar * remove weights from text header in the excel spreadsheet code
1.0
Change "weights" text to "relative importance" - * change "weights" label in sidebar * remove weights from text header in the excel spreadsheet code
priority
change weights text to relative importance change weights label in sidebar remove weights from text header in the excel spreadsheet code
1
203,208
7,058,651,583
IssuesEvent
2018-01-04 21:17:44
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
Make sure that none of the MergeOperations implement MutatingOperation
Priority: High Team: Core
David, please make sure that none of the MergeOperations in the codebase implement `MutatingOperation` after you finish the MergePolicies PRD. Thx a lot!
1.0
Make sure that none of the MergeOperations implement MutatingOperation - David, please make sure that none of the MergeOperations in the codebase implement `MutatingOperation` after you finish the MergePolicies PRD. Thx a lot!
priority
make sure that none of the mergeoperations implement mutatingoperation david please make sure that none of the mergeoperations in the codebase implement mutatingoperation after you finish the mergepolicies prd thx a lot
1
704,361
24,194,142,647
IssuesEvent
2022-09-23 21:04:10
returntocorp/semgrep
https://api.github.com/repos/returntocorp/semgrep
closed
CLI hangs after scan
priority:high
**Describe the bug** I run semgrep as part of a bash script, and for some reason it does not hand control back to the parent script. **To Reproduce** Steps to reproduce the behavior, ideally a link to https://semgrep.dev: ~/.local/bin/semgrep --config auto --json -o ../atd_results/$repo_name.findings.json ./ I'm targeting a clone of ["Small Test Repo"](https://github.com/rtyley/small-test-repo) **Expected behavior** I get the report on STDERR, I see the json file, but the program just hangs, never terminating. **Screenshots** ``` ./test-agent.sh mkdir: cannot create directory ‘atd_results’: File exists mkdir: cannot create directory ‘temp_repo’: File exists Initialized empty Git repository in /home/jason/atd/temp_repo/.git/ url: git@github.com:rtyley/small-test-repo.git Fetching origin Switched to branch 'master' Gathering Code Metrics Static Code Analysis Semgrep rule registry URL is https://semgrep.dev/registry. Scanning 2 files with 52 <multilang> rules. 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████|2/2 tasks Some files were skipped or only partially analyzed. Scan was limited to files tracked by git. (need more rules? `semgrep login` for additional free Semgrep Registry rules) Ran 1036 rules on 1 file: 0 findings. If Semgrep missed a finding, please send us feedback to let us know! $ semgrep shouldafound --help ^C ``` **What is the priority of the bug to you?** - [x] P0: blocking your adoption of Semgrep or workflow - [ ] P1: important to fix or quite annoying - [ ] P2: regular bug that should get fixed **Environment** brew installed semgrep on Ubuntu 22.04 in WSL2.0 Windows 11. **Use case** I'm trying to get semgrep reports alongside pymetrics into our dashboard.
1.0
CLI hangs after scan - **Describe the bug** I run semgrep as part of a bash script, and for some reason it does not hand control back to the parent script. **To Reproduce** Steps to reproduce the behavior, ideally a link to https://semgrep.dev: ~/.local/bin/semgrep --config auto --json -o ../atd_results/$repo_name.findings.json ./ I'm targeting a clone of ["Small Test Repo"](https://github.com/rtyley/small-test-repo) **Expected behavior** I get the report on STDERR, I see the json file, but the program just hangs, never terminating. **Screenshots** ``` ./test-agent.sh mkdir: cannot create directory ‘atd_results’: File exists mkdir: cannot create directory ‘temp_repo’: File exists Initialized empty Git repository in /home/jason/atd/temp_repo/.git/ url: git@github.com:rtyley/small-test-repo.git Fetching origin Switched to branch 'master' Gathering Code Metrics Static Code Analysis Semgrep rule registry URL is https://semgrep.dev/registry. Scanning 2 files with 52 <multilang> rules. 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████|2/2 tasks Some files were skipped or only partially analyzed. Scan was limited to files tracked by git. (need more rules? `semgrep login` for additional free Semgrep Registry rules) Ran 1036 rules on 1 file: 0 findings. If Semgrep missed a finding, please send us feedback to let us know! $ semgrep shouldafound --help ^C ``` **What is the priority of the bug to you?** - [x] P0: blocking your adoption of Semgrep or workflow - [ ] P1: important to fix or quite annoying - [ ] P2: regular bug that should get fixed **Environment** brew installed semgrep on Ubuntu 22.04 in WSL2.0 Windows 11. **Use case** I'm trying to get semgrep reports alongside pymetrics into our dashboard.
priority
cli hangs after scan describe the bug i run semgrep as part of a bash script and for some reason it does not hand control back to the parent script to reproduce steps to reproduce the behavior ideally a link to local bin semgrep config auto json o atd results repo name findings json i m targeting a clone of expected behavior i get the report on stderr i see the json file but the program just hangs never terminating screenshots test agent sh mkdir cannot create directory ‘atd results’ file exists mkdir cannot create directory ‘temp repo’ file exists initialized empty git repository in home jason atd temp repo git url git github com rtyley small test repo git fetching origin switched to branch master gathering code metrics static code analysis semgrep rule registry url is scanning files with rules ███████████████████████████████████████████████████████████████████████████████████████████████████████ tasks some files were skipped or only partially analyzed scan was limited to files tracked by git need more rules semgrep login for additional free semgrep registry rules ran rules on file findings if semgrep missed a finding please send us feedback to let us know semgrep shouldafound help c what is the priority of the bug to you blocking your adoption of semgrep or workflow important to fix or quite annoying regular bug that should get fixed environment brew installed semgrep on ubuntu in windows use case i m trying to get semgrep reports alongside pymetrics into our dashboard
1