Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
12,896
15,284,184,903
IssuesEvent
2021-02-23 11:53:16
Bedrohung-der-Bienen/Transformationsfelder-Digitalisierung
https://api.github.com/repos/Bedrohung-der-Bienen/Transformationsfelder-Digitalisierung
opened
Benutzer in der Datenbank anlegen
backburner backend login process register process
Es muss in der Datenbank die Tabelle Benutzer erstellt werden mit ihren Attributen
2.0
Benutzer in der Datenbank anlegen - Es muss in der Datenbank die Tabelle Benutzer erstellt werden mit ihren Attributen
process
benutzer in der datenbank anlegen es muss in der datenbank die tabelle benutzer erstellt werden mit ihren attributen
1
21,345
11,198,425,054
IssuesEvent
2020-01-03 15:54:59
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
File low-hanging fruit of HomeActivity creation
P3 eng:performance
Let's file actionable bugs to address the low-hanging fruit of so we can 1) prioritize them & ensure we're working on the most important things and 2) ensure our backlog has actionable works for devs. ### Acceptance criteria - File bugs to address the low-hanging fruit (and some larger diminishing returns) in `HomeActivity` creation - Notify Slack channel when issues are filed --- Overall startup time is around 300ms (at this point and on a specific device). `HomeActivity.onCreate` takes 307ms. Layout inflation alone takes 126ms. `setupToolbarAndNavigation` takes 70ms. @colintheshots @csadilek can you tag appropriately? Thanks!
True
File low-hanging fruit of HomeActivity creation - Let's file actionable bugs to address the low-hanging fruit of so we can 1) prioritize them & ensure we're working on the most important things and 2) ensure our backlog has actionable works for devs. ### Acceptance criteria - File bugs to address the low-hanging fruit (and some larger diminishing returns) in `HomeActivity` creation - Notify Slack channel when issues are filed --- Overall startup time is around 300ms (at this point and on a specific device). `HomeActivity.onCreate` takes 307ms. Layout inflation alone takes 126ms. `setupToolbarAndNavigation` takes 70ms. @colintheshots @csadilek can you tag appropriately? Thanks!
non_process
file low hanging fruit of homeactivity creation let s file actionable bugs to address the low hanging fruit of so we can prioritize them ensure we re working on the most important things and ensure our backlog has actionable works for devs acceptance criteria file bugs to address the low hanging fruit and some larger diminishing returns in homeactivity creation notify slack channel when issues are filed overall startup time is around at this point and on a specific device homeactivity oncreate takes layout inflation alone takes setuptoolbarandnavigation takes colintheshots csadilek can you tag appropriately thanks
0
209,804
23,730,845,641
IssuesEvent
2022-08-31 01:27:39
vipinsun/blockchain-carbon-accounting
https://api.github.com/repos/vipinsun/blockchain-carbon-accounting
closed
CVE-2020-11023 (Medium) detected in multiple libraries - autoclosed
security vulnerability
## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-3.2.1.min.js</b>, <b>jquery-1.7.2.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary> <p> <details><summary><b>jquery-3.2.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.2.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.2.1/jquery.min.js</a></p> <p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/superagent/docs/tail.html</p> <p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/superagent/docs/tail.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.2.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/jmespath/index.html</p> <p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/jmespath/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/dot-qs/test/test-dot-querystring.html</p> <p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/dot-qs/test/test-dot-querystring.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/vipinsun/blockchain-carbon-accounting/commit/d388e16464e00b9ce84df0d247029f534a429b90">d388e16464e00b9ce84df0d247029f534a429b90</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11023 (Medium) detected in multiple libraries - autoclosed - ## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-3.2.1.min.js</b>, <b>jquery-1.7.2.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary> <p> <details><summary><b>jquery-3.2.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.2.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.2.1/jquery.min.js</a></p> <p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/superagent/docs/tail.html</p> <p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/superagent/docs/tail.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.2.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/jmespath/index.html</p> <p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/jmespath/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/dot-qs/test/test-dot-querystring.html</p> <p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/dot-qs/test/test-dot-querystring.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/vipinsun/blockchain-carbon-accounting/commit/d388e16464e00b9ce84df0d247029f534a429b90">d388e16464e00b9ce84df0d247029f534a429b90</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in multiple libraries autoclosed cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file utility emissions channel typescript app node modules superagent docs tail html path to vulnerable library utility emissions channel typescript app node modules superagent docs tail html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file utility emissions channel typescript app node modules jmespath index html path to vulnerable library utility emissions channel typescript app node modules jmespath index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file utility emissions channel typescript app node modules dot qs test test dot querystring html path to vulnerable library utility emissions channel typescript app node modules dot qs test test dot querystring html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails step up your open source security game with mend
0
12,916
15,288,166,606
IssuesEvent
2021-02-23 16:33:47
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
ARM: VMUL not decompiled properly with VFPv2/VFPv3
Feature: Processor/ARM Type: Bug
**Describe the bug** When decompiling any binary built with VFPv2/VFPv3 vector extensions for the 32-bit variant of ARM, these instructions get decompiled as CDP instructions. While not necessarily wrong, this does make the instructions significantly more annoying to read. This [appears](https://github.com/NationalSecurityAgency/ghidra/blob/master/Ghidra/Processors/ARM/data/languages/ARMneon.sinc#L2992-L3002) to be due to the fact that Sleigh constructors only exist for the NEON A1/T1 variants of VMUL. There doesn't seem to be constructors for the VFPv2/VFPv3/VFPv4 A2/T2 variants that allow a condition code on the A2 variant. Just figured that I'd report this in case it was an oversight. **To Reproduce** 1. Decompile any application variant that makes use of VFPv2/VFPv3 variants of VMUL. 2. Notice that it gets decompiled as a CDP instruction. 3. Despair **Expected behavior** The instructions get decompiled as VMUL variants, and not as CDP instructions. **Environment (please complete the following information):** - OS: Windows 10 - Java Version: 12.0 - Ghidra Version: 9.1
1.0
ARM: VMUL not decompiled properly with VFPv2/VFPv3 - **Describe the bug** When decompiling any binary built with VFPv2/VFPv3 vector extensions for the 32-bit variant of ARM, these instructions get decompiled as CDP instructions. While not necessarily wrong, this does make the instructions significantly more annoying to read. This [appears](https://github.com/NationalSecurityAgency/ghidra/blob/master/Ghidra/Processors/ARM/data/languages/ARMneon.sinc#L2992-L3002) to be due to the fact that Sleigh constructors only exist for the NEON A1/T1 variants of VMUL. There doesn't seem to be constructors for the VFPv2/VFPv3/VFPv4 A2/T2 variants that allow a condition code on the A2 variant. Just figured that I'd report this in case it was an oversight. **To Reproduce** 1. Decompile any application variant that makes use of VFPv2/VFPv3 variants of VMUL. 2. Notice that it gets decompiled as a CDP instruction. 3. Despair **Expected behavior** The instructions get decompiled as VMUL variants, and not as CDP instructions. **Environment (please complete the following information):** - OS: Windows 10 - Java Version: 12.0 - Ghidra Version: 9.1
process
arm vmul not decompiled properly with describe the bug when decompiling any binary built with vector extensions for the bit variant of arm these instructions get decompiled as cdp instructions while not necessarily wrong this does make the instructions significantly more annoying to read this to be due to the fact that sleigh constructors only exist for the neon variants of vmul there doesn t seem to be constructors for the variants that allow a condition code on the variant just figured that i d report this in case it was an oversight to reproduce decompile any application variant that makes use of variants of vmul notice that it gets decompiled as a cdp instruction despair expected behavior the instructions get decompiled as vmul variants and not as cdp instructions environment please complete the following information os windows java version ghidra version
1
18,961
24,921,876,549
IssuesEvent
2022-10-31 01:30:37
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Child process IPC option for plain text serialization
child_process feature request stale
### Is your feature request related to a problem? Please describe. Spawning a child process with IPC serialization currently only has [two serialization modes](https://nodejs.org/api/child_process.html#child_process_child_process_spawn_command_args_options), 'json' and 'advanced'. 'advanced' is specific to node/V8, which works great when the subprocess is another Node.js process, but for non node.js processes it's not an option. JSON serialization mode currently has no option for a custom serializer, so it's hard to send complex data (bigint, function, etc) without first JSON.serializaing with a custom encoder, JSON.parsing it, then having it serialized back to JSON internally. ### Describe the solution you'd like This overhead can be reduced by adding a plaintext option (probably the easiest) to let the user do the encoding. Though maybe this introduces issues with buffering? If so maybe some way to add a custom JSON replacer. ### Describe alternatives you've considered Some other IPC mode? Currently I have to resort to a really bad hack to expose the internal writeUtf8String method in child_process/serialization. https://github.com/extremeheat/JSPyBridge/blob/master/src/pythonia/IpcPipeCom.js#L15
1.0
Child process IPC option for plain text serialization - ### Is your feature request related to a problem? Please describe. Spawning a child process with IPC serialization currently only has [two serialization modes](https://nodejs.org/api/child_process.html#child_process_child_process_spawn_command_args_options), 'json' and 'advanced'. 'advanced' is specific to node/V8, which works great when the subprocess is another Node.js process, but for non node.js processes it's not an option. JSON serialization mode currently has no option for a custom serializer, so it's hard to send complex data (bigint, function, etc) without first JSON.serializaing with a custom encoder, JSON.parsing it, then having it serialized back to JSON internally. ### Describe the solution you'd like This overhead can be reduced by adding a plaintext option (probably the easiest) to let the user do the encoding. Though maybe this introduces issues with buffering? If so maybe some way to add a custom JSON replacer. ### Describe alternatives you've considered Some other IPC mode? Currently I have to resort to a really bad hack to expose the internal writeUtf8String method in child_process/serialization. https://github.com/extremeheat/JSPyBridge/blob/master/src/pythonia/IpcPipeCom.js#L15
process
child process ipc option for plain text serialization is your feature request related to a problem please describe spawning a child process with ipc serialization currently only has json and advanced advanced is specific to node which works great when the subprocess is another node js process but for non node js processes it s not an option json serialization mode currently has no option for a custom serializer so it s hard to send complex data bigint function etc without first json serializaing with a custom encoder json parsing it then having it serialized back to json internally describe the solution you d like this overhead can be reduced by adding a plaintext option probably the easiest to let the user do the encoding though maybe this introduces issues with buffering if so maybe some way to add a custom json replacer describe alternatives you ve considered some other ipc mode currently i have to resort to a really bad hack to expose the internal method in child process serialization
1
5,139
7,922,824,471
IssuesEvent
2018-07-05 12:13:54
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
Prompts are not displayed on a newline after evaluating empty string
process:eval
With cursor at an empty prompt: ``` > *cursor* ``` Hitting RET several times produces this: ``` > > > *cursor* ``` Instead of ``` > > > *cursor* ```
1.0
Prompts are not displayed on a newline after evaluating empty string - With cursor at an empty prompt: ``` > *cursor* ``` Hitting RET several times produces this: ``` > > > *cursor* ``` Instead of ``` > > > *cursor* ```
process
prompts are not displayed on a newline after evaluating empty string with cursor at an empty prompt cursor hitting ret several times produces this cursor instead of cursor
1
11,734
14,576,707,972
IssuesEvent
2020-12-18 00:08:14
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
how to parse custom log format that contains a pipe as delimiter
bug duplicate log-processing
my log format as follow: **03/Sep/2019:10:04:08 +0800 | 10.10.10.10 | http | 304 | 192.168.2.12 | - | www.hello.com | - | - | GET / HTTP/1.1 | 0 | - | Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.110 Safari/547.36 | - | 0.017 | - | - | -** Please can you help me with the right command ? Thanks in advance. Regards
1.0
how to parse custom log format that contains a pipe as delimiter - my log format as follow: **03/Sep/2019:10:04:08 +0800 | 10.10.10.10 | http | 304 | 192.168.2.12 | - | www.hello.com | - | - | GET / HTTP/1.1 | 0 | - | Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.110 Safari/547.36 | - | 0.017 | - | - | -** Please can you help me with the right command ? Thanks in advance. Regards
process
how to parse custom log format that contains a pipe as delimiter my log format as follow sep http get http mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari please can you help me with the right command thanks in advance regards
1
15,647
19,846,426,421
IssuesEvent
2022-01-21 07:06:28
ooi-data/RS01SBPD-DP01A-06-DOSTAD104-recovered_wfp-dpc_optode_instrument_recovered
https://api.github.com/repos/ooi-data/RS01SBPD-DP01A-06-DOSTAD104-recovered_wfp-dpc_optode_instrument_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:06:28.296042. ## Details Flow name: `RS01SBPD-DP01A-06-DOSTAD104-recovered_wfp-dpc_optode_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:06:28.296042. ## Details Flow name: `RS01SBPD-DP01A-06-DOSTAD104-recovered_wfp-dpc_optode_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered wfp dpc optode instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
17,872
23,817,012,201
IssuesEvent
2022-09-05 07:47:43
Tencent/tdesign-miniprogram
https://api.github.com/repos/Tencent/tdesign-miniprogram
closed
[t-tabs] t-tab-panel 里面内容绑定事件没有触发
bug good first issue in process
### tdesign-miniprogram 版本 0.19.3 ### 重现链接 _No response_ ### 重现步骤 <t-tabs value="{{typeid}}" placement="left" bind:change="onTabsChange"> <t-tab-panel wx:for="{{list}}" wx:key="id" label="{{item.text}}" value="{{item.id}}"> <view bindtap="onchange">内容这里事件无法触发</view> </t-tab-panel> </t-tabs> ### 期望结果 _No response_ ### 实际结果 _No response_ ### 框架版本 _No response_ ### 浏览器版本 _No response_ ### 系统版本 _No response_ ### Node版本 _No response_ ### 补充说明 请尽快修复,完全不能用影响挺大的
1.0
[t-tabs] t-tab-panel 里面内容绑定事件没有触发 - ### tdesign-miniprogram 版本 0.19.3 ### 重现链接 _No response_ ### 重现步骤 <t-tabs value="{{typeid}}" placement="left" bind:change="onTabsChange"> <t-tab-panel wx:for="{{list}}" wx:key="id" label="{{item.text}}" value="{{item.id}}"> <view bindtap="onchange">内容这里事件无法触发</view> </t-tab-panel> </t-tabs> ### 期望结果 _No response_ ### 实际结果 _No response_ ### 框架版本 _No response_ ### 浏览器版本 _No response_ ### 系统版本 _No response_ ### Node版本 _No response_ ### 补充说明 请尽快修复,完全不能用影响挺大的
process
t tab panel 里面内容绑定事件没有触发 tdesign miniprogram 版本 重现链接 no response 重现步骤 内容这里事件无法触发 期望结果 no response 实际结果 no response 框架版本 no response 浏览器版本 no response 系统版本 no response node版本 no response 补充说明 请尽快修复,完全不能用影响挺大的
1
41,399
8,970,606,115
IssuesEvent
2019-01-29 14:04:08
surge-synthesizer/surge
https://api.github.com/repos/surge-synthesizer/surge
opened
using namespace in header files
code cleanup
`using namespace` in header files is dangerous. Google will tell you why but here's a good stack overflow https://stackoverflow.com/questions/4872373/why-is-including-using-namespace-into-a-header-file-a-bad-idea-in-c We have these: ``` (p3env) paul:~/dev/music/surge$ grep -r "using namespace" src | grep \.h: src/vst3/SurgeVst3EditController.h:using namespace Steinberg; src/vst3/SurgeVst3EditController.h:using namespace Steinberg::Vst; src/vst3/SurgeVst3Processor.h:using namespace Steinberg; src/common/SurgeStorage.h:using namespace std; src/common/SurgeSynthesizer.h:using namespace std; src/common/globals.h:using namespace std; src/common/gui/PopupEditorSpawner.h:using namespace std; src/common/gui/SurgeGUIEditor.h:using namespace std; ``` which we should clean up one day. My view is using std:: as a non-used namespace is now becoming the way most folks write, so I'd just eliminate those and fix compile errors. The ones in vst3 are a bit less worrying - especially since the edit controller isn't used.
1.0
using namespace in header files - `using namespace` in header files is dangerous. Google will tell you why but here's a good stack overflow https://stackoverflow.com/questions/4872373/why-is-including-using-namespace-into-a-header-file-a-bad-idea-in-c We have these: ``` (p3env) paul:~/dev/music/surge$ grep -r "using namespace" src | grep \.h: src/vst3/SurgeVst3EditController.h:using namespace Steinberg; src/vst3/SurgeVst3EditController.h:using namespace Steinberg::Vst; src/vst3/SurgeVst3Processor.h:using namespace Steinberg; src/common/SurgeStorage.h:using namespace std; src/common/SurgeSynthesizer.h:using namespace std; src/common/globals.h:using namespace std; src/common/gui/PopupEditorSpawner.h:using namespace std; src/common/gui/SurgeGUIEditor.h:using namespace std; ``` which we should clean up one day. My view is using std:: as a non-used namespace is now becoming the way most folks write, so I'd just eliminate those and fix compile errors. The ones in vst3 are a bit less worrying - especially since the edit controller isn't used.
non_process
using namespace in header files using namespace in header files is dangerous google will tell you why but here s a good stack overflow we have these paul dev music surge grep r using namespace src grep h src h using namespace steinberg src h using namespace steinberg vst src h using namespace steinberg src common surgestorage h using namespace std src common surgesynthesizer h using namespace std src common globals h using namespace std src common gui popupeditorspawner h using namespace std src common gui surgeguieditor h using namespace std which we should clean up one day my view is using std as a non used namespace is now becoming the way most folks write so i d just eliminate those and fix compile errors the ones in are a bit less worrying especially since the edit controller isn t used
0
22,204
30,758,658,676
IssuesEvent
2023-07-29 12:02:29
DxytJuly3/gitalk_blog
https://api.github.com/repos/DxytJuly3/gitalk_blog
opened
[Linux] 守护进程介绍、服务器的部署、日志文件... - July.cc Blogs
Gitalk /posts/Linux-Daemon-Process
https://www.julysblog.cn/posts/Linux-Daemon-Process 我们使用的系统中, 一般以服务器的方式工作 对外提供服务的服务器, 都是以守护进程的方式在系统中工作的 比如, 我们使用Linux服务器时, 大多都会使用一些终端软件通过ssh远程连接服务器使用 守护进程, 一旦启动之后. 除非用户手动关闭, 否则不会被关闭 会一直运行
1.0
[Linux] 守护进程介绍、服务器的部署、日志文件... - July.cc Blogs - https://www.julysblog.cn/posts/Linux-Daemon-Process 我们使用的系统中, 一般以服务器的方式工作 对外提供服务的服务器, 都是以守护进程的方式在系统中工作的 比如, 我们使用Linux服务器时, 大多都会使用一些终端软件通过ssh远程连接服务器使用 守护进程, 一旦启动之后. 除非用户手动关闭, 否则不会被关闭 会一直运行
process
守护进程介绍、服务器的部署、日志文件 july cc blogs 我们使用的系统中 一般以服务器的方式工作 对外提供服务的服务器 都是以守护进程的方式在系统中工作的 比如 我们使用linux服务器时 大多都会使用一些终端软件通过ssh远程连接服务器使用 守护进程 一旦启动之后 除非用户手动关闭 否则不会被关闭 会一直运行
1
132,658
18,268,821,301
IssuesEvent
2021-10-04 11:40:28
artsking/linux-3.0.35
https://api.github.com/repos/artsking/linux-3.0.35
opened
CVE-2017-13695 (Medium) detected in linux-stable-rtv3.8.6
security vulnerability
## CVE-2017-13695 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35/commit/5992fa81c6ac1b4e9db13f5408d914525c5b7875">5992fa81c6ac1b4e9db13f5408d914525c5b7875</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The acpi_ns_evaluate() function in drivers/acpi/acpica/nseval.c in the Linux kernel through 4.12.9 does not flush the operand cache and causes a kernel stack dump, which allows local users to obtain sensitive information from kernel memory and bypass the KASLR protection mechanism (in the kernel through 4.9) via a crafted ACPI table. <p>Publish Date: 2017-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-13695>CVE-2017-13695</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/torvalds/linux/commit/97f3c0a4b0579b646b6b10ae5a3d59f0441cc12c#diff-2a91ba8eef66d4e148b6c2abcd63107b">https://github.com/torvalds/linux/commit/97f3c0a4b0579b646b6b10ae5a3d59f0441cc12c#diff-2a91ba8eef66d4e148b6c2abcd63107b</a></p> <p>Release Date: 2017-08-25</p> <p>Fix Resolution: v4.17-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-13695 (Medium) detected in linux-stable-rtv3.8.6 - ## CVE-2017-13695 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35/commit/5992fa81c6ac1b4e9db13f5408d914525c5b7875">5992fa81c6ac1b4e9db13f5408d914525c5b7875</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The acpi_ns_evaluate() function in drivers/acpi/acpica/nseval.c in the Linux kernel through 4.12.9 does not flush the operand cache and causes a kernel stack dump, which allows local users to obtain sensitive information from kernel memory and bypass the KASLR protection mechanism (in the kernel through 4.9) via a crafted ACPI table. <p>Publish Date: 2017-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-13695>CVE-2017-13695</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/torvalds/linux/commit/97f3c0a4b0579b646b6b10ae5a3d59f0441cc12c#diff-2a91ba8eef66d4e148b6c2abcd63107b">https://github.com/torvalds/linux/commit/97f3c0a4b0579b646b6b10ae5a3d59f0441cc12c#diff-2a91ba8eef66d4e148b6c2abcd63107b</a></p> <p>Release Date: 2017-08-25</p> <p>Fix Resolution: v4.17-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers acpi acpica nseval c drivers acpi acpica nseval c drivers acpi acpica nseval c vulnerability details the acpi ns evaluate function in drivers acpi acpica nseval c in the linux kernel through does not flush the operand cache and causes a kernel stack dump which allows local users to obtain sensitive information from kernel memory and bypass the kaslr protection mechanism in the kernel through via a crafted acpi table publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
411,692
27,828,104,238
IssuesEvent
2023-03-20 00:16:28
miguelgrinberg/python-socketio
https://api.github.com/repos/miguelgrinberg/python-socketio
closed
docs: please make sticky sessions requirement louder
documentation
**Is your feature request related to a problem? Please describe.** When run as a cluster, this library requires sticky sessions to be enabled. This wasn't obvious at first and so I deployed a buggy cluster. Please improve the docs or errors to make this requirement clearer (and I'm happy to submit a docs / error message PR, if helpful). (Also thank for for providing this library -- I am very happy to be able to run my socketio backend in my existing fastapi stack). **Describe the solution you'd like** **In the codebase**: - remove the sticky sessions requirement if possible (I don't understand the internals of this project or the socketio protocol, but I assume the sticky sessions requirement is in here for a legit reason, and removing it would be a large project if possible) - explain the situation in the error message, if possible. In my case the error is usually 'Invalid session', presumably from 2 places in [handle_request](https://github.com/miguelgrinberg/python-engineio/blob/3af6bf2b4b28dd1d3534ad21b78366ecb60d5503/src/engineio/server.py#L405) **In docs**: In the [using a message queue](https://python-socketio.readthedocs.io/en/latest/server.html#using-a-message-queue) section, can you link to the [scalability notes warning](https://python-socketio.readthedocs.io/en/latest/server.html#scalability-notes) (which says to use sticky sessions), and make it clear your cluster will fail if you don't meet the reqs? **Describe alternatives you've considered** The alternative I tried first was having my chat intermittently fail in prod. Because my cluster is small, it usually connected eventually, so for a while I assumed it was an error in the ordering of my authorization logic + socket connection. **Logs** n/a, but let me know if you want me to capture something **Additional context** Did a readthrough of other issues mentioning this: - #371 'invalid session' with sanic in multiple workers mode - #437, #219
1.0
docs: please make sticky sessions requirement louder - **Is your feature request related to a problem? Please describe.** When run as a cluster, this library requires sticky sessions to be enabled. This wasn't obvious at first and so I deployed a buggy cluster. Please improve the docs or errors to make this requirement clearer (and I'm happy to submit a docs / error message PR, if helpful). (Also thank for for providing this library -- I am very happy to be able to run my socketio backend in my existing fastapi stack). **Describe the solution you'd like** **In the codebase**: - remove the sticky sessions requirement if possible (I don't understand the internals of this project or the socketio protocol, but I assume the sticky sessions requirement is in here for a legit reason, and removing it would be a large project if possible) - explain the situation in the error message, if possible. In my case the error is usually 'Invalid session', presumably from 2 places in [handle_request](https://github.com/miguelgrinberg/python-engineio/blob/3af6bf2b4b28dd1d3534ad21b78366ecb60d5503/src/engineio/server.py#L405) **In docs**: In the [using a message queue](https://python-socketio.readthedocs.io/en/latest/server.html#using-a-message-queue) section, can you link to the [scalability notes warning](https://python-socketio.readthedocs.io/en/latest/server.html#scalability-notes) (which says to use sticky sessions), and make it clear your cluster will fail if you don't meet the reqs? **Describe alternatives you've considered** The alternative I tried first was having my chat intermittently fail in prod. Because my cluster is small, it usually connected eventually, so for a while I assumed it was an error in the ordering of my authorization logic + socket connection. **Logs** n/a, but let me know if you want me to capture something **Additional context** Did a readthrough of other issues mentioning this: - #371 'invalid session' with sanic in multiple workers mode - #437, #219
non_process
docs please make sticky sessions requirement louder is your feature request related to a problem please describe when run as a cluster this library requires sticky sessions to be enabled this wasn t obvious at first and so i deployed a buggy cluster please improve the docs or errors to make this requirement clearer and i m happy to submit a docs error message pr if helpful also thank for for providing this library i am very happy to be able to run my socketio backend in my existing fastapi stack describe the solution you d like in the codebase remove the sticky sessions requirement if possible i don t understand the internals of this project or the socketio protocol but i assume the sticky sessions requirement is in here for a legit reason and removing it would be a large project if possible explain the situation in the error message if possible in my case the error is usually invalid session presumably from places in in docs in the section can you link to the which says to use sticky sessions and make it clear your cluster will fail if you don t meet the reqs describe alternatives you ve considered the alternative i tried first was having my chat intermittently fail in prod because my cluster is small it usually connected eventually so for a while i assumed it was an error in the ordering of my authorization logic socket connection logs n a but let me know if you want me to capture something additional context did a readthrough of other issues mentioning this invalid session with sanic in multiple workers mode
0
20,163
26,716,686,143
IssuesEvent
2023-01-28 16:05:01
LLazyEmail/nomoretogo_email_template
https://api.github.com/repos/LLazyEmail/nomoretogo_email_template
closed
rename to Weekly Menu button and move to config file
todo in process
https://github.com/LLazyEmail/nomoretogo_email_template/blob/9c1f21518fc1e72bfa95baa91db45fc2d23b822c/src/display/displayContent.js#L175 ```javascript // TODO rename to Weekly Menu button and move to config file const contentButtonParams = { id: '04-22-2021', href: 'https://www.nomoretogo.com/weekly-menu/', ```
1.0
rename to Weekly Menu button and move to config file - https://github.com/LLazyEmail/nomoretogo_email_template/blob/9c1f21518fc1e72bfa95baa91db45fc2d23b822c/src/display/displayContent.js#L175 ```javascript // TODO rename to Weekly Menu button and move to config file const contentButtonParams = { id: '04-22-2021', href: 'https://www.nomoretogo.com/weekly-menu/', ```
process
rename to weekly menu button and move to config file javascript todo rename to weekly menu button and move to config file const contentbuttonparams id href
1
112,038
14,212,892,203
IssuesEvent
2020-11-17 01:10:53
phetsims/energy-skate-park
https://api.github.com/repos/phetsims/energy-skate-park
opened
Slider pixel polishes
design:polish
@KatieWoe reported in https://github.com/phetsims/QA/issues/578#issuecomment-728427972: >On MacOS 11 Safari, I noticed a bit of overlap with longer strings and the thumbs of sliders. ><img width="254" alt="Screen Shot 2020-11-16 at 4 35 37 PM" src="https://user-images.githubusercontent.com/41024075/99321450-12db5180-282b-11eb-93fa-8b796a1734a6.png"> This issue is not limited to macOS/Safari or i18n. The slider thumb is a bit too big, and the tick mark length is a bit too short. The result is that the thumb touches letters that descend below the baseline. This affects both ESP and ESPB. ![image](https://user-images.githubusercontent.com/8419308/99325622-ff80b400-2833-11eb-929b-67b9795929b1.png) ![image](https://user-images.githubusercontent.com/8419308/99325633-060f2b80-2834-11eb-8bfa-9c21a03e0d9b.png) Luckily, the slider options are factored out into [EnergySkatParkConstants.js](https://github.com/phetsims/energy-skate-park/blob/master/js/common/EnergySkateParkConstants.js#L39-L43) which made it really easy to test out a few options locally. I reduced the `thumbSize` height to 23 and increased the `majorTickLength` to 11. This removes the overlap between the slider label and thumb, but does not change the overall height of the panel on the Intro Screen (where vertical space is tight). | Before | After | | --- | --- | | ![image](https://user-images.githubusercontent.com/8419308/99326199-107df500-2835-11eb-8d1d-107d078023e3.png) | ![image](https://user-images.githubusercontent.com/8419308/99326146-f3e1bd00-2834-11eb-9aca-d5264e02db2a.png) | This is honestly quite minor, so I'm hesitant to do anything so close to the finish line. @ariel-phet what do you think?
1.0
Slider pixel polishes - @KatieWoe reported in https://github.com/phetsims/QA/issues/578#issuecomment-728427972: >On MacOS 11 Safari, I noticed a bit of overlap with longer strings and the thumbs of sliders. ><img width="254" alt="Screen Shot 2020-11-16 at 4 35 37 PM" src="https://user-images.githubusercontent.com/41024075/99321450-12db5180-282b-11eb-93fa-8b796a1734a6.png"> This issue is not limited to macOS/Safari or i18n. The slider thumb is a bit too big, and the tick mark length is a bit too short. The result is that the thumb touches letters that descend below the baseline. This affects both ESP and ESPB. ![image](https://user-images.githubusercontent.com/8419308/99325622-ff80b400-2833-11eb-929b-67b9795929b1.png) ![image](https://user-images.githubusercontent.com/8419308/99325633-060f2b80-2834-11eb-8bfa-9c21a03e0d9b.png) Luckily, the slider options are factored out into [EnergySkatParkConstants.js](https://github.com/phetsims/energy-skate-park/blob/master/js/common/EnergySkateParkConstants.js#L39-L43) which made it really easy to test out a few options locally. I reduced the `thumbSize` height to 23 and increased the `majorTickLength` to 11. This removes the overlap between the slider label and thumb, but does not change the overall height of the panel on the Intro Screen (where vertical space is tight). | Before | After | | --- | --- | | ![image](https://user-images.githubusercontent.com/8419308/99326199-107df500-2835-11eb-8d1d-107d078023e3.png) | ![image](https://user-images.githubusercontent.com/8419308/99326146-f3e1bd00-2834-11eb-9aca-d5264e02db2a.png) | This is honestly quite minor, so I'm hesitant to do anything so close to the finish line. @ariel-phet what do you think?
non_process
slider pixel polishes katiewoe reported in on macos safari i noticed a bit of overlap with longer strings and the thumbs of sliders img width alt screen shot at pm src this issue is not limited to macos safari or the slider thumb is a bit too big and the tick mark length is a bit too short the result is that the thumb touches letters that descend below the baseline this affects both esp and espb luckily the slider options are factored out into which made it really easy to test out a few options locally i reduced the thumbsize height to and increased the majorticklength to this removes the overlap between the slider label and thumb but does not change the overall height of the panel on the intro screen where vertical space is tight before after this is honestly quite minor so i m hesitant to do anything so close to the finish line ariel phet what do you think
0
385,358
26,633,218,537
IssuesEvent
2023-01-24 19:33:42
ethyca/fidesdocs
https://api.github.com/repos/ethyca/fidesdocs
closed
update `FIDES__LOG_PII` reference to FIDES__LOGGING__LOG_PII`
documentation
### Docs Update Description We used to control whether PII was masked in our logs via an env variable (and corresponding application config property) that was `FIDES__LOG_PII`. That was changed to `FIDES__LOGGING__LOG_PII` in [this PR](https://github.com/ethyca/fides/pull/2031). We should update our documentation accordingly. ### Additional context (Internal helm charts have already been updated accordingly: https://github.com/ethyca/fides-helm/pull/13)
1.0
update `FIDES__LOG_PII` reference to FIDES__LOGGING__LOG_PII` - ### Docs Update Description We used to control whether PII was masked in our logs via an env variable (and corresponding application config property) that was `FIDES__LOG_PII`. That was changed to `FIDES__LOGGING__LOG_PII` in [this PR](https://github.com/ethyca/fides/pull/2031). We should update our documentation accordingly. ### Additional context (Internal helm charts have already been updated accordingly: https://github.com/ethyca/fides-helm/pull/13)
non_process
update fides log pii reference to fides logging log pii docs update description we used to control whether pii was masked in our logs via an env variable and corresponding application config property that was fides log pii that was changed to fides logging log pii in we should update our documentation accordingly additional context internal helm charts have already been updated accordingly
0
45,894
7,208,625,526
IssuesEvent
2018-02-07 04:15:20
GoogleCloudPlatform/google-cloud-java
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-java
closed
BigQuery Data Transfer: ListDataSourcesRequest not documented / not linked to
api: bigquery documentation priority: p2 type: process
See: https://googlecloudplatform.github.io/google-cloud-java/latest/apidocs/com/google/cloud/bigquery/datatransfer/v1/DataTransferServiceClient.html#listDataSources-com.google.cloud.bigquery.datatransfer.v1.ListDataSourcesRequest- I would expect `com.google.cloud.bigquery.datatransfer.v1.ListDataSourcesRequest` to be a hyperlink to the ListDataSourcesRequest class, but it is not. A search within that page for ListDataSourcesRequest does not show any documentation for that class.
1.0
BigQuery Data Transfer: ListDataSourcesRequest not documented / not linked to - See: https://googlecloudplatform.github.io/google-cloud-java/latest/apidocs/com/google/cloud/bigquery/datatransfer/v1/DataTransferServiceClient.html#listDataSources-com.google.cloud.bigquery.datatransfer.v1.ListDataSourcesRequest- I would expect `com.google.cloud.bigquery.datatransfer.v1.ListDataSourcesRequest` to be a hyperlink to the ListDataSourcesRequest class, but it is not. A search within that page for ListDataSourcesRequest does not show any documentation for that class.
non_process
bigquery data transfer listdatasourcesrequest not documented not linked to see i would expect com google cloud bigquery datatransfer listdatasourcesrequest to be a hyperlink to the listdatasourcesrequest class but it is not a search within that page for listdatasourcesrequest does not show any documentation for that class
0
59,878
3,117,581,479
IssuesEvent
2015-09-04 02:52:40
framingeinstein/issues-test
https://api.github.com/repos/framingeinstein/issues-test
opened
SPK-215: Add to Wishlist: Message contains a link that merges with background color
priority:normal priority:normal priority:normal priority:normal priority:normal resolution:fixed resolution:fixed
"Item has been added to your wishlist. Click here to continue shopping" message contains a link ("here") that has the same color as the background. Steps: 1. Go to http://qa.speakman.rallyapp.co/shower-and-bath/ 2. Add any item to Wishlist 3. Note message that appears on top of the page Browsers/OS: Firefox 37 Mac OS X 10.9.5 Safari 7.1 Mac OS X 10.9.5 Chrome 41 Mac OS X 10.9.4
5.0
SPK-215: Add to Wishlist: Message contains a link that merges with background color - "Item has been added to your wishlist. Click here to continue shopping" message contains a link ("here") that has the same color as the background. Steps: 1. Go to http://qa.speakman.rallyapp.co/shower-and-bath/ 2. Add any item to Wishlist 3. Note message that appears on top of the page Browsers/OS: Firefox 37 Mac OS X 10.9.5 Safari 7.1 Mac OS X 10.9.5 Chrome 41 Mac OS X 10.9.4
non_process
spk add to wishlist message contains a link that merges with background color item has been added to your wishlist click here to continue shopping message contains a link here that has the same color as the background steps go to add any item to wishlist note message that appears on top of the page browsers os firefox mac os x safari mac os x chrome mac os x
0
3,670
6,705,664,517
IssuesEvent
2017-10-12 01:49:32
zero-os/0-Disk
https://api.github.com/repos/zero-os/0-Disk
closed
Add some testable (godoc) examples to 0-Disk packages
process_duplicate process_wontfix type_feature
A golang feature we haven't made use of yet, is testable examples. It allows us to write examples which will be available in the godoc documentation for the relevant package, while the `go test` tool ensures those examples are actually valid. More information: https://blog.golang.org/examples Packages where we should add such examples: - [x] github.com/zero-os/0-Disk + how to create a new hash, how to create a hash from bytes, how would you use the Hash methods afterwards; + how to create and use a hasher; + how to create and use a keyed hasher; - [x] ~github.com/zero-os/0-Disk/log~ + ~how to create a logger and how to use it;~ + ~how to use and configure the global logger;~ + ~how to broadcast errors;~ - [x] github.com/zero-os/0-Disk/nbd/ardb/storage + how to create a storage for a given vdisk; + how to use a (block)storage;
2.0
Add some testable (godoc) examples to 0-Disk packages - A golang feature we haven't made use of yet, is testable examples. It allows us to write examples which will be available in the godoc documentation for the relevant package, while the `go test` tool ensures those examples are actually valid. More information: https://blog.golang.org/examples Packages where we should add such examples: - [x] github.com/zero-os/0-Disk + how to create a new hash, how to create a hash from bytes, how would you use the Hash methods afterwards; + how to create and use a hasher; + how to create and use a keyed hasher; - [x] ~github.com/zero-os/0-Disk/log~ + ~how to create a logger and how to use it;~ + ~how to use and configure the global logger;~ + ~how to broadcast errors;~ - [x] github.com/zero-os/0-Disk/nbd/ardb/storage + how to create a storage for a given vdisk; + how to use a (block)storage;
process
add some testable godoc examples to disk packages a golang feature we haven t made use of yet is testable examples it allows us to write examples which will be available in the godoc documentation for the relevant package while the go test tool ensures those examples are actually valid more information packages where we should add such examples github com zero os disk how to create a new hash how to create a hash from bytes how would you use the hash methods afterwards how to create and use a hasher how to create and use a keyed hasher github com zero os disk log how to create a logger and how to use it how to use and configure the global logger how to broadcast errors github com zero os disk nbd ardb storage how to create a storage for a given vdisk how to use a block storage
1
1,743
10,678,143,835
IssuesEvent
2019-10-21 16:41:49
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
Set flank version in ui-test/Dockerfile as ENV var
eng:automation 🌟 feature request
Currently we download the flank jar file once when the Docker image is built. To have more control over this, we'd like to pass in the desired flank version num to the Dockerfile as an env var.
1.0
Set flank version in ui-test/Dockerfile as ENV var - Currently we download the flank jar file once when the Docker image is built. To have more control over this, we'd like to pass in the desired flank version num to the Dockerfile as an env var.
non_process
set flank version in ui test dockerfile as env var currently we download the flank jar file once when the docker image is built to have more control over this we d like to pass in the desired flank version num to the dockerfile as an env var
0
16,173
20,618,541,731
IssuesEvent
2022-03-07 15:21:39
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject test failed
area-System.Diagnostics.Process blocking-clean-ci untriaged
Configuration: `net7.0-Linux-Release-x64-CoreCLR_checked-Ubuntu.1804.Amd64.Open` Build: [runtime pipeline of the rolling build](https://dev.azure.com/dnceng/public/_build/results?buildId=1635569&view=ms.vss-test-web.build-test-results-tab&runId=45291136&paneView=dotnet-dnceng.dnceng-build-release-tasks.helix-test-information-tab&resultId=147215), [console log](https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-f713537715f04c66ac/System.Diagnostics.Process.Tests/1/console.34f7ff4c.log?sv=2019-07-07&se=2022-03-20T08%3A53%3A47Z&sr=c&sp=rl&sig=%2FCxSnOmgSEXdBGlx6bMmwpMV0AFOlnBRyg8Qd6IQSU8%3D) ``` System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject(invokeRequired: False) [FAIL] Assert.Equal() Failure ↓ (pos 105) Expected: ··· closed error stream Actual: ··· closed error streamAssert failure(PID 18855 [0x000049a7], Th··· ↑ (pos 105) Stack Trace: /_/src/libraries/System.Diagnostics.Process/tests/ProcessStreamReadTests.cs(85,0): at System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject(Boolean invokeRequired) ... Finished: System.Diagnostics.Process.Tests === TEST EXECUTION SUMMARY === System.Diagnostics.Process.Tests Total: 317, Errors: 0, Failed: 1, Skipped: 2, Time: 42.933s ```
1.0
System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject test failed - Configuration: `net7.0-Linux-Release-x64-CoreCLR_checked-Ubuntu.1804.Amd64.Open` Build: [runtime pipeline of the rolling build](https://dev.azure.com/dnceng/public/_build/results?buildId=1635569&view=ms.vss-test-web.build-test-results-tab&runId=45291136&paneView=dotnet-dnceng.dnceng-build-release-tasks.helix-test-information-tab&resultId=147215), [console log](https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-f713537715f04c66ac/System.Diagnostics.Process.Tests/1/console.34f7ff4c.log?sv=2019-07-07&se=2022-03-20T08%3A53%3A47Z&sr=c&sp=rl&sig=%2FCxSnOmgSEXdBGlx6bMmwpMV0AFOlnBRyg8Qd6IQSU8%3D) ``` System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject(invokeRequired: False) [FAIL] Assert.Equal() Failure ↓ (pos 105) Expected: ··· closed error stream Actual: ··· closed error streamAssert failure(PID 18855 [0x000049a7], Th··· ↑ (pos 105) Stack Trace: /_/src/libraries/System.Diagnostics.Process/tests/ProcessStreamReadTests.cs(85,0): at System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject(Boolean invokeRequired) ... Finished: System.Diagnostics.Process.Tests === TEST EXECUTION SUMMARY === System.Diagnostics.Process.Tests Total: 317, Errors: 0, Failed: 1, Skipped: 2, Time: 42.933s ```
process
system diagnostics tests processstreamreadtests testasyncerrorstream synchronizingobject test failed configuration linux release coreclr checked ubuntu open build system diagnostics tests processstreamreadtests testasyncerrorstream synchronizingobject invokerequired false assert equal failure ↓ pos expected ··· closed error stream actual ··· closed error streamassert failure pid th··· ↑ pos stack trace src libraries system diagnostics process tests processstreamreadtests cs at system diagnostics tests processstreamreadtests testasyncerrorstream synchronizingobject boolean invokerequired finished system diagnostics process tests test execution summary system diagnostics process tests total errors failed skipped time
1
21,356
29,189,101,062
IssuesEvent
2023-05-19 18:08:01
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Implement `remove` and `replace`
.metabase-lib .Team/QueryProcessor :hammer_and_wrench:
### Related PRs - https://github.com/metabase/metabase/pull/29625 - https://github.com/metabase/metabase/pull/29657 - https://github.com/metabase/metabase/pull/29775 - https://github.com/metabase/metabase/pull/29777 - https://github.com/metabase/metabase/pull/29803 - https://github.com/metabase/metabase/pull/29841 - https://github.com/metabase/metabase/pull/30002 - https://github.com/metabase/metabase/pull/30074 - https://github.com/metabase/metabase/pull/30373 - https://github.com/metabase/metabase/pull/30775 ### `remove` takes a clause and removes it. I think this should probably take `stage-number` as an optional parameter, defaulting to `-1` (last stage of the query) if not explicitly specified, the same way literally everything else does. Since we have UUIDs (see below) I guess it would be possible to walk all the stages and find something to remove, but this seems more complicated than we need right now ```clj (let [my-order-by (first (order-bys query -1))] (remove query my-order-by)) ``` or ```js const orderBy = query.orderBys()?.[0]; query = query.remove(orderBy); ``` The whole reason we added `:lib/uuid` in the options in the first place was to facilitate this sort of stuff, so we should obviously leverage this info. I think we'll want to have some sort of underlying `remove*` multimethod that we can implement to handle various MBQL clauses and maps. e.g. we'd have an `:asc` method which would then go look in `:order-by` and find the matching `:asc` with the same `:lib/uuid` and remove that from the query. Some stuff like `:field` clauses would probably have to look in multiple places, e.g. both `:fields` and `:breakout`. On that note, if we're not already doing it, we should enforce a rule against duplicate UUIDs in a query stage (ignoring metadata) Removal methods that affect the columns that will be in the response should update `:lib/stage-metadata` for the affected and subsequent stages, e.g. if I remove something from `:fields` or `:breakout` the `:lib/stage-metadata` should reflect this. Open question: should we "clean" the query and remove orphaned references? E.g. if I have a join and add a filter against a joined field, then remove the join, should we do a "cascading delete" and remove the filter as well? I am about 95% sure the answer should be yes here. So let's implement that too. This will have to be done on a case-by-case basis for different `remove*` impls ### `replace` This is basically the same as `remove`, but instead of removing it just replaces the clause in that location with a different clause ```clj (let [my-order-by (first (lib/order-bys query -1)) new-order-by (lib/order-by-clause (lib/field whatever))] (lib/replace query my-order-by new-order-by)) ``` I guess this should probably be implemented with an underlying multimethod as well. This also needs to update metadata and "clean" the query where appropriate. Question: can we be smart and instead of removing stuff, update it to the equivalent when possible? E.g. if I join table X and have a filter where X.ID = 1, and I replace the join with one against Table Y, which also has ID, can we update the filter to be Y.ID = 1? That would be **smart AF**. I think we should try to do that as a best-effort thing.
1.0
[MLv2] Implement `remove` and `replace` - ### Related PRs - https://github.com/metabase/metabase/pull/29625 - https://github.com/metabase/metabase/pull/29657 - https://github.com/metabase/metabase/pull/29775 - https://github.com/metabase/metabase/pull/29777 - https://github.com/metabase/metabase/pull/29803 - https://github.com/metabase/metabase/pull/29841 - https://github.com/metabase/metabase/pull/30002 - https://github.com/metabase/metabase/pull/30074 - https://github.com/metabase/metabase/pull/30373 - https://github.com/metabase/metabase/pull/30775 ### `remove` takes a clause and removes it. I think this should probably take `stage-number` as an optional parameter, defaulting to `-1` (last stage of the query) if not explicitly specified, the same way literally everything else does. Since we have UUIDs (see below) I guess it would be possible to walk all the stages and find something to remove, but this seems more complicated than we need right now ```clj (let [my-order-by (first (order-bys query -1))] (remove query my-order-by)) ``` or ```js const orderBy = query.orderBys()?.[0]; query = query.remove(orderBy); ``` The whole reason we added `:lib/uuid` in the options in the first place was to facilitate this sort of stuff, so we should obviously leverage this info. I think we'll want to have some sort of underlying `remove*` multimethod that we can implement to handle various MBQL clauses and maps. e.g. we'd have an `:asc` method which would then go look in `:order-by` and find the matching `:asc` with the same `:lib/uuid` and remove that from the query. Some stuff like `:field` clauses would probably have to look in multiple places, e.g. both `:fields` and `:breakout`. On that note, if we're not already doing it, we should enforce a rule against duplicate UUIDs in a query stage (ignoring metadata) Removal methods that affect the columns that will be in the response should update `:lib/stage-metadata` for the affected and subsequent stages, e.g. if I remove something from `:fields` or `:breakout` the `:lib/stage-metadata` should reflect this. Open question: should we "clean" the query and remove orphaned references? E.g. if I have a join and add a filter against a joined field, then remove the join, should we do a "cascading delete" and remove the filter as well? I am about 95% sure the answer should be yes here. So let's implement that too. This will have to be done on a case-by-case basis for different `remove*` impls ### `replace` This is basically the same as `remove`, but instead of removing it just replaces the clause in that location with a different clause ```clj (let [my-order-by (first (lib/order-bys query -1)) new-order-by (lib/order-by-clause (lib/field whatever))] (lib/replace query my-order-by new-order-by)) ``` I guess this should probably be implemented with an underlying multimethod as well. This also needs to update metadata and "clean" the query where appropriate. Question: can we be smart and instead of removing stuff, update it to the equivalent when possible? E.g. if I join table X and have a filter where X.ID = 1, and I replace the join with one against Table Y, which also has ID, can we update the filter to be Y.ID = 1? That would be **smart AF**. I think we should try to do that as a best-effort thing.
process
implement remove and replace related prs remove takes a clause and removes it i think this should probably take stage number as an optional parameter defaulting to last stage of the query if not explicitly specified the same way literally everything else does since we have uuids see below i guess it would be possible to walk all the stages and find something to remove but this seems more complicated than we need right now clj let remove query my order by or js const orderby query orderbys query query remove orderby the whole reason we added lib uuid in the options in the first place was to facilitate this sort of stuff so we should obviously leverage this info i think we ll want to have some sort of underlying remove multimethod that we can implement to handle various mbql clauses and maps e g we d have an asc method which would then go look in order by and find the matching asc with the same lib uuid and remove that from the query some stuff like field clauses would probably have to look in multiple places e g both fields and breakout on that note if we re not already doing it we should enforce a rule against duplicate uuids in a query stage ignoring metadata removal methods that affect the columns that will be in the response should update lib stage metadata for the affected and subsequent stages e g if i remove something from fields or breakout the lib stage metadata should reflect this open question should we clean the query and remove orphaned references e g if i have a join and add a filter against a joined field then remove the join should we do a cascading delete and remove the filter as well i am about sure the answer should be yes here so let s implement that too this will have to be done on a case by case basis for different remove impls replace this is basically the same as remove but instead of removing it just replaces the clause in that location with a different clause clj let my order by first lib order bys query new order by lib order by clause lib field whatever lib replace query my order by new order by i guess this should probably be implemented with an underlying multimethod as well this also needs to update metadata and clean the query where appropriate question can we be smart and instead of removing stuff update it to the equivalent when possible e g if i join table x and have a filter where x id and i replace the join with one against table y which also has id can we update the filter to be y id that would be smart af i think we should try to do that as a best effort thing
1
21,708
30,208,063,370
IssuesEvent
2023-07-05 10:48:58
q191201771/lal
https://api.github.com/repos/q191201771/lal
closed
demo中的pullrtmp2hls无法工作
#Bug *In process * help wanted
pullrtmp2hls 和 -i rtsp 做连接 直接crash。 ### 2021/07/23 23:05:19.665373  INFO [RTMPPULL1] < R _result(). - client_session.go:479 2021/07/23 23:05:19.665373  INFO [RTMPPULL1] > W play('mystream'). - client_session.go:482 2021/07/23 23:05:19.665891  WARN [RTMPPULL1] read user control message, ignore. buf=00000000 00 04 00 00 00 01 |......| - client_session.go:350 2021/07/23 23:05:19.666407  WARN [RTMPPULL1] read user control message, ignore. buf=00000000 00 00 00 00 00 01 |......| - client_session.go:350 2021/07/23 23:05:19.666925  INFO [RTMPPULL1] < R onStatus('NetStream.Play.Start'). - client_session.go:435 panic: runtime error: invalid memory address or nil pointer dereference [signal 0xc0000005 code=0x0 addr=0x18 pc=0xdb73a8] ### 比较容易重现的。 lalserver做服务,pullrtmp2hls 和lalserver连接。
1.0
demo中的pullrtmp2hls无法工作 - pullrtmp2hls 和 -i rtsp 做连接 直接crash。 ### 2021/07/23 23:05:19.665373  INFO [RTMPPULL1] < R _result(). - client_session.go:479 2021/07/23 23:05:19.665373  INFO [RTMPPULL1] > W play('mystream'). - client_session.go:482 2021/07/23 23:05:19.665891  WARN [RTMPPULL1] read user control message, ignore. buf=00000000 00 04 00 00 00 01 |......| - client_session.go:350 2021/07/23 23:05:19.666407  WARN [RTMPPULL1] read user control message, ignore. buf=00000000 00 00 00 00 00 01 |......| - client_session.go:350 2021/07/23 23:05:19.666925  INFO [RTMPPULL1] < R onStatus('NetStream.Play.Start'). - client_session.go:435 panic: runtime error: invalid memory address or nil pointer dereference [signal 0xc0000005 code=0x0 addr=0x18 pc=0xdb73a8] ### 比较容易重现的。 lalserver做服务,pullrtmp2hls 和lalserver连接。
process
和 i rtsp 做连接 直接crash。  r result client session go  w play mystream client session go  read user control message ignore buf client session go  read user control message ignore buf client session go  r onstatus netstream play start client session go panic runtime error invalid memory address or nil pointer dereference 比较容易重现的。 lalserver做服务, 和lalserver连接。
1
20,109
26,648,502,260
IssuesEvent
2023-01-25 11:53:56
firebase/firebase-cpp-sdk
https://api.github.com/repos/firebase/firebase-cpp-sdk
reopened
[C++] Nightly Integration Testing Report for Firestore
type: process nightly-testing
<hidden value="integration-test-status-comment"></hidden> ### ✅&nbsp; [build against repo] Integration test succeeded! Requested by @sunmou99 on commit 9c494f708c5918d1b82e9624460b7ab4bbaf8431 Last updated: Tue Jan 24 04:03 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3995022157)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit 9c494f708c5918d1b82e9624460b7ab4bbaf8431 Last updated: Tue Jan 24 15:17 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3996121210)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against tip] Integration test succeeded! Requested by @sunmou99 on commit 21f954fa9d0300d3e68ee16d557eac451136da26 Last updated: Wed Jan 25 03:51 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4005239522)**
1.0
[C++] Nightly Integration Testing Report for Firestore - <hidden value="integration-test-status-comment"></hidden> ### ✅&nbsp; [build against repo] Integration test succeeded! Requested by @sunmou99 on commit 9c494f708c5918d1b82e9624460b7ab4bbaf8431 Last updated: Tue Jan 24 04:03 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3995022157)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit 9c494f708c5918d1b82e9624460b7ab4bbaf8431 Last updated: Tue Jan 24 15:17 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3996121210)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against tip] Integration test succeeded! Requested by @sunmou99 on commit 21f954fa9d0300d3e68ee16d557eac451136da26 Last updated: Wed Jan 25 03:51 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4005239522)**
process
nightly integration testing report for firestore ✅ nbsp integration test succeeded requested by on commit last updated tue jan pst ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated tue jan pst ✅ nbsp integration test succeeded requested by on commit last updated wed jan pst
1
77,820
3,507,278,186
IssuesEvent
2016-01-08 12:21:14
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Client - Lagging of core (BB #802)
Category: Exploit migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** smoldar **Original Date:** 29.01.2015 22:32:32 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** closed **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/802 <hr> Player tryes to login again, AccountId = XXX Player tryes to login again, AccountId = XXX Player tryes to login again, AccountId = XXX Player tryes to login again, AccountId = XXX 100x spammed in server.log / sec i think about packet attack of core, when this spam runned, core was laggy 50-5000ms I think this problem is here, this may be blocked by different way i think. Destroying of client session, or something else to avoid laggy, because client can repeatly spam core. CharacterHandler.cpp ``` #!c++ void WorldSession::HandlePlayerLoginOpcode(WorldPacket & recv_data) { if (PlayerLoading() || GetPlayer() != NULL) { sLog.outError("Player tryes to login again, AccountId = %d",GetAccountId()); return; } ```
1.0
Client - Lagging of core (BB #802) - This issue was migrated from bitbucket. **Original Reporter:** smoldar **Original Date:** 29.01.2015 22:32:32 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** closed **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/802 <hr> Player tryes to login again, AccountId = XXX Player tryes to login again, AccountId = XXX Player tryes to login again, AccountId = XXX Player tryes to login again, AccountId = XXX 100x spammed in server.log / sec i think about packet attack of core, when this spam runned, core was laggy 50-5000ms I think this problem is here, this may be blocked by different way i think. Destroying of client session, or something else to avoid laggy, because client can repeatly spam core. CharacterHandler.cpp ``` #!c++ void WorldSession::HandlePlayerLoginOpcode(WorldPacket & recv_data) { if (PlayerLoading() || GetPlayer() != NULL) { sLog.outError("Player tryes to login again, AccountId = %d",GetAccountId()); return; } ```
non_process
client lagging of core bb this issue was migrated from bitbucket original reporter smoldar original date gmt original priority major original type bug original state closed direct link player tryes to login again accountid xxx player tryes to login again accountid xxx player tryes to login again accountid xxx player tryes to login again accountid xxx spammed in server log sec i think about packet attack of core when this spam runned core was laggy i think this problem is here this may be blocked by different way i think destroying of client session or something else to avoid laggy because client can repeatly spam core characterhandler cpp c void worldsession handleplayerloginopcode worldpacket recv data if playerloading getplayer null slog outerror player tryes to login again accountid d getaccountid return
0
17,860
23,807,284,441
IssuesEvent
2022-09-04 08:22:10
anitsh/til
https://api.github.com/repos/anitsh/til
opened
Kanban Maturity Model and The 4-Step Process to defect management by Nave
agile quality qa process
Introduce agile techniques like Definition of Done (DoD), code reviews and pair programming to make sure you eliminate defects at an early development stage. Your team will become more self-managed, motivated and engaged. Shorter lead times lead to faster innovation, greater customer satisfaction and substantial profitability. Focus on making sure the defects do not occur in the first place or, at least, the defects are significantly reduced before making their way into production. Defect management will help you increase the consistency, predictability and effectiveness of your development efforts, and it will ultimately reduce your delivery times. The Kanban Maturity Model describes the general practices of reducing defects to help you improve your business outcomes. Here is the 4-step process which will enable you to achieve a more sustainable delivery of quality software. Step 1: Visualize Defects Step 2: Track Defects Step 3: Perform Defect Clustering Analysis Step 4: Define Relevant Policies Step 1: Visualize Defects We cannot manage what we cannot see. Increasing transparency is the first step towards reducing defects and raising awareness of the disruptive impact they may have. When you’re testing a work item, and a new defect is discovered, create a separate ticket and mark it as a defect. Add the following essential information to the ticket: Add a brief description of the defect. Add a link to the work item the defect is associated with. Specify the reason that has caused the defect. Assign the people responsible for fixing the defect. Assign severity. Use a different color for defects to make them easily recognizable from other types of work. Step 2: Track Defects In the book “The Economics of Software Quality”, defect tracking is used in more than 55% of military and defense software applications, but in less than 15% of IT, commercial, web or embedded applications. Given their prevention effectiveness, it is surprising that defect measurement is not used more often. The second step of the defect management process is to track and analyze software defects on a regular basis. The Throughput Run Chart displays the throughput of your team on a daily, weekly or monthly basis and compares its values over time. Use it to keep track of the number of defects you resolve and assess how trends have developed. The Throughput Breakdown Chart can help you evaluate what type of work is demanding the bulk of the time. It is especially useful when you want to emphasize the high volume of defects you’re handling. Step 3: Perform Defect Clustering Analysis The third step of the defect management process is to identify the clusters of causes hindering your delivery times. To perform defect clustering analysis: Evaluate the causes of defects and their resolution times. Group the causes by clustering the defects which are similar in nature. Capture the total cycle time for each group to discover which are the most critical causes of defects. Report the defect clustering results at the Service Delivery Review, Risk Review, and Operations Review. That way, you’ll keep your attention on the defects with the highest impact on your development process, and you’ll emphasize the urgency to act upon them sooner. Step 4: Define Relevant Policies The goal of this step is to bring continuous improvement to your development practices and help you maintain a stable, predictable workflow in the long run. Introduce policies on how defects should be handled. A defect ticket should be attached to the work item that it belongs to. The work item should then remain in the same process state where the problem was first discovered. Create a new ticket for your defect and associate it with the work item that spawned it. Place the new ticket in the rework queue state (i.e. Development Queue). Block the work item in the current process state (i.e. Testing). By adding new items into your workflow, you count them against your WIP limit. An excessive number of defects will then block the pull of new work until any quality issues are first resolved. This approach will also help you avoid the common anti-pattern of moving work backward on your Kanban board. Make sure your activities are not just targeted at finding defects after they have happened. https://www.kanbanmaturitymodel.com https://getnave.com/throughput-run-chart https://www.infoq.com/articles/blockers-defects-process-improvement
1.0
Kanban Maturity Model and The 4-Step Process to defect management by Nave - Introduce agile techniques like Definition of Done (DoD), code reviews and pair programming to make sure you eliminate defects at an early development stage. Your team will become more self-managed, motivated and engaged. Shorter lead times lead to faster innovation, greater customer satisfaction and substantial profitability. Focus on making sure the defects do not occur in the first place or, at least, the defects are significantly reduced before making their way into production. Defect management will help you increase the consistency, predictability and effectiveness of your development efforts, and it will ultimately reduce your delivery times. The Kanban Maturity Model describes the general practices of reducing defects to help you improve your business outcomes. Here is the 4-step process which will enable you to achieve a more sustainable delivery of quality software. Step 1: Visualize Defects Step 2: Track Defects Step 3: Perform Defect Clustering Analysis Step 4: Define Relevant Policies Step 1: Visualize Defects We cannot manage what we cannot see. Increasing transparency is the first step towards reducing defects and raising awareness of the disruptive impact they may have. When you’re testing a work item, and a new defect is discovered, create a separate ticket and mark it as a defect. Add the following essential information to the ticket: Add a brief description of the defect. Add a link to the work item the defect is associated with. Specify the reason that has caused the defect. Assign the people responsible for fixing the defect. Assign severity. Use a different color for defects to make them easily recognizable from other types of work. Step 2: Track Defects In the book “The Economics of Software Quality”, defect tracking is used in more than 55% of military and defense software applications, but in less than 15% of IT, commercial, web or embedded applications. Given their prevention effectiveness, it is surprising that defect measurement is not used more often. The second step of the defect management process is to track and analyze software defects on a regular basis. The Throughput Run Chart displays the throughput of your team on a daily, weekly or monthly basis and compares its values over time. Use it to keep track of the number of defects you resolve and assess how trends have developed. The Throughput Breakdown Chart can help you evaluate what type of work is demanding the bulk of the time. It is especially useful when you want to emphasize the high volume of defects you’re handling. Step 3: Perform Defect Clustering Analysis The third step of the defect management process is to identify the clusters of causes hindering your delivery times. To perform defect clustering analysis: Evaluate the causes of defects and their resolution times. Group the causes by clustering the defects which are similar in nature. Capture the total cycle time for each group to discover which are the most critical causes of defects. Report the defect clustering results at the Service Delivery Review, Risk Review, and Operations Review. That way, you’ll keep your attention on the defects with the highest impact on your development process, and you’ll emphasize the urgency to act upon them sooner. Step 4: Define Relevant Policies The goal of this step is to bring continuous improvement to your development practices and help you maintain a stable, predictable workflow in the long run. Introduce policies on how defects should be handled. A defect ticket should be attached to the work item that it belongs to. The work item should then remain in the same process state where the problem was first discovered. Create a new ticket for your defect and associate it with the work item that spawned it. Place the new ticket in the rework queue state (i.e. Development Queue). Block the work item in the current process state (i.e. Testing). By adding new items into your workflow, you count them against your WIP limit. An excessive number of defects will then block the pull of new work until any quality issues are first resolved. This approach will also help you avoid the common anti-pattern of moving work backward on your Kanban board. Make sure your activities are not just targeted at finding defects after they have happened. https://www.kanbanmaturitymodel.com https://getnave.com/throughput-run-chart https://www.infoq.com/articles/blockers-defects-process-improvement
process
kanban maturity model and the step process to defect management by nave introduce agile techniques like definition of done dod code reviews and pair programming to make sure you eliminate defects at an early development stage your team will become more self managed motivated and engaged shorter lead times lead to faster innovation greater customer satisfaction and substantial profitability focus on making sure the defects do not occur in the first place or at least the defects are significantly reduced before making their way into production defect management will help you increase the consistency predictability and effectiveness of your development efforts and it will ultimately reduce your delivery times the kanban maturity model describes the general practices of reducing defects to help you improve your business outcomes here is the step process which will enable you to achieve a more sustainable delivery of quality software step visualize defects step track defects step perform defect clustering analysis step define relevant policies step visualize defects we cannot manage what we cannot see increasing transparency is the first step towards reducing defects and raising awareness of the disruptive impact they may have when you’re testing a work item and a new defect is discovered create a separate ticket and mark it as a defect add the following essential information to the ticket add a brief description of the defect add a link to the work item the defect is associated with specify the reason that has caused the defect assign the people responsible for fixing the defect assign severity use a different color for defects to make them easily recognizable from other types of work step track defects in the book “the economics of software quality” defect tracking is used in more than of military and defense software applications but in less than of it commercial web or embedded applications given their prevention effectiveness it is surprising that defect measurement is not used more often the second step of the defect management process is to track and analyze software defects on a regular basis the throughput run chart displays the throughput of your team on a daily weekly or monthly basis and compares its values over time use it to keep track of the number of defects you resolve and assess how trends have developed the throughput breakdown chart can help you evaluate what type of work is demanding the bulk of the time it is especially useful when you want to emphasize the high volume of defects you’re handling step perform defect clustering analysis the third step of the defect management process is to identify the clusters of causes hindering your delivery times to perform defect clustering analysis evaluate the causes of defects and their resolution times group the causes by clustering the defects which are similar in nature capture the total cycle time for each group to discover which are the most critical causes of defects report the defect clustering results at the service delivery review risk review and operations review that way you’ll keep your attention on the defects with the highest impact on your development process and you’ll emphasize the urgency to act upon them sooner step define relevant policies the goal of this step is to bring continuous improvement to your development practices and help you maintain a stable predictable workflow in the long run introduce policies on how defects should be handled a defect ticket should be attached to the work item that it belongs to the work item should then remain in the same process state where the problem was first discovered create a new ticket for your defect and associate it with the work item that spawned it place the new ticket in the rework queue state i e development queue block the work item in the current process state i e testing by adding new items into your workflow you count them against your wip limit an excessive number of defects will then block the pull of new work until any quality issues are first resolved this approach will also help you avoid the common anti pattern of moving work backward on your kanban board make sure your activities are not just targeted at finding defects after they have happened
1
19,135
25,190,171,054
IssuesEvent
2022-11-11 23:14:27
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Release 0.11.0
enhancement P1 process
### Problem v0.11.0 is not fully released ### Solution Deploy 0.11.0 ### Alternatives _No response_
1.0
Release 0.11.0 - ### Problem v0.11.0 is not fully released ### Solution Deploy 0.11.0 ### Alternatives _No response_
process
release problem is not fully released solution deploy alternatives no response
1
252,547
21,582,953,335
IssuesEvent
2022-05-02 20:52:03
damccorm/test-migration-target
https://api.github.com/repos/damccorm/test-migration-target
opened
ExternalPythonTransformTest.trivialPythonTransform flaky
bug test-failures cross-language P2
Example run: https://ci-beam.apache.org/job/beam_PreCommit_Java_Phrase/4806/testReport/junit/org.apache.beam.sdk.extensions.python/ExternalPythonTransformTest/trivialPythonTransform/ {noformat} java.lang.RuntimeException: java.util.concurrent.TimeoutException: Timeout waiting for Python service startup after 16616 seconds. at org.apache.beam.sdk.extensions.python.ExternalPythonTransform.expand(ExternalPythonTransform.java:107) at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:548) at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:482) at org.apache.beam.sdk.values.PCollection.apply(PCollection.java:363) at org.apache.beam.sdk.extensions.python.ExternalPythonTransformTest.trivialPythonTransform(ExternalPythonTransformTest.java:41) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) {noformat} Imported from Jira [BEAM-14148](https://issues.apache.org/jira/browse/BEAM-14148). Original Jira may contain additional context. Reported by: lcwik.
1.0
ExternalPythonTransformTest.trivialPythonTransform flaky - Example run: https://ci-beam.apache.org/job/beam_PreCommit_Java_Phrase/4806/testReport/junit/org.apache.beam.sdk.extensions.python/ExternalPythonTransformTest/trivialPythonTransform/ {noformat} java.lang.RuntimeException: java.util.concurrent.TimeoutException: Timeout waiting for Python service startup after 16616 seconds. at org.apache.beam.sdk.extensions.python.ExternalPythonTransform.expand(ExternalPythonTransform.java:107) at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:548) at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:482) at org.apache.beam.sdk.values.PCollection.apply(PCollection.java:363) at org.apache.beam.sdk.extensions.python.ExternalPythonTransformTest.trivialPythonTransform(ExternalPythonTransformTest.java:41) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) {noformat} Imported from Jira [BEAM-14148](https://issues.apache.org/jira/browse/BEAM-14148). Original Jira may contain additional context. Reported by: lcwik.
non_process
externalpythontransformtest trivialpythontransform flaky example run noformat java lang runtimeexception java util concurrent timeoutexception timeout waiting for python service startup after seconds at org apache beam sdk extensions python externalpythontransform expand externalpythontransform java at org apache beam sdk pipeline applyinternal pipeline java at org apache beam sdk pipeline applytransform pipeline java at org apache beam sdk values pcollection apply pcollection java at org apache beam sdk extensions python externalpythontransformtest trivialpythontransform externalpythontransformtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java noformat imported from jira original jira may contain additional context reported by lcwik
0
210,905
7,196,432,181
IssuesEvent
2018-02-05 02:54:12
wso2/message-broker
https://api.github.com/repos/wso2/message-broker
closed
Implement AMQP queue.delete frames
Complexity/Moderate Module/broker-amqp Priority/High Severity/Major Type/Improvement
**Description:** <!-- Give a brief description of the issue --> Implement queue.delete amqp frame to delete queue through an AMQP client
1.0
Implement AMQP queue.delete frames - **Description:** <!-- Give a brief description of the issue --> Implement queue.delete amqp frame to delete queue through an AMQP client
non_process
implement amqp queue delete frames description implement queue delete amqp frame to delete queue through an amqp client
0
91,870
3,863,516,295
IssuesEvent
2016-04-08 09:45:33
iamxavier/elmah
https://api.github.com/repos/iamxavier/elmah
closed
Add BCC to ELMAH
auto-migrated Priority-Medium Type-Enhancement
``` What new or enhanced feature are you proposing? Add the capability to use BCC for emailing errorlogs. What goal would this enhancement help you achieve? This would allow us to hide a developers distribution list. I have downloaded the source ELMAH-1.2-sp2-src and made changes to the attached file for your review as well. ``` Original issue reported on code.google.com by `michael....@gmail.com` on 8 Aug 2012 at 6:54 Attachments: * [ErrorMailModule.cs](https://storage.googleapis.com/google-code-attachments/elmah/issue-305/comment-0/ErrorMailModule.cs)
1.0
Add BCC to ELMAH - ``` What new or enhanced feature are you proposing? Add the capability to use BCC for emailing errorlogs. What goal would this enhancement help you achieve? This would allow us to hide a developers distribution list. I have downloaded the source ELMAH-1.2-sp2-src and made changes to the attached file for your review as well. ``` Original issue reported on code.google.com by `michael....@gmail.com` on 8 Aug 2012 at 6:54 Attachments: * [ErrorMailModule.cs](https://storage.googleapis.com/google-code-attachments/elmah/issue-305/comment-0/ErrorMailModule.cs)
non_process
add bcc to elmah what new or enhanced feature are you proposing add the capability to use bcc for emailing errorlogs what goal would this enhancement help you achieve this would allow us to hide a developers distribution list i have downloaded the source elmah src and made changes to the attached file for your review as well original issue reported on code google com by michael gmail com on aug at attachments
0
246,617
20,888,045,567
IssuesEvent
2022-03-23 08:08:01
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
Fail to import one .csv file in one SAS attached table with permissions 'Query, Add'
🧪 testing :gear: tables :beetle: regression
**Storage Explorer Version:** 1.24.0-dev **Build Number:** 20220322.15 **Branch:** main **Platform/OS:** Windows 10/Linux Ubuntu 20.04/MacOS Monterey 12.1 (Apple M1 Pro) **Architecture:** ia32\x64 **How Found:** Ad-hoc testing **Regression From:** Previous release (1.22.1) ## Steps to Reproduce ## 1. Expand one storage account -> Tables -> Create a table. 2. Right click the table -> Click 'Get Shared Access Signature...'. 3. Select 'Query, Add' permissions -> Create the SAS URL and copy it. 4. Attach the table using the SAS URL. 5. Go to the attached table -> Try to import one .csv file. 6. Check whether succeeds to import the file. ## Expected Experience ## Succeed to import the file. ## Actual Experience ## Fail to import the file. ![image](https://user-images.githubusercontent.com/87792676/159649021-d55a7df7-968f-4239-b8e4-c0624d006c84.png)
1.0
Fail to import one .csv file in one SAS attached table with permissions 'Query, Add' - **Storage Explorer Version:** 1.24.0-dev **Build Number:** 20220322.15 **Branch:** main **Platform/OS:** Windows 10/Linux Ubuntu 20.04/MacOS Monterey 12.1 (Apple M1 Pro) **Architecture:** ia32\x64 **How Found:** Ad-hoc testing **Regression From:** Previous release (1.22.1) ## Steps to Reproduce ## 1. Expand one storage account -> Tables -> Create a table. 2. Right click the table -> Click 'Get Shared Access Signature...'. 3. Select 'Query, Add' permissions -> Create the SAS URL and copy it. 4. Attach the table using the SAS URL. 5. Go to the attached table -> Try to import one .csv file. 6. Check whether succeeds to import the file. ## Expected Experience ## Succeed to import the file. ## Actual Experience ## Fail to import the file. ![image](https://user-images.githubusercontent.com/87792676/159649021-d55a7df7-968f-4239-b8e4-c0624d006c84.png)
non_process
fail to import one csv file in one sas attached table with permissions query add storage explorer version dev build number branch main platform os windows linux ubuntu macos monterey apple pro architecture how found ad hoc testing regression from previous release steps to reproduce expand one storage account tables create a table right click the table click get shared access signature select query add permissions create the sas url and copy it attach the table using the sas url go to the attached table try to import one csv file check whether succeeds to import the file expected experience succeed to import the file actual experience fail to import the file
0
1,941
4,769,492,271
IssuesEvent
2016-10-26 12:46:13
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
Increase amount of alba statistics in GUI to make it more transparent
process_wontfix type_enhancement
Currently you see the preset in use and policy in use. But this is too little information to keep the system rolling. A vital part to know if your backend is still ok, is to know: * How much data has been written in a certain policy * How healthy is my policy? (e.g. policy is will stop working after 6 disks or 2 nodes stop working) * ...
1.0
Increase amount of alba statistics in GUI to make it more transparent - Currently you see the preset in use and policy in use. But this is too little information to keep the system rolling. A vital part to know if your backend is still ok, is to know: * How much data has been written in a certain policy * How healthy is my policy? (e.g. policy is will stop working after 6 disks or 2 nodes stop working) * ...
process
increase amount of alba statistics in gui to make it more transparent currently you see the preset in use and policy in use but this is too little information to keep the system rolling a vital part to know if your backend is still ok is to know how much data has been written in a certain policy how healthy is my policy e g policy is will stop working after disks or nodes stop working
1
18,770
10,224,254,861
IssuesEvent
2019-08-16 12:12:27
ppy/osu
https://api.github.com/repos/ppy/osu
closed
Seeking in editor is laggy
editor pending review performance
**Describe the bug:** Map seeking is very laggy in the editor. **Screenshots or videos showing encountered issue:** https://streamable.com/0my1e **osu!lazer version:** 2019.807.0.0 **Logs:** [runtime.log](https://github.com/ppy/osu/files/3482264/runtime.log) [performance.log](https://github.com/ppy/osu/files/3482266/performance.log) [network.log](https://github.com/ppy/osu/files/3482267/network.log) [database.log](https://github.com/ppy/osu/files/3482268/database.log)
True
Seeking in editor is laggy - **Describe the bug:** Map seeking is very laggy in the editor. **Screenshots or videos showing encountered issue:** https://streamable.com/0my1e **osu!lazer version:** 2019.807.0.0 **Logs:** [runtime.log](https://github.com/ppy/osu/files/3482264/runtime.log) [performance.log](https://github.com/ppy/osu/files/3482266/performance.log) [network.log](https://github.com/ppy/osu/files/3482267/network.log) [database.log](https://github.com/ppy/osu/files/3482268/database.log)
non_process
seeking in editor is laggy describe the bug map seeking is very laggy in the editor screenshots or videos showing encountered issue osu lazer version logs
0
69,172
22,264,137,647
IssuesEvent
2022-06-10 05:23:32
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
"Mark as read" should not be shown after "Quick reply" from notification
T-Defect
### Steps to reproduce ![image](https://user-images.githubusercontent.com/14617699/172995986-95c249ea-64f1-480a-9de0-f5d40063fab4.png) 1. Receive an incoming message notification 2. Reply from notification ### Outcome #### What did you expect? Mark as read should not be offered after reply. #### What happened instead? Mark as read is offered after reply. ### Your phone model OP3T ### Operating system version Lineage+microG 18.1 on Android 11 ### Application version and app store Element 1.4.16 ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? No
1.0
"Mark as read" should not be shown after "Quick reply" from notification - ### Steps to reproduce ![image](https://user-images.githubusercontent.com/14617699/172995986-95c249ea-64f1-480a-9de0-f5d40063fab4.png) 1. Receive an incoming message notification 2. Reply from notification ### Outcome #### What did you expect? Mark as read should not be offered after reply. #### What happened instead? Mark as read is offered after reply. ### Your phone model OP3T ### Operating system version Lineage+microG 18.1 on Android 11 ### Application version and app store Element 1.4.16 ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? No
non_process
mark as read should not be shown after quick reply from notification steps to reproduce receive an incoming message notification reply from notification outcome what did you expect mark as read should not be offered after reply what happened instead mark as read is offered after reply your phone model operating system version lineage microg on android application version and app store element homeserver no response will you send logs no are you willing to provide a pr no
0
29,812
24,301,302,790
IssuesEvent
2022-09-29 14:03:25
ephios-dev/ephios
https://api.github.com/repos/ephios-dev/ephios
closed
Derive version from git tag
[P] nice to have [C] infrastructure
To prevent version bump commits it would be great to determine the target version from the current git tag. Poetry doesn't support that natively, but it will be possible with poetry 1.2 once it is released. Someone is building a plugin for that, the progess is tracked in https://github.com/mtkennerly/poetry-dynamic-versioning/issues/39
1.0
Derive version from git tag - To prevent version bump commits it would be great to determine the target version from the current git tag. Poetry doesn't support that natively, but it will be possible with poetry 1.2 once it is released. Someone is building a plugin for that, the progess is tracked in https://github.com/mtkennerly/poetry-dynamic-versioning/issues/39
non_process
derive version from git tag to prevent version bump commits it would be great to determine the target version from the current git tag poetry doesn t support that natively but it will be possible with poetry once it is released someone is building a plugin for that the progess is tracked in
0
430,125
12,440,737,397
IssuesEvent
2020-05-26 12:32:36
BgeeDB/bgee_apps
https://api.github.com/repos/BgeeDB/bgee_apps
closed
Provide anatomical homologies
priority: medium ui enhancement
In GitLab by @marcrr on Oct 15, 2015, 11:48 Provide ancestral_taxa_homology_annotations.tsv file on Bgee homepage. Provide small file with only multiple-entity annotations, under name such as "non trivial homologies".
1.0
Provide anatomical homologies - In GitLab by @marcrr on Oct 15, 2015, 11:48 Provide ancestral_taxa_homology_annotations.tsv file on Bgee homepage. Provide small file with only multiple-entity annotations, under name such as "non trivial homologies".
non_process
provide anatomical homologies in gitlab by marcrr on oct provide ancestral taxa homology annotations tsv file on bgee homepage provide small file with only multiple entity annotations under name such as non trivial homologies
0
2,826
5,780,515,406
IssuesEvent
2017-04-29 01:10:51
gaocegege/maintainer
https://api.github.com/repos/gaocegege/maintainer
closed
Fix Docker image because of gh-md-toc
priority/P1 process/wip type/bug
Now maintainer supports `readme toc`, but the related dependencies is not installed in Dockerfile.
1.0
Fix Docker image because of gh-md-toc - Now maintainer supports `readme toc`, but the related dependencies is not installed in Dockerfile.
process
fix docker image because of gh md toc now maintainer supports readme toc but the related dependencies is not installed in dockerfile
1
32,567
15,439,661,901
IssuesEvent
2021-03-08 01:01:03
labstack/echo
https://api.github.com/repos/labstack/echo
closed
Performance regression with echo `4.2.0`
performance
### Issue Description We have just updated to echo 4.2.0 and have seen a relatively high performance regression with that release. With bisecting it seems like commit 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4 has introduced a perfomance penalty of about 30-40%. The following benchstat has been generated with https://github.com/vishr/web-framework-benchmark comparing commit cf2fcad01b833d4bac6deb916940759cef3fb376 to 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4. ``` name old time/op new time/op delta EchoStatic-8 29.5µs ± 2% 38.1µs ± 2% +29.26% (p=0.000 n=9+8) EchoGitHubAPI-8 46.2µs ± 1% 62.8µs ± 1% +35.93% (p=0.000 n=8+9) EchoGplusAPI-8 2.29µs ± 4% 3.23µs ± 1% +41.25% (p=0.000 n=9+9) EchoParseAPI-8 3.99µs ± 1% 5.67µs ± 1% +42.00% (p=0.000 n=9+9) ``` ### Steps to reproduce Run https://github.com/vishr/web-framework-benchmark benchmark with commit 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4 and cf2fcad01b833d4bac6deb916940759cef3fb376 ### Version/commit `4.2.0` first commit with regression 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4
True
Performance regression with echo `4.2.0` - ### Issue Description We have just updated to echo 4.2.0 and have seen a relatively high performance regression with that release. With bisecting it seems like commit 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4 has introduced a perfomance penalty of about 30-40%. The following benchstat has been generated with https://github.com/vishr/web-framework-benchmark comparing commit cf2fcad01b833d4bac6deb916940759cef3fb376 to 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4. ``` name old time/op new time/op delta EchoStatic-8 29.5µs ± 2% 38.1µs ± 2% +29.26% (p=0.000 n=9+8) EchoGitHubAPI-8 46.2µs ± 1% 62.8µs ± 1% +35.93% (p=0.000 n=8+9) EchoGplusAPI-8 2.29µs ± 4% 3.23µs ± 1% +41.25% (p=0.000 n=9+9) EchoParseAPI-8 3.99µs ± 1% 5.67µs ± 1% +42.00% (p=0.000 n=9+9) ``` ### Steps to reproduce Run https://github.com/vishr/web-framework-benchmark benchmark with commit 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4 and cf2fcad01b833d4bac6deb916940759cef3fb376 ### Version/commit `4.2.0` first commit with regression 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4
non_process
performance regression with echo issue description we have just updated to echo and have seen a relatively high performance regression with that release with bisecting it seems like commit has introduced a perfomance penalty of about the following benchstat has been generated with comparing commit to name old time op new time op delta echostatic ± ± p n echogithubapi ± ± p n echogplusapi ± ± p n echoparseapi ± ± p n steps to reproduce run benchmark with commit and version commit first commit with regression
0
13,642
16,331,832,916
IssuesEvent
2021-05-12 10:10:00
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Merge GO:0052413 metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction & related
multi-species process
Pectin is only found in plants. ~Plants are not symbionts of other organisms.~ But there are no examples I know of where a plant synthesises or degrades another plant's cell wall. So those 4 terms should be merged or obsoleted: * GO:0052413 metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction * GO:0052341 catabolism by organism of cell wall pectin in other organism involved in symbiotic interaction * GO:0052412 metabolism by host of symbiont cell wall pectin * GO:0052357 catabolism by host of symbiont cell wall pectin Those can remain: * GO:0052179 metabolism by symbiont of host cell wall pectin * GO:0052011 catabolism by symbiont of host cell wall pectin @ValWood @mgiglio99 @dsiegele any preference for merge/obsolete? There are no annotations (I moved annotations from GO:0052413 to GO:0052011) Thanks, Pascale
1.0
Merge GO:0052413 metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction & related - Pectin is only found in plants. ~Plants are not symbionts of other organisms.~ But there are no examples I know of where a plant synthesises or degrades another plant's cell wall. So those 4 terms should be merged or obsoleted: * GO:0052413 metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction * GO:0052341 catabolism by organism of cell wall pectin in other organism involved in symbiotic interaction * GO:0052412 metabolism by host of symbiont cell wall pectin * GO:0052357 catabolism by host of symbiont cell wall pectin Those can remain: * GO:0052179 metabolism by symbiont of host cell wall pectin * GO:0052011 catabolism by symbiont of host cell wall pectin @ValWood @mgiglio99 @dsiegele any preference for merge/obsolete? There are no annotations (I moved annotations from GO:0052413 to GO:0052011) Thanks, Pascale
process
merge go metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction related pectin is only found in plants plants are not symbionts of other organisms but there are no examples i know of where a plant synthesises or degrades another plant s cell wall so those terms should be merged or obsoleted go metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction go catabolism by organism of cell wall pectin in other organism involved in symbiotic interaction go metabolism by host of symbiont cell wall pectin go catabolism by host of symbiont cell wall pectin those can remain go metabolism by symbiont of host cell wall pectin go catabolism by symbiont of host cell wall pectin valwood dsiegele any preference for merge obsolete there are no annotations i moved annotations from go to go thanks pascale
1
558
3,020,669,461
IssuesEvent
2015-07-31 09:34:28
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
Make use_calibration option compatible with ML
estimation preprocessor
ML currently requires to specify explicit starting values inside of the ```estimated_params``` block, rendering the ```estimated_params_init``` block useless.
1.0
Make use_calibration option compatible with ML - ML currently requires to specify explicit starting values inside of the ```estimated_params``` block, rendering the ```estimated_params_init``` block useless.
process
make use calibration option compatible with ml ml currently requires to specify explicit starting values inside of the estimated params block rendering the estimated params init block useless
1
21,597
30,000,989,022
IssuesEvent
2023-06-26 09:16:13
metallb/metallb
https://api.github.com/repos/metallb/metallb
reopened
Update MetalLB info in CNCF landscape
process lifecycle-stale
The information about MetalLB in the CNCF [landscape](https://landscape.cncf.io/) is outdated. https://github.com/cncf/landscape/blob/1873345e220ab784e20435768ad59b7b7fc3659a/landscape.yml#L1909-L1915 We should open a PR to address the following: - The project's homepage should be metallb.io (blocked by https://github.com/metallb/metallb/issues/691). - The logo should be updated in case we get a new design from the CNCF. - The Twitter handle (which is currently Dave's personal handle) should be either removed or updated in case we decide to manage a MetalLB Twitter account. - The Crunchbase link should be updated to `https://www.crunchbase.com/organization/cloud-native-computing-foundation`.
1.0
Update MetalLB info in CNCF landscape - The information about MetalLB in the CNCF [landscape](https://landscape.cncf.io/) is outdated. https://github.com/cncf/landscape/blob/1873345e220ab784e20435768ad59b7b7fc3659a/landscape.yml#L1909-L1915 We should open a PR to address the following: - The project's homepage should be metallb.io (blocked by https://github.com/metallb/metallb/issues/691). - The logo should be updated in case we get a new design from the CNCF. - The Twitter handle (which is currently Dave's personal handle) should be either removed or updated in case we decide to manage a MetalLB Twitter account. - The Crunchbase link should be updated to `https://www.crunchbase.com/organization/cloud-native-computing-foundation`.
process
update metallb info in cncf landscape the information about metallb in the cncf is outdated we should open a pr to address the following the project s homepage should be metallb io blocked by the logo should be updated in case we get a new design from the cncf the twitter handle which is currently dave s personal handle should be either removed or updated in case we decide to manage a metallb twitter account the crunchbase link should be updated to
1
18,950
24,910,734,528
IssuesEvent
2022-10-29 20:40:05
TUM-Dev/NavigaTUM
https://api.github.com/repos/TUM-Dev/NavigaTUM
closed
[Bug] WZW Bereich Mitte
bug webform delete-after-processing
Das Gewächshauslaborzentrum (und auch andere Gebäude) ist falsch verortet. Es liegt rund 2 km außerhalb von Freising. Ferner ist der Begriff WZW veraltet, denn das Wissenschaftszentrum Weihenstephan ist in der TUM School of Life Sciences aufgegangen.
1.0
[Bug] WZW Bereich Mitte - Das Gewächshauslaborzentrum (und auch andere Gebäude) ist falsch verortet. Es liegt rund 2 km außerhalb von Freising. Ferner ist der Begriff WZW veraltet, denn das Wissenschaftszentrum Weihenstephan ist in der TUM School of Life Sciences aufgegangen.
process
wzw bereich mitte das gewächshauslaborzentrum und auch andere gebäude ist falsch verortet es liegt rund km außerhalb von freising ferner ist der begriff wzw veraltet denn das wissenschaftszentrum weihenstephan ist in der tum school of life sciences aufgegangen
1
129,347
27,447,624,946
IssuesEvent
2023-03-02 15:16:27
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
Sourcegraph for VSCode extension doesn't render search page
bug vscode-extension team/code-exploration
- **Sourcegraph version:** <!-- the version of Sourcegraph or "Sourcegraph.com" --> VSCE v2.2.14 - **Platform information:** <!-- OS version, cloud provider, web browser version, Docker version, etc., depending on the issue --> VSCode v1.74.2 #### Steps to reproduce: 1. Install the latest VSCE extension (v2.2.14 for the time being). 2. Open it by clicking the Sourcegraph icon in the sidebar. 3. Check out the search page. #### Expected behavior: The search page is rendered (v2.2.13). <img width="1524" alt="Screenshot 2023-01-09 at 15 54 25" src="https://user-images.githubusercontent.com/25318659/211324192-4547d8ba-7340-4676-90e8-e5a440d440d0.png"> #### Actual behavior: The search page is not rendered (v2.2.14). <img width="1524" alt="Screenshot 2023-01-09 at 15 53 06" src="https://user-images.githubusercontent.com/25318659/211324254-04c3cd0f-4f88-433b-9d88-0a170f4729d7.png"> The following errors are thrown in the console in dev mode. <img width="1215" alt="Screenshot 2023-01-09 at 12 12 05" src="https://user-images.githubusercontent.com/25318659/211324352-2fcf53ce-2495-402e-b467-c2f85555fa0b.png"> If you would like immediate help on this, please email support@sourcegraph.com (you can still create the issue, but there are no [SLAs](https://about.sourcegraph.com/support/) on issues like there are for support requests). /cc @felixfbecker
2.0
Sourcegraph for VSCode extension doesn't render search page - - **Sourcegraph version:** <!-- the version of Sourcegraph or "Sourcegraph.com" --> VSCE v2.2.14 - **Platform information:** <!-- OS version, cloud provider, web browser version, Docker version, etc., depending on the issue --> VSCode v1.74.2 #### Steps to reproduce: 1. Install the latest VSCE extension (v2.2.14 for the time being). 2. Open it by clicking the Sourcegraph icon in the sidebar. 3. Check out the search page. #### Expected behavior: The search page is rendered (v2.2.13). <img width="1524" alt="Screenshot 2023-01-09 at 15 54 25" src="https://user-images.githubusercontent.com/25318659/211324192-4547d8ba-7340-4676-90e8-e5a440d440d0.png"> #### Actual behavior: The search page is not rendered (v2.2.14). <img width="1524" alt="Screenshot 2023-01-09 at 15 53 06" src="https://user-images.githubusercontent.com/25318659/211324254-04c3cd0f-4f88-433b-9d88-0a170f4729d7.png"> The following errors are thrown in the console in dev mode. <img width="1215" alt="Screenshot 2023-01-09 at 12 12 05" src="https://user-images.githubusercontent.com/25318659/211324352-2fcf53ce-2495-402e-b467-c2f85555fa0b.png"> If you would like immediate help on this, please email support@sourcegraph.com (you can still create the issue, but there are no [SLAs](https://about.sourcegraph.com/support/) on issues like there are for support requests). /cc @felixfbecker
non_process
sourcegraph for vscode extension doesn t render search page sourcegraph version vsce platform information vscode steps to reproduce install the latest vsce extension for the time being open it by clicking the sourcegraph icon in the sidebar check out the search page expected behavior the search page is rendered img width alt screenshot at src actual behavior the search page is not rendered img width alt screenshot at src the following errors are thrown in the console in dev mode img width alt screenshot at src if you would like immediate help on this please email support sourcegraph com you can still create the issue but there are no on issues like there are for support requests cc felixfbecker
0
5,120
7,890,513,098
IssuesEvent
2018-06-28 09:03:06
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Memory usage (external) reports a negative value?
memory process
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: v10.5.0 Platform: Linux PZW2 4.9.35+ #1014 Fri Jun 30 14:34:49 BST 2017 armv6l GNU/Linux Subsystem: **?d** If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: v10.5.0 * **Platform**: Linux PZW2 4.9.35+ #1014 Fri Jun 30 14:34:49 BST 2017 armv6l GNU/Linux * **Subsystem**: REPL in Node.js <!-- Enter your issue details below this comment. --> Running some http server process in node.js in REPL environment (under tmux), the function: process.memoryUsage() function can return a negative value. Is that a bug? The process (server) was run for many days, and I sometimes use process.memoryUsage() to check the values and I found: ``` > process.memoryUsage() { rss: 33452032, heapTotal: 9031680, heapUsed: 5112064, external: -503806 } ``` > The external memory is negative, but it was still postive 2 days ago when I checked. Notes, I run the node.js with flags: ` --expose-gc --max-old-space-size=256` This is running on a Raspberry Pi ZERO W. ****
1.0
Memory usage (external) reports a negative value? - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: v10.5.0 Platform: Linux PZW2 4.9.35+ #1014 Fri Jun 30 14:34:49 BST 2017 armv6l GNU/Linux Subsystem: **?d** If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: v10.5.0 * **Platform**: Linux PZW2 4.9.35+ #1014 Fri Jun 30 14:34:49 BST 2017 armv6l GNU/Linux * **Subsystem**: REPL in Node.js <!-- Enter your issue details below this comment. --> Running some http server process in node.js in REPL environment (under tmux), the function: process.memoryUsage() function can return a negative value. Is that a bug? The process (server) was run for many days, and I sometimes use process.memoryUsage() to check the values and I found: ``` > process.memoryUsage() { rss: 33452032, heapTotal: 9031680, heapUsed: 5112064, external: -503806 } ``` > The external memory is negative, but it was still postive 2 days ago when I checked. Notes, I run the node.js with flags: ` --expose-gc --max-old-space-size=256` This is running on a Raspberry Pi ZERO W. ****
process
memory usage external reports a negative value thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version platform linux fri jun bst gnu linux subsystem d if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version platform linux fri jun bst gnu linux subsystem repl in node js running some http server process in node js in repl environment under tmux the function process memoryusage function can return a negative value is that a bug the process server was run for many days and i sometimes use process memoryusage to check the values and i found process memoryusage rss heaptotal heapused external the external memory is negative but it was still postive days ago when i checked notes i run the node js with flags expose gc max old space size this is running on a raspberry pi zero w
1
13,477
15,987,347,271
IssuesEvent
2021-04-19 00:13:32
shirou/gopsutil
https://api.github.com/repos/shirou/gopsutil
closed
Get Windows Processes Name Take Too Much Time
os:windows package:process
**Describe the bug** It takes too much time to get the windows process name. In general, my windows has 260 processes, and When I get the names of these processes, it usually takes 8-9 seconds. It seems execute func getFromSnapProcess(pid int32) (int32, int32, string, error) 260*260 times, not 260 times. **To Reproduce** ```go processInfos, _ := process.Processes() for _, processInfo := range processInfos { name, _ := processInfo.Name() } ``` **Expected behavior** It should execute func getFromSnapProcess(pid int32) (int32, int32, string, error) 260 times and takse less than 1s **Environment (please complete the following information):** - [ ] Windows: Microsoft Windows [version 10.0.18363.1440] **Additional context** command: go build C:\WINDOWS\system32>go env set GO111MODULE=on set GOARCH=amd64 set GOBIN= set GOCACHE=C:\Users\xiaotao.hu\AppData\Local\go-build set GOENV=C:\Users\xiaotao.hu\AppData\Roaming\go\env set GOEXE=.exe set GOFLAGS= set GOHOSTARCH=amd64 set GOHOSTOS=windows set GOINSECURE= set GOMODCACHE=C:\Users\xiaotao.hu\go\pkg\mod set GONOPROXY= set GONOSUMDB= set GOOS=windows set GOPATH=C:\Users\xiaotao.hu\go set GOPRIVATE= set GOPROXY=https://goproxy.io set GOROOT=c:\go set GOSUMDB=sum.golang.org set GOTMPDIR= set GOTOOLDIR=c:\go\pkg\tool\windows_amd64 set GCCGO=gccgo set AR=ar set CC=x86_64-w64-mingw32-gcc set CXX=g++ set CGO_ENABLED=1 set GOMOD=NUL set CGO_CFLAGS=-g -O2 set CGO_CPPFLAGS= set CGO_CXXFLAGS=-g -O2 set CGO_FFLAGS=-g -O2 set CGO_LDFLAGS=-g -O2 set PKG_CONFIG=pkg-config set GOGCCFLAGS=-m64 -mthreads -fmessage-length=0 -fdebug-prefix-map=C:\Users\xiaotao.hu\AppData\Local\Temp\go-build909459346=/tmp/go-build -gno-record-gcc-switches
1.0
Get Windows Processes Name Take Too Much Time - **Describe the bug** It takes too much time to get the windows process name. In general, my windows has 260 processes, and When I get the names of these processes, it usually takes 8-9 seconds. It seems execute func getFromSnapProcess(pid int32) (int32, int32, string, error) 260*260 times, not 260 times. **To Reproduce** ```go processInfos, _ := process.Processes() for _, processInfo := range processInfos { name, _ := processInfo.Name() } ``` **Expected behavior** It should execute func getFromSnapProcess(pid int32) (int32, int32, string, error) 260 times and takse less than 1s **Environment (please complete the following information):** - [ ] Windows: Microsoft Windows [version 10.0.18363.1440] **Additional context** command: go build C:\WINDOWS\system32>go env set GO111MODULE=on set GOARCH=amd64 set GOBIN= set GOCACHE=C:\Users\xiaotao.hu\AppData\Local\go-build set GOENV=C:\Users\xiaotao.hu\AppData\Roaming\go\env set GOEXE=.exe set GOFLAGS= set GOHOSTARCH=amd64 set GOHOSTOS=windows set GOINSECURE= set GOMODCACHE=C:\Users\xiaotao.hu\go\pkg\mod set GONOPROXY= set GONOSUMDB= set GOOS=windows set GOPATH=C:\Users\xiaotao.hu\go set GOPRIVATE= set GOPROXY=https://goproxy.io set GOROOT=c:\go set GOSUMDB=sum.golang.org set GOTMPDIR= set GOTOOLDIR=c:\go\pkg\tool\windows_amd64 set GCCGO=gccgo set AR=ar set CC=x86_64-w64-mingw32-gcc set CXX=g++ set CGO_ENABLED=1 set GOMOD=NUL set CGO_CFLAGS=-g -O2 set CGO_CPPFLAGS= set CGO_CXXFLAGS=-g -O2 set CGO_FFLAGS=-g -O2 set CGO_LDFLAGS=-g -O2 set PKG_CONFIG=pkg-config set GOGCCFLAGS=-m64 -mthreads -fmessage-length=0 -fdebug-prefix-map=C:\Users\xiaotao.hu\AppData\Local\Temp\go-build909459346=/tmp/go-build -gno-record-gcc-switches
process
get windows processes name take too much time describe the bug it takes too much time to get the windows process name in general my windows has processes and when i get the names of these processes it usually takes seconds it seems execute func getfromsnapprocess pid string error times not times to reproduce go processinfos process processes for processinfo range processinfos name processinfo name expected behavior it should execute func getfromsnapprocess pid string error times and takse less than environment please complete the following information windows microsoft windows additional context command go build c windows go env set on set goarch set gobin set gocache c users xiaotao hu appdata local go build set goenv c users xiaotao hu appdata roaming go env set goexe exe set goflags set gohostarch set gohostos windows set goinsecure set gomodcache c users xiaotao hu go pkg mod set gonoproxy set gonosumdb set goos windows set gopath c users xiaotao hu go set goprivate set goproxy set goroot c go set gosumdb sum golang org set gotmpdir set gotooldir c go pkg tool windows set gccgo gccgo set ar ar set cc gcc set cxx g set cgo enabled set gomod nul set cgo cflags g set cgo cppflags set cgo cxxflags g set cgo fflags g set cgo ldflags g set pkg config pkg config set gogccflags mthreads fmessage length fdebug prefix map c users xiaotao hu appdata local temp go tmp go build gno record gcc switches
1
88,572
17,610,951,097
IssuesEvent
2021-08-18 01:05:50
jhona-tam/Farmacia
https://api.github.com/repos/jhona-tam/Farmacia
closed
Usando libreria
documentation code
Usamos librería `animate.css` para una mejor presentacion en los movimientos.
1.0
Usando libreria - Usamos librería `animate.css` para una mejor presentacion en los movimientos.
non_process
usando libreria usamos librería animate css para una mejor presentacion en los movimientos
0
274,696
8,564,351,057
IssuesEvent
2018-11-09 16:30:01
mozilla/MozDef
https://api.github.com/repos/mozilla/MozDef
closed
Feature flagging
category:enhancement priority:medium
#### ISSUE TYPE - Feature Idea #### COMPONENT NAME Meteor/Mozdef #### CONFIGURATION Will add capability to do feature flagging to do things like: - turn off block IP (already exists) - turn off the ip blocklist - turn off the fqdn blocklist - turn off individual vizualizations #### DESCRIPTION Currently it's an all or nothing experience, but maybe folks just want to use Mozdef to track investigations/incidents, or maybe they just want alerts, or just visualizations. This will allow folks to turn on and off the bits they want. This will start with the UI first, then expand to other components if possible.
1.0
Feature flagging - #### ISSUE TYPE - Feature Idea #### COMPONENT NAME Meteor/Mozdef #### CONFIGURATION Will add capability to do feature flagging to do things like: - turn off block IP (already exists) - turn off the ip blocklist - turn off the fqdn blocklist - turn off individual vizualizations #### DESCRIPTION Currently it's an all or nothing experience, but maybe folks just want to use Mozdef to track investigations/incidents, or maybe they just want alerts, or just visualizations. This will allow folks to turn on and off the bits they want. This will start with the UI first, then expand to other components if possible.
non_process
feature flagging issue type feature idea component name meteor mozdef configuration will add capability to do feature flagging to do things like turn off block ip already exists turn off the ip blocklist turn off the fqdn blocklist turn off individual vizualizations description currently it s an all or nothing experience but maybe folks just want to use mozdef to track investigations incidents or maybe they just want alerts or just visualizations this will allow folks to turn on and off the bits they want this will start with the ui first then expand to other components if possible
0
59,958
14,679,657,881
IssuesEvent
2020-12-31 07:38:04
google/xls
https://api.github.com/repos/google/xls
closed
Ubuntu 18.04: undeclared inclusion(s) in rule '@com_google_absl//absl/base:log_severity'
build
Here's the error: ``` ❯ bazel build //xls/tools:codegen_main INFO: Analyzed target //xls/tools:codegen_main (0 packages loaded, 0 targets configured). INFO: Found 1 target... ERROR: /home/lachlan/.cache/bazel/_bazel_lachlan/d8365cac006c3b38c2028f7288f6928e/external/com_google_absl/absl/base/BUILD.bazel:54:11: undeclared inclusion(s) in rule '@com_google_absl//absl/base:log_severity': this rule is missing dependency declarations for the following files included by 'com_google_absl/absl/base/log_severity.cc': '/usr/lib/gcc/x86_64-linux-gnu/9/include/stddef.h' '/usr/lib/gcc/x86_64-linux-gnu/9/include/stdarg.h' '/usr/lib/gcc/x86_64-linux-gnu/9/include/stdint.h' '/usr/lib/gcc/x86_64-linux-gnu/9/include/limits.h' '/usr/lib/gcc/x86_64-linux-gnu/9/include/syslimits.h' Target //xls/tools:codegen_main failed to build Use --verbose_failures to see the command lines of failed build steps. INFO: Elapsed time: 3.369s, Critical Path: 3.01s INFO: 13 processes: 13 internal. FAILED: Build did NOT complete successfully ``` I'm using ubuntu running within wsl.
1.0
Ubuntu 18.04: undeclared inclusion(s) in rule '@com_google_absl//absl/base:log_severity' - Here's the error: ``` ❯ bazel build //xls/tools:codegen_main INFO: Analyzed target //xls/tools:codegen_main (0 packages loaded, 0 targets configured). INFO: Found 1 target... ERROR: /home/lachlan/.cache/bazel/_bazel_lachlan/d8365cac006c3b38c2028f7288f6928e/external/com_google_absl/absl/base/BUILD.bazel:54:11: undeclared inclusion(s) in rule '@com_google_absl//absl/base:log_severity': this rule is missing dependency declarations for the following files included by 'com_google_absl/absl/base/log_severity.cc': '/usr/lib/gcc/x86_64-linux-gnu/9/include/stddef.h' '/usr/lib/gcc/x86_64-linux-gnu/9/include/stdarg.h' '/usr/lib/gcc/x86_64-linux-gnu/9/include/stdint.h' '/usr/lib/gcc/x86_64-linux-gnu/9/include/limits.h' '/usr/lib/gcc/x86_64-linux-gnu/9/include/syslimits.h' Target //xls/tools:codegen_main failed to build Use --verbose_failures to see the command lines of failed build steps. INFO: Elapsed time: 3.369s, Critical Path: 3.01s INFO: 13 processes: 13 internal. FAILED: Build did NOT complete successfully ``` I'm using ubuntu running within wsl.
non_process
ubuntu undeclared inclusion s in rule com google absl absl base log severity here s the error ❯ bazel build xls tools codegen main info analyzed target xls tools codegen main packages loaded targets configured info found target error home lachlan cache bazel bazel lachlan external com google absl absl base build bazel undeclared inclusion s in rule com google absl absl base log severity this rule is missing dependency declarations for the following files included by com google absl absl base log severity cc usr lib gcc linux gnu include stddef h usr lib gcc linux gnu include stdarg h usr lib gcc linux gnu include stdint h usr lib gcc linux gnu include limits h usr lib gcc linux gnu include syslimits h target xls tools codegen main failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes internal failed build did not complete successfully i m using ubuntu running within wsl
0
22,589
31,814,482,806
IssuesEvent
2023-09-13 19:20:52
googleapis/sdk-platform-java
https://api.github.com/repos/googleapis/sdk-platform-java
closed
Dependency Dashboard
type: process priority: p4
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Rate-Limited These updates are currently rate-limited. Click on a checkbox below to force their creation now. - [x] <!-- unlimit-branch=renovate/org.jacoco-jacoco-maven-plugin-0.x -->build(deps): update dependency org.jacoco:jacoco-maven-plugin to v0.8.9 - [x] <!-- unlimit-branch=renovate/netty-dependencies -->deps: update netty dependencies to v4.1.91.final (`io.netty:netty-buffer`, `io.netty:netty-codec`, `io.netty:netty-codec-http`, `io.netty:netty-codec-http2`, `io.netty:netty-codec-socks`, `io.netty:netty-common`, `io.netty:netty-handler`, `io.netty:netty-handler-proxy`, `io.netty:netty-resolver`, `io.netty:netty-transport`, `io.netty:netty-transport-native-unix-common`) - [ ] <!-- unlimit-branch=renovate/attrs-22.x -->deps: update dependency attrs to v22.2.0 - [ ] <!-- unlimit-branch=renovate/charset-normalizer-2.x -->deps: update dependency charset-normalizer to v2.1.1 - [ ] <!-- unlimit-branch=renovate/click-8.x -->deps: update dependency click to v8.1.3 - [ ] <!-- unlimit-branch=renovate/gcp-releasetool-1.x -->deps: update dependency gcp-releasetool to v1.11.0 - [ ] <!-- unlimit-branch=renovate/google-api-core-2.x -->deps: update dependency google-api-core to v2.11.0 - [ ] <!-- unlimit-branch=renovate/google-auth-2.x -->deps: update dependency google-auth to v2.17.1 - [ ] <!-- unlimit-branch=renovate/google-cloud-storage-2.x -->deps: update dependency google-cloud-storage to v2.8.0 - [ ] <!-- unlimit-branch=renovate/google-crc32c-1.x -->deps: update dependency google-crc32c to v1.5.0 - [ ] <!-- unlimit-branch=renovate/google-resumable-media-2.x -->deps: update dependency google-resumable-media to v2.4.1 - [ ] <!-- unlimit-branch=renovate/googleapis-common-protos-1.x -->deps: update dependency googleapis-common-protos to v1.59.0 - [ ] <!-- unlimit-branch=renovate/importlib-metadata-4.x -->deps: update dependency importlib-metadata to v4.13.0 - [ ] <!-- unlimit-branch=renovate/jeepney-0.x -->deps: update dependency jeepney to v0.8.0 - [ ] <!-- unlimit-branch=renovate/jinja2-3.x -->deps: update dependency jinja2 to v3.1.2 - [ ] <!-- unlimit-branch=renovate/keyring-23.x -->deps: update dependency keyring to v23.13.1 - [ ] <!-- unlimit-branch=renovate/markupsafe-2.x -->deps: update dependency markupsafe to v2.1.2 - [ ] <!-- unlimit-branch=renovate/protobuf-3.x -->deps: update dependency protobuf to v3.20.3 - [ ] <!-- unlimit-branch=renovate/pyjwt-2.x -->deps: update dependency pyjwt to v2.6.0 - [ ] <!-- unlimit-branch=renovate/requests-2.x -->deps: update dependency requests to v2.28.2 - [ ] <!-- unlimit-branch=renovate/typing-extensions-4.x -->deps: update dependency typing-extensions to v4.5.0 - [ ] <!-- unlimit-branch=renovate/zipp-3.x -->deps: update dependency zipp to v3.15.0 - [ ] <!-- create-all-rate-limited-prs -->🔐 **Create all rate-limited PRs at once** 🔐 --- ### ⚠ Dependency Lookup Warnings ⚠ - Renovate failed to look up the following dependencies: `Failed to look up maven package com.google.cloud:gapic-showcase`, `Failed to look up maven package com.google.cloud:gapic-showcase-parent`, `Failed to look up maven package com.google.api.grpc:proto-gapic-showcase-v1beta1`, `Failed to look up maven package com.google.api.grpc:grpc-gapic-showcase-v1beta1`. Files affected: `coverage-report/pom.xml`, `showcase/gapic-showcase/pom.xml`, `showcase/grpc-gapic-showcase-v1beta1/pom.xml`, `showcase/pom.xml`, `showcase/proto-gapic-showcase-v1beta1/pom.xml` --- ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/cryptography-39.x -->[deps: update dependency cryptography to v39.0.2](../pull/1592) - [ ] <!-- rebase-branch=renovate/gcp-docuploader-0.x -->[deps: update dependency gcp-docuploader to v0.6.5](../pull/1593) - [ ] <!-- rebase-branch=renovate/google-cloud-core-2.x -->[deps: update dependency google-cloud-core to v2.3.2](../pull/1595) - [ ] <!-- rebase-branch=renovate/net.bytebuddy-byte-buddy-1.x -->[deps: update dependency net.bytebuddy:byte-buddy to v1.14.3](../pull/1567) - [ ] <!-- rebase-branch=renovate/org.threeten-threetenbp-1.x -->[deps: update dependency org.threeten:threetenbp to v1.6.8](../pull/1590) - [ ] <!-- rebase-branch=renovate/threeten.version -->[deps: update dependency org.threeten:threetenbp to v1.6.8](../pull/1591) - [ ] <!-- rebase-branch=renovate/urllib3-1.x -->[deps: update dependency urllib3 to v1.26.15](../pull/1596) - [ ] <!-- rebase-branch=renovate/google-api-dependencies -->[deps: update google api dependencies](../pull/1578) (`com.google.api.grpc:grpc-google-common-protos`, `com.google.api.grpc:grpc-google-iam-v1`, `com.google.api.grpc:proto-google-common-protos`, `com.google.api.grpc:proto-google-iam-v1`, `com.google.api:api-common`) - [ ] <!-- rebase-branch=renovate/protobuf-dependencies -->[deps: update protobuf dependencies to v3.22.2](../pull/1369) (`com.google.protobuf:protobuf-java`, `com.google.protobuf:protobuf-bom`) - [x] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once** ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/org.apache.maven.plugins-maven-surefire-plugin-3.x -->[build(deps): update dependency org.apache.maven.plugins:maven-surefire-plugin to v3.0.0](../pull/1488) - [ ] <!-- recreate-branch=renovate/com.coveo-fmt-maven-plugin-2.x -->[build(deps): update dependency com.coveo:fmt-maven-plugin to v2.13](../pull/1363) - [ ] <!-- recreate-branch=renovate/com.google.googlejavaformat-google-java-format-1.x -->[deps: update dependency com.google.googlejavaformat:google-java-format to v1.16.0](../pull/1422) ## Detected dependencies <details><summary>bazel</summary> <blockquote> <details><summary>WORKSPACE</summary> </details> </blockquote> </details> <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/ci.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/checkout v3` - `actions/setup-java v3` </details> <details><summary>.github/workflows/downstream-native-image.yaml</summary> - `actions/checkout v3` - `stCarolas/setup-maven v4` - `ayltai/setup-graalvm v1` </details> <details><summary>.github/workflows/nightly.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/setup-java v3` </details> <details><summary>.github/workflows/shared_dependencies.yaml</summary> - `actions/checkout v2` - `stCarolas/setup-maven v4` - `actions/setup-java v1` </details> <details><summary>.github/workflows/sonar.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` - `actions/cache v3` - `actions/cache v3` </details> </blockquote> </details> <details><summary>gradle-wrapper</summary> <blockquote> <details><summary>rules_java_gapic/resources/gradle/gradle/wrapper/gradle-wrapper.properties</summary> - `gradle 7.6.1` </details> </blockquote> </details> <details><summary>maven</summary> <blockquote> <details><summary>api-common-java/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.auto.value:auto-value-annotations 1.10.1` - `com.google.auto.value:auto-value 1.10.1` - `com.google.code.findbugs:jsr305 3.0.2` - `com.google.errorprone:error_prone_annotations 2.18.0` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` - `com.google.auto.value:auto-value 1.10.1` </details> <details><summary>coverage-report/pom.xml</summary> - `com.google.cloud:gapic-showcase 0.0.1-SNAPSHOT` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax-grpc 2.24.1-SNAPSHOT` - `com.google.api:gax-httpjson 0.109.1-SNAPSHOT` - `com.google.api:api-common 2.7.1-SNAPSHOT` - `org.jacoco:jacoco-maven-plugin 0.8.8` - `com.coveo:fmt-maven-plugin 2.9` </details> <details><summary>gapic-generator-java-bom/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api:api-common 2.7.1-SNAPSHOT` - `com.google.api:gax-bom 2.24.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v1 1.10.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v2 1.10.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v2beta 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v1 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v2 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v2beta 1.10.1-SNAPSHOT` </details> <details><summary>gapic-generator-java-pom-parent/pom.xml</summary> - `com.google.guava:guava 31.1-jre` - `javax.annotation:javax.annotation-api 1.3.2` - `com.google.auth:google-auth-library-bom 1.16.0` - `com.google.code.gson:gson 2.10.1` - `com.google.guava:guava-bom 31.1-jre` - `com.google.protobuf:protobuf-bom 3.21.12` - `io.grpc:grpc-bom 1.54.0` - `com.google.cloud:google-cloud-shared-config 1.5.5` - `org.apache.maven.plugins:maven-surefire-plugin 3.0.0-M8` - `org.apache.maven.plugins:maven-failsafe-plugin 3.0.0` - `org.jacoco:jacoco-maven-plugin 0.8.8` - `javax.annotation:javax.annotation-api 1.3.2` - `com.google.auth:google-auth-library-bom 1.16.0` - `com.google.guava:guava 31.1-jre` - `io.grpc:grpc-bom 1.54.0` - `com.google.code.gson:gson 2.10.1` - `com.google.protobuf:protobuf-bom 3.21.12` - `javax.annotation:javax.annotation-api 1.3.2` - `com.google.protobuf:protobuf-bom 3.21.12` - `io.grpc:grpc-bom 1.54.0` - `javax.annotation:javax.annotation-api 1.3.2` - `com.google.guava:guava-bom 31.1-jre` - `com.google.protobuf:protobuf-bom 3.21.12` - `io.grpc:grpc-bom 1.54.0` - `javax.annotation:javax.annotation-api 1.3.2` </details> <details><summary>gapic-generator-java/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` - `kr.motd.maven:os-maven-plugin 1.7.1` - `org.apache.maven.plugins:maven-compiler-plugin 3.11.0` - `com.google.auto.value:auto-value 1.10.1` - `com.coveo:fmt-maven-plugin 2.9.1` - `com.googlecode.maven-download-plugin:download-maven-plugin 1.6.8` - `org.xolstice.maven.plugins:protobuf-maven-plugin 0.6.1` - `org.apache.maven.plugins:maven-surefire-plugin 3.0.0` - `org.jacoco:jacoco-maven-plugin 0.8.8` - `org.apache.maven.plugins:maven-shade-plugin 3.4.1` - `com.google.guava:guava 31.1-jre` - `com.google.auto.value:auto-value-annotations 1.10.1` - `org.yaml:snakeyaml 2.0` - `com.google.googlejavaformat:google-java-format 1.7` - `junit:junit 4.13.2` - `com.google.errorprone:error_prone_annotations 2.18.0` - `com.google.truth:truth 1.1.3` - `io.github.java-diff-utils:java-diff-utils 4.12` </details> <details><summary>gax-java/gax-bom/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.5` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax-grpc 2.24.1-SNAPSHOT` - `com.google.api:gax-grpc 2.24.1-SNAPSHOT` - `com.google.api:gax-grpc 2.24.1-SNAPSHOT` - `com.google.api:gax-httpjson 0.109.1-SNAPSHOT` - `com.google.api:gax-httpjson 0.109.1-SNAPSHOT` - `com.google.api:gax-httpjson 0.109.1-SNAPSHOT` </details> <details><summary>gax-java/gax-grpc/pom.xml</summary> - `com.google.api:gax-parent 2.24.1-SNAPSHOT` </details> <details><summary>gax-java/gax-httpjson/pom.xml</summary> - `com.google.api:gax-parent 2.24.1-SNAPSHOT` </details> <details><summary>gax-java/gax/pom.xml</summary> - `com.google.api:gax-parent 2.24.1-SNAPSHOT` </details> <details><summary>gax-java/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api:api-common 2.7.1-SNAPSHOT` - `org.threeten:threetenbp 1.6.7` - `com.google.code.findbugs:jsr305 3.0.2` - `io.opencensus:opencensus-api 0.31.1` - `com.google.auto.value:auto-value-annotations 1.10.1` - `com.google.auto.value:auto-value 1.10.1` - `junit:junit 4.13.2` - `org.mockito:mockito-core 4.11.0` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT` - `org.graalvm.sdk:graal-sdk 22.3.1` - `com.google.http-client:google-http-client-bom 1.43.1` - `com.google.truth:truth 1.1.3` - `com.google.auto.value:auto-value 1.10.1` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` </details> <details><summary>java-common-protos/grpc-google-common-protos/pom.xml</summary> - `com.google.api.grpc:google-common-protos-parent 2.15.1-SNAPSHOT` </details> <details><summary>java-common-protos/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT` - `com.google.guava:guava 31.1-jre` - `junit:junit 4.13.2` - `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.2` - `org.apache.maven.plugins:maven-javadoc-plugin 3.5.0` </details> <details><summary>java-common-protos/proto-google-common-protos/pom.xml</summary> - `com.google.api.grpc:google-common-protos-parent 2.15.1-SNAPSHOT` </details> <details><summary>java-core/google-cloud-core-bom/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT` - `com.google.cloud:google-cloud-core-grpc 2.14.1-SNAPSHOT` - `com.google.cloud:google-cloud-core-http 2.14.1-SNAPSHOT` </details> <details><summary>java-core/google-cloud-core-grpc/pom.xml</summary> - `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT` </details> <details><summary>java-core/google-cloud-core-http/pom.xml</summary> - `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT` </details> <details><summary>java-core/google-cloud-core/pom.xml</summary> - `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT` </details> <details><summary>java-core/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.cloud:google-cloud-shared-dependencies 3.6.1-SNAPSHOT` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` - `org.easymock:easymock 5.1.0` - `org.mockito:mockito-all 1.10.19` </details> <details><summary>java-iam/grpc-google-iam-v1/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/grpc-google-iam-v2/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/grpc-google-iam-v2beta/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api:gax-bom 2.24.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v2 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v2 1.10.1-SNAPSHOT` - `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v2beta 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v1 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v2beta 1.10.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v1 1.10.1-SNAPSHOT` - `junit:junit 4.13.2` - `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.2` - `org.apache.maven.plugins:maven-javadoc-plugin 3.5.0` </details> <details><summary>java-iam/proto-google-iam-v1/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/proto-google-iam-v2/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/proto-google-iam-v2beta/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-shared-dependencies/dependency-convergence-check/pom.xml</summary> - `com.google.guava:guava 31.0.1-jre` - `com.google.cloud.tools:dependencies 1.5.13` </details> <details><summary>java-shared-dependencies/first-party-dependencies/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.5` - `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT` - `com.google.cloud:grpc-gcp 1.4.1` - `com.google.code.gson:gson 2.10.1` - `com.google.cloud:google-cloud-core-bom 2.14.1-SNAPSHOT` - `com.google.http-client:google-http-client-bom 1.43.1` - `com.google.oauth-client:google-oauth-client-bom 1.34.1` - `com.google.api-client:google-api-client-bom 2.2.0` - `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT` - `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT` </details> <details><summary>java-shared-dependencies/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.cloud:first-party-dependencies 3.6.1-SNAPSHOT` - `com.google.cloud:third-party-dependencies 3.6.1-SNAPSHOT` </details> <details><summary>java-shared-dependencies/third-party-dependencies/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.5` - `org.apache.httpcomponents:httpcore 4.4.16` - `org.apache.httpcomponents:httpclient 4.5.14` - `org.threeten:threetenbp 1.6.7` - `javax.annotation:javax.annotation-api 1.3.2` - `org.codehaus.mojo:animal-sniffer-annotations 1.23` - `com.google.code.findbugs:jsr305 3.0.2` - `com.google.errorprone:error_prone_annotations 2.18.0` - `com.fasterxml.jackson:jackson-bom 2.14.2` - `commons-codec:commons-codec 1.15` - `io.opencensus:opencensus-api 0.31.1` - `io.opencensus:opencensus-contrib-grpc-util 0.31.1` - `io.opencensus:opencensus-contrib-http-util 0.31.1` - `io.opencensus:opencensus-contrib-zpages 0.31.1` - `io.opencensus:opencensus-exporter-stats-stackdriver 0.31.1` - `io.opencensus:opencensus-exporter-trace-stackdriver 0.31.1` - `io.opencensus:opencensus-impl 0.31.1` - `io.opencensus:opencensus-impl-core 0.31.1` - `org.checkerframework:checker-qual 3.32.0` - `io.perfmark:perfmark-api 0.26.0` </details> <details><summary>java-shared-dependencies/upper-bound-check/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.5` - `com.google.cloud:google-cloud-shared-dependencies 3.6.1-SNAPSHOT` </details> <details><summary>pom.xml</summary> - `com.coveo:fmt-maven-plugin 2.9` - `org.apache.maven.plugins:maven-deploy-plugin 3.1.1` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` </details> <details><summary>showcase/gapic-showcase/pom.xml</summary> - `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` - `org.jacoco:jacoco-maven-plugin 0.8.8` - `com.coveo:fmt-maven-plugin 2.9` - `com.googlecode.maven-download-plugin:download-maven-plugin 1.6.8` - `com.google.truth:truth 1.1.3` </details> <details><summary>showcase/grpc-gapic-showcase-v1beta1/pom.xml</summary> - `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` </details> <details><summary>showcase/pom.xml</summary> - `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT` - `com.google.api.grpc:proto-gapic-showcase-v1beta1 0.0.1-SNAPSHOT` - `com.google.api.grpc:grpc-gapic-showcase-v1beta1 0.0.1-SNAPSHOT` - `com.google.cloud:gapic-showcase 0.0.1-SNAPSHOT` - `junit:junit 4.13.2` - `org.apache.maven.plugins:maven-surefire-plugin 3.0.0-M8` </details> <details><summary>showcase/proto-gapic-showcase-v1beta1/pom.xml</summary> - `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>.kokoro/requirements.txt</summary> - `attrs ==22.1.0` - `cachetools ==4.2.4` - `certifi ==2022.12.7` - `cffi ==1.15.1` - `charset-normalizer ==2.0.12` - `click ==8.0.4` - `colorlog ==6.7.0` - `cryptography ==39.0.1` - `gcp-docuploader ==0.6.3` - `gcp-releasetool ==1.10.5` - `google-api-core ==2.8.2` - `google-auth ==2.14.1` - `google-cloud-core ==2.3.1` - `google-cloud-storage ==2.0.0` - `google-crc32c ==1.3.0` - `google-resumable-media ==2.3.3` - `googleapis-common-protos ==1.56.3` - `idna ==3.4` - `importlib-metadata ==4.8.3` - `jeepney ==0.7.1` - `jinja2 ==3.0.3` - `keyring ==23.4.1` - `markupsafe ==2.0.1` - `packaging ==21.3` - `protobuf ==3.19.5` - `pyasn1 ==0.4.8` - `pyasn1-modules ==0.2.8` - `pycparser ==2.21` - `pyjwt ==2.4.0` - `pyparsing ==3.0.9` - `pyperclip ==1.8.2` - `python-dateutil ==2.8.2` - `requests ==2.27.1` - `rsa ==4.9` - `secretstorage ==3.3.3` - `six ==1.16.0` - `typing-extensions ==4.1.1` - `urllib3 ==1.26.12` - `zipp ==3.6.0` </details> </blockquote> </details> <details><summary>regex</summary> <blockquote> <details><summary>gax-java/dependencies.properties</summary> - `com.google.api.grpc:proto-google-common-protos 2.14.3` - `com.google.api.grpc:grpc-google-common-protos 2.14.3` - `com.google.auth:google-auth-library-oauth2-http 1.16.0` - `com.google.auth:google-auth-library-credentials 1.16.0` - `io.opencensus:opencensus-api 0.31.1` - `io.opencensus:opencensus-contrib-grpc-metrics 0.31.1` - `io.opencensus:opencensus-contrib-http-util 0.31.1` - `io.netty:netty-tcnative-boringssl-static 2.0.59.Final` - `io.netty:netty-handler 4.1.90.Final` - `io.netty:netty-common 4.1.90.Final` - `io.netty:netty-codec-socks 4.1.90.Final` - `io.netty:netty-codec-http2 4.1.90.Final` - `io.netty:netty-codec-http 4.1.90.Final` - `io.netty:netty-codec 4.1.90.Final` - `io.netty:netty-buffer 4.1.90.Final` - `io.netty:netty-resolver 4.1.90.Final` - `io.netty:netty-transport 4.1.90.Final` - `io.netty:netty-handler-proxy 4.1.90.Final` - `io.netty:netty-transport-native-unix-common 4.1.90.Final` - `io.perfmark:perfmark-api 0.26.0` - `org.apache.tomcat:annotations-api 6.0.53` - `com.google.code.gson:gson 2.10.1` - `com.google.guava:guava 31.1-jre` - `com.google.guava:failureaccess 1.0.1` - `org.apache.commons:commons-lang3 3.12.0` - `com.google.android:annotations 4.1.1.4` - `com.google.code.findbugs:jsr305 3.0.2` - `com.google.errorprone:error_prone_annotations 2.18.0` - `com.google.j2objc:j2objc-annotations 1.3` - `com.google.auto.value:auto-value 1.10.1` - `com.google.auto.value:auto-value-annotations 1.10.1` - `com.google.api:api-common 2.6.3` - `org.threeten:threetenbp 1.6.7` - `com.google.api.grpc:grpc-google-iam-v1 1.9.3` - `com.google.api.grpc:proto-google-iam-v1 1.9.3` - `com.google.http-client:google-http-client 1.43.1` - `com.google.http-client:google-http-client-gson 1.43.1` - `org.codehaus.mojo:animal-sniffer-annotations 1.23` - `javax.annotation:javax.annotation-api 1.3.2` - `org.graalvm.sdk:graal-sdk 22.3.1` - `junit:junit 4.13.2` - `org.mockito:mockito-core 2.28.2` - `org.hamcrest:hamcrest-core 1.3` - `com.google.truth:truth 1.1.3` - `com.googlecode.java-diff-utils:diffutils 1.3.0` - `net.bytebuddy:byte-buddy 1.14.2` - `org.objenesis:objenesis 2.6` </details> <details><summary>gax-java/dependencies.properties</summary> - `com.google.protobuf:protobuf-java 3.21.10` </details> <details><summary>gax-java/dependencies.properties</summary> - `com.google.googlejavaformat:google-java-format 1.15.0` </details> <details><summary>gax-java/dependencies.properties</summary> - `io.grpc:grpc-core 1.54.0` </details> </blockquote> </details> --- - [x] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Rate-Limited These updates are currently rate-limited. Click on a checkbox below to force their creation now. - [x] <!-- unlimit-branch=renovate/org.jacoco-jacoco-maven-plugin-0.x -->build(deps): update dependency org.jacoco:jacoco-maven-plugin to v0.8.9 - [x] <!-- unlimit-branch=renovate/netty-dependencies -->deps: update netty dependencies to v4.1.91.final (`io.netty:netty-buffer`, `io.netty:netty-codec`, `io.netty:netty-codec-http`, `io.netty:netty-codec-http2`, `io.netty:netty-codec-socks`, `io.netty:netty-common`, `io.netty:netty-handler`, `io.netty:netty-handler-proxy`, `io.netty:netty-resolver`, `io.netty:netty-transport`, `io.netty:netty-transport-native-unix-common`) - [ ] <!-- unlimit-branch=renovate/attrs-22.x -->deps: update dependency attrs to v22.2.0 - [ ] <!-- unlimit-branch=renovate/charset-normalizer-2.x -->deps: update dependency charset-normalizer to v2.1.1 - [ ] <!-- unlimit-branch=renovate/click-8.x -->deps: update dependency click to v8.1.3 - [ ] <!-- unlimit-branch=renovate/gcp-releasetool-1.x -->deps: update dependency gcp-releasetool to v1.11.0 - [ ] <!-- unlimit-branch=renovate/google-api-core-2.x -->deps: update dependency google-api-core to v2.11.0 - [ ] <!-- unlimit-branch=renovate/google-auth-2.x -->deps: update dependency google-auth to v2.17.1 - [ ] <!-- unlimit-branch=renovate/google-cloud-storage-2.x -->deps: update dependency google-cloud-storage to v2.8.0 - [ ] <!-- unlimit-branch=renovate/google-crc32c-1.x -->deps: update dependency google-crc32c to v1.5.0 - [ ] <!-- unlimit-branch=renovate/google-resumable-media-2.x -->deps: update dependency google-resumable-media to v2.4.1 - [ ] <!-- unlimit-branch=renovate/googleapis-common-protos-1.x -->deps: update dependency googleapis-common-protos to v1.59.0 - [ ] <!-- unlimit-branch=renovate/importlib-metadata-4.x -->deps: update dependency importlib-metadata to v4.13.0 - [ ] <!-- unlimit-branch=renovate/jeepney-0.x -->deps: update dependency jeepney to v0.8.0 - [ ] <!-- unlimit-branch=renovate/jinja2-3.x -->deps: update dependency jinja2 to v3.1.2 - [ ] <!-- unlimit-branch=renovate/keyring-23.x -->deps: update dependency keyring to v23.13.1 - [ ] <!-- unlimit-branch=renovate/markupsafe-2.x -->deps: update dependency markupsafe to v2.1.2 - [ ] <!-- unlimit-branch=renovate/protobuf-3.x -->deps: update dependency protobuf to v3.20.3 - [ ] <!-- unlimit-branch=renovate/pyjwt-2.x -->deps: update dependency pyjwt to v2.6.0 - [ ] <!-- unlimit-branch=renovate/requests-2.x -->deps: update dependency requests to v2.28.2 - [ ] <!-- unlimit-branch=renovate/typing-extensions-4.x -->deps: update dependency typing-extensions to v4.5.0 - [ ] <!-- unlimit-branch=renovate/zipp-3.x -->deps: update dependency zipp to v3.15.0 - [ ] <!-- create-all-rate-limited-prs -->🔐 **Create all rate-limited PRs at once** 🔐 --- ### ⚠ Dependency Lookup Warnings ⚠ - Renovate failed to look up the following dependencies: `Failed to look up maven package com.google.cloud:gapic-showcase`, `Failed to look up maven package com.google.cloud:gapic-showcase-parent`, `Failed to look up maven package com.google.api.grpc:proto-gapic-showcase-v1beta1`, `Failed to look up maven package com.google.api.grpc:grpc-gapic-showcase-v1beta1`. Files affected: `coverage-report/pom.xml`, `showcase/gapic-showcase/pom.xml`, `showcase/grpc-gapic-showcase-v1beta1/pom.xml`, `showcase/pom.xml`, `showcase/proto-gapic-showcase-v1beta1/pom.xml` --- ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/cryptography-39.x -->[deps: update dependency cryptography to v39.0.2](../pull/1592) - [ ] <!-- rebase-branch=renovate/gcp-docuploader-0.x -->[deps: update dependency gcp-docuploader to v0.6.5](../pull/1593) - [ ] <!-- rebase-branch=renovate/google-cloud-core-2.x -->[deps: update dependency google-cloud-core to v2.3.2](../pull/1595) - [ ] <!-- rebase-branch=renovate/net.bytebuddy-byte-buddy-1.x -->[deps: update dependency net.bytebuddy:byte-buddy to v1.14.3](../pull/1567) - [ ] <!-- rebase-branch=renovate/org.threeten-threetenbp-1.x -->[deps: update dependency org.threeten:threetenbp to v1.6.8](../pull/1590) - [ ] <!-- rebase-branch=renovate/threeten.version -->[deps: update dependency org.threeten:threetenbp to v1.6.8](../pull/1591) - [ ] <!-- rebase-branch=renovate/urllib3-1.x -->[deps: update dependency urllib3 to v1.26.15](../pull/1596) - [ ] <!-- rebase-branch=renovate/google-api-dependencies -->[deps: update google api dependencies](../pull/1578) (`com.google.api.grpc:grpc-google-common-protos`, `com.google.api.grpc:grpc-google-iam-v1`, `com.google.api.grpc:proto-google-common-protos`, `com.google.api.grpc:proto-google-iam-v1`, `com.google.api:api-common`) - [ ] <!-- rebase-branch=renovate/protobuf-dependencies -->[deps: update protobuf dependencies to v3.22.2](../pull/1369) (`com.google.protobuf:protobuf-java`, `com.google.protobuf:protobuf-bom`) - [x] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once** ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/org.apache.maven.plugins-maven-surefire-plugin-3.x -->[build(deps): update dependency org.apache.maven.plugins:maven-surefire-plugin to v3.0.0](../pull/1488) - [ ] <!-- recreate-branch=renovate/com.coveo-fmt-maven-plugin-2.x -->[build(deps): update dependency com.coveo:fmt-maven-plugin to v2.13](../pull/1363) - [ ] <!-- recreate-branch=renovate/com.google.googlejavaformat-google-java-format-1.x -->[deps: update dependency com.google.googlejavaformat:google-java-format to v1.16.0](../pull/1422) ## Detected dependencies <details><summary>bazel</summary> <blockquote> <details><summary>WORKSPACE</summary> </details> </blockquote> </details> <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/ci.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/checkout v3` - `actions/setup-java v3` </details> <details><summary>.github/workflows/downstream-native-image.yaml</summary> - `actions/checkout v3` - `stCarolas/setup-maven v4` - `ayltai/setup-graalvm v1` </details> <details><summary>.github/workflows/nightly.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/setup-java v3` </details> <details><summary>.github/workflows/shared_dependencies.yaml</summary> - `actions/checkout v2` - `stCarolas/setup-maven v4` - `actions/setup-java v1` </details> <details><summary>.github/workflows/sonar.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` - `actions/cache v3` - `actions/cache v3` </details> </blockquote> </details> <details><summary>gradle-wrapper</summary> <blockquote> <details><summary>rules_java_gapic/resources/gradle/gradle/wrapper/gradle-wrapper.properties</summary> - `gradle 7.6.1` </details> </blockquote> </details> <details><summary>maven</summary> <blockquote> <details><summary>api-common-java/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.auto.value:auto-value-annotations 1.10.1` - `com.google.auto.value:auto-value 1.10.1` - `com.google.code.findbugs:jsr305 3.0.2` - `com.google.errorprone:error_prone_annotations 2.18.0` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` - `com.google.auto.value:auto-value 1.10.1` </details> <details><summary>coverage-report/pom.xml</summary> - `com.google.cloud:gapic-showcase 0.0.1-SNAPSHOT` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax-grpc 2.24.1-SNAPSHOT` - `com.google.api:gax-httpjson 0.109.1-SNAPSHOT` - `com.google.api:api-common 2.7.1-SNAPSHOT` - `org.jacoco:jacoco-maven-plugin 0.8.8` - `com.coveo:fmt-maven-plugin 2.9` </details> <details><summary>gapic-generator-java-bom/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api:api-common 2.7.1-SNAPSHOT` - `com.google.api:gax-bom 2.24.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v1 1.10.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v2 1.10.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v2beta 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v1 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v2 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v2beta 1.10.1-SNAPSHOT` </details> <details><summary>gapic-generator-java-pom-parent/pom.xml</summary> - `com.google.guava:guava 31.1-jre` - `javax.annotation:javax.annotation-api 1.3.2` - `com.google.auth:google-auth-library-bom 1.16.0` - `com.google.code.gson:gson 2.10.1` - `com.google.guava:guava-bom 31.1-jre` - `com.google.protobuf:protobuf-bom 3.21.12` - `io.grpc:grpc-bom 1.54.0` - `com.google.cloud:google-cloud-shared-config 1.5.5` - `org.apache.maven.plugins:maven-surefire-plugin 3.0.0-M8` - `org.apache.maven.plugins:maven-failsafe-plugin 3.0.0` - `org.jacoco:jacoco-maven-plugin 0.8.8` - `javax.annotation:javax.annotation-api 1.3.2` - `com.google.auth:google-auth-library-bom 1.16.0` - `com.google.guava:guava 31.1-jre` - `io.grpc:grpc-bom 1.54.0` - `com.google.code.gson:gson 2.10.1` - `com.google.protobuf:protobuf-bom 3.21.12` - `javax.annotation:javax.annotation-api 1.3.2` - `com.google.protobuf:protobuf-bom 3.21.12` - `io.grpc:grpc-bom 1.54.0` - `javax.annotation:javax.annotation-api 1.3.2` - `com.google.guava:guava-bom 31.1-jre` - `com.google.protobuf:protobuf-bom 3.21.12` - `io.grpc:grpc-bom 1.54.0` - `javax.annotation:javax.annotation-api 1.3.2` </details> <details><summary>gapic-generator-java/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` - `kr.motd.maven:os-maven-plugin 1.7.1` - `org.apache.maven.plugins:maven-compiler-plugin 3.11.0` - `com.google.auto.value:auto-value 1.10.1` - `com.coveo:fmt-maven-plugin 2.9.1` - `com.googlecode.maven-download-plugin:download-maven-plugin 1.6.8` - `org.xolstice.maven.plugins:protobuf-maven-plugin 0.6.1` - `org.apache.maven.plugins:maven-surefire-plugin 3.0.0` - `org.jacoco:jacoco-maven-plugin 0.8.8` - `org.apache.maven.plugins:maven-shade-plugin 3.4.1` - `com.google.guava:guava 31.1-jre` - `com.google.auto.value:auto-value-annotations 1.10.1` - `org.yaml:snakeyaml 2.0` - `com.google.googlejavaformat:google-java-format 1.7` - `junit:junit 4.13.2` - `com.google.errorprone:error_prone_annotations 2.18.0` - `com.google.truth:truth 1.1.3` - `io.github.java-diff-utils:java-diff-utils 4.12` </details> <details><summary>gax-java/gax-bom/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.5` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax-grpc 2.24.1-SNAPSHOT` - `com.google.api:gax-grpc 2.24.1-SNAPSHOT` - `com.google.api:gax-grpc 2.24.1-SNAPSHOT` - `com.google.api:gax-httpjson 0.109.1-SNAPSHOT` - `com.google.api:gax-httpjson 0.109.1-SNAPSHOT` - `com.google.api:gax-httpjson 0.109.1-SNAPSHOT` </details> <details><summary>gax-java/gax-grpc/pom.xml</summary> - `com.google.api:gax-parent 2.24.1-SNAPSHOT` </details> <details><summary>gax-java/gax-httpjson/pom.xml</summary> - `com.google.api:gax-parent 2.24.1-SNAPSHOT` </details> <details><summary>gax-java/gax/pom.xml</summary> - `com.google.api:gax-parent 2.24.1-SNAPSHOT` </details> <details><summary>gax-java/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api:api-common 2.7.1-SNAPSHOT` - `org.threeten:threetenbp 1.6.7` - `com.google.code.findbugs:jsr305 3.0.2` - `io.opencensus:opencensus-api 0.31.1` - `com.google.auto.value:auto-value-annotations 1.10.1` - `com.google.auto.value:auto-value 1.10.1` - `junit:junit 4.13.2` - `org.mockito:mockito-core 4.11.0` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api:gax 2.24.1-SNAPSHOT` - `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT` - `org.graalvm.sdk:graal-sdk 22.3.1` - `com.google.http-client:google-http-client-bom 1.43.1` - `com.google.truth:truth 1.1.3` - `com.google.auto.value:auto-value 1.10.1` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` </details> <details><summary>java-common-protos/grpc-google-common-protos/pom.xml</summary> - `com.google.api.grpc:google-common-protos-parent 2.15.1-SNAPSHOT` </details> <details><summary>java-common-protos/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT` - `com.google.guava:guava 31.1-jre` - `junit:junit 4.13.2` - `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.2` - `org.apache.maven.plugins:maven-javadoc-plugin 3.5.0` </details> <details><summary>java-common-protos/proto-google-common-protos/pom.xml</summary> - `com.google.api.grpc:google-common-protos-parent 2.15.1-SNAPSHOT` </details> <details><summary>java-core/google-cloud-core-bom/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT` - `com.google.cloud:google-cloud-core-grpc 2.14.1-SNAPSHOT` - `com.google.cloud:google-cloud-core-http 2.14.1-SNAPSHOT` </details> <details><summary>java-core/google-cloud-core-grpc/pom.xml</summary> - `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT` </details> <details><summary>java-core/google-cloud-core-http/pom.xml</summary> - `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT` </details> <details><summary>java-core/google-cloud-core/pom.xml</summary> - `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT` </details> <details><summary>java-core/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.cloud:google-cloud-shared-dependencies 3.6.1-SNAPSHOT` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` - `org.easymock:easymock 5.1.0` - `org.mockito:mockito-all 1.10.19` </details> <details><summary>java-iam/grpc-google-iam-v1/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/grpc-google-iam-v2/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/grpc-google-iam-v2beta/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.api:gax-bom 2.24.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v2 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v2 1.10.1-SNAPSHOT` - `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v2beta 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v1 1.10.1-SNAPSHOT` - `com.google.api.grpc:grpc-google-iam-v2beta 1.10.1-SNAPSHOT` - `com.google.api.grpc:proto-google-iam-v1 1.10.1-SNAPSHOT` - `junit:junit 4.13.2` - `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.2` - `org.apache.maven.plugins:maven-javadoc-plugin 3.5.0` </details> <details><summary>java-iam/proto-google-iam-v1/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/proto-google-iam-v2/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-iam/proto-google-iam-v2beta/pom.xml</summary> - `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT` </details> <details><summary>java-shared-dependencies/dependency-convergence-check/pom.xml</summary> - `com.google.guava:guava 31.0.1-jre` - `com.google.cloud.tools:dependencies 1.5.13` </details> <details><summary>java-shared-dependencies/first-party-dependencies/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.5` - `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT` - `com.google.cloud:grpc-gcp 1.4.1` - `com.google.code.gson:gson 2.10.1` - `com.google.cloud:google-cloud-core-bom 2.14.1-SNAPSHOT` - `com.google.http-client:google-http-client-bom 1.43.1` - `com.google.oauth-client:google-oauth-client-bom 1.34.1` - `com.google.api-client:google-api-client-bom 2.2.0` - `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT` - `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT` </details> <details><summary>java-shared-dependencies/pom.xml</summary> - `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT` - `com.google.cloud:first-party-dependencies 3.6.1-SNAPSHOT` - `com.google.cloud:third-party-dependencies 3.6.1-SNAPSHOT` </details> <details><summary>java-shared-dependencies/third-party-dependencies/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.5` - `org.apache.httpcomponents:httpcore 4.4.16` - `org.apache.httpcomponents:httpclient 4.5.14` - `org.threeten:threetenbp 1.6.7` - `javax.annotation:javax.annotation-api 1.3.2` - `org.codehaus.mojo:animal-sniffer-annotations 1.23` - `com.google.code.findbugs:jsr305 3.0.2` - `com.google.errorprone:error_prone_annotations 2.18.0` - `com.fasterxml.jackson:jackson-bom 2.14.2` - `commons-codec:commons-codec 1.15` - `io.opencensus:opencensus-api 0.31.1` - `io.opencensus:opencensus-contrib-grpc-util 0.31.1` - `io.opencensus:opencensus-contrib-http-util 0.31.1` - `io.opencensus:opencensus-contrib-zpages 0.31.1` - `io.opencensus:opencensus-exporter-stats-stackdriver 0.31.1` - `io.opencensus:opencensus-exporter-trace-stackdriver 0.31.1` - `io.opencensus:opencensus-impl 0.31.1` - `io.opencensus:opencensus-impl-core 0.31.1` - `org.checkerframework:checker-qual 3.32.0` - `io.perfmark:perfmark-api 0.26.0` </details> <details><summary>java-shared-dependencies/upper-bound-check/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.5` - `com.google.cloud:google-cloud-shared-dependencies 3.6.1-SNAPSHOT` </details> <details><summary>pom.xml</summary> - `com.coveo:fmt-maven-plugin 2.9` - `org.apache.maven.plugins:maven-deploy-plugin 3.1.1` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` </details> <details><summary>showcase/gapic-showcase/pom.xml</summary> - `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` - `org.jacoco:jacoco-maven-plugin 0.8.8` - `com.coveo:fmt-maven-plugin 2.9` - `com.googlecode.maven-download-plugin:download-maven-plugin 1.6.8` - `com.google.truth:truth 1.1.3` </details> <details><summary>showcase/grpc-gapic-showcase-v1beta1/pom.xml</summary> - `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` </details> <details><summary>showcase/pom.xml</summary> - `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT` - `com.google.api.grpc:proto-gapic-showcase-v1beta1 0.0.1-SNAPSHOT` - `com.google.api.grpc:grpc-gapic-showcase-v1beta1 0.0.1-SNAPSHOT` - `com.google.cloud:gapic-showcase 0.0.1-SNAPSHOT` - `junit:junit 4.13.2` - `org.apache.maven.plugins:maven-surefire-plugin 3.0.0-M8` </details> <details><summary>showcase/proto-gapic-showcase-v1beta1/pom.xml</summary> - `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` - `org.codehaus.mojo:exec-maven-plugin 3.1.0` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>.kokoro/requirements.txt</summary> - `attrs ==22.1.0` - `cachetools ==4.2.4` - `certifi ==2022.12.7` - `cffi ==1.15.1` - `charset-normalizer ==2.0.12` - `click ==8.0.4` - `colorlog ==6.7.0` - `cryptography ==39.0.1` - `gcp-docuploader ==0.6.3` - `gcp-releasetool ==1.10.5` - `google-api-core ==2.8.2` - `google-auth ==2.14.1` - `google-cloud-core ==2.3.1` - `google-cloud-storage ==2.0.0` - `google-crc32c ==1.3.0` - `google-resumable-media ==2.3.3` - `googleapis-common-protos ==1.56.3` - `idna ==3.4` - `importlib-metadata ==4.8.3` - `jeepney ==0.7.1` - `jinja2 ==3.0.3` - `keyring ==23.4.1` - `markupsafe ==2.0.1` - `packaging ==21.3` - `protobuf ==3.19.5` - `pyasn1 ==0.4.8` - `pyasn1-modules ==0.2.8` - `pycparser ==2.21` - `pyjwt ==2.4.0` - `pyparsing ==3.0.9` - `pyperclip ==1.8.2` - `python-dateutil ==2.8.2` - `requests ==2.27.1` - `rsa ==4.9` - `secretstorage ==3.3.3` - `six ==1.16.0` - `typing-extensions ==4.1.1` - `urllib3 ==1.26.12` - `zipp ==3.6.0` </details> </blockquote> </details> <details><summary>regex</summary> <blockquote> <details><summary>gax-java/dependencies.properties</summary> - `com.google.api.grpc:proto-google-common-protos 2.14.3` - `com.google.api.grpc:grpc-google-common-protos 2.14.3` - `com.google.auth:google-auth-library-oauth2-http 1.16.0` - `com.google.auth:google-auth-library-credentials 1.16.0` - `io.opencensus:opencensus-api 0.31.1` - `io.opencensus:opencensus-contrib-grpc-metrics 0.31.1` - `io.opencensus:opencensus-contrib-http-util 0.31.1` - `io.netty:netty-tcnative-boringssl-static 2.0.59.Final` - `io.netty:netty-handler 4.1.90.Final` - `io.netty:netty-common 4.1.90.Final` - `io.netty:netty-codec-socks 4.1.90.Final` - `io.netty:netty-codec-http2 4.1.90.Final` - `io.netty:netty-codec-http 4.1.90.Final` - `io.netty:netty-codec 4.1.90.Final` - `io.netty:netty-buffer 4.1.90.Final` - `io.netty:netty-resolver 4.1.90.Final` - `io.netty:netty-transport 4.1.90.Final` - `io.netty:netty-handler-proxy 4.1.90.Final` - `io.netty:netty-transport-native-unix-common 4.1.90.Final` - `io.perfmark:perfmark-api 0.26.0` - `org.apache.tomcat:annotations-api 6.0.53` - `com.google.code.gson:gson 2.10.1` - `com.google.guava:guava 31.1-jre` - `com.google.guava:failureaccess 1.0.1` - `org.apache.commons:commons-lang3 3.12.0` - `com.google.android:annotations 4.1.1.4` - `com.google.code.findbugs:jsr305 3.0.2` - `com.google.errorprone:error_prone_annotations 2.18.0` - `com.google.j2objc:j2objc-annotations 1.3` - `com.google.auto.value:auto-value 1.10.1` - `com.google.auto.value:auto-value-annotations 1.10.1` - `com.google.api:api-common 2.6.3` - `org.threeten:threetenbp 1.6.7` - `com.google.api.grpc:grpc-google-iam-v1 1.9.3` - `com.google.api.grpc:proto-google-iam-v1 1.9.3` - `com.google.http-client:google-http-client 1.43.1` - `com.google.http-client:google-http-client-gson 1.43.1` - `org.codehaus.mojo:animal-sniffer-annotations 1.23` - `javax.annotation:javax.annotation-api 1.3.2` - `org.graalvm.sdk:graal-sdk 22.3.1` - `junit:junit 4.13.2` - `org.mockito:mockito-core 2.28.2` - `org.hamcrest:hamcrest-core 1.3` - `com.google.truth:truth 1.1.3` - `com.googlecode.java-diff-utils:diffutils 1.3.0` - `net.bytebuddy:byte-buddy 1.14.2` - `org.objenesis:objenesis 2.6` </details> <details><summary>gax-java/dependencies.properties</summary> - `com.google.protobuf:protobuf-java 3.21.10` </details> <details><summary>gax-java/dependencies.properties</summary> - `com.google.googlejavaformat:google-java-format 1.15.0` </details> <details><summary>gax-java/dependencies.properties</summary> - `io.grpc:grpc-core 1.54.0` </details> </blockquote> </details> --- - [x] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more rate limited these updates are currently rate limited click on a checkbox below to force their creation now build deps update dependency org jacoco jacoco maven plugin to deps update netty dependencies to final io netty netty buffer io netty netty codec io netty netty codec http io netty netty codec io netty netty codec socks io netty netty common io netty netty handler io netty netty handler proxy io netty netty resolver io netty netty transport io netty netty transport native unix common deps update dependency attrs to deps update dependency charset normalizer to deps update dependency click to deps update dependency gcp releasetool to deps update dependency google api core to deps update dependency google auth to deps update dependency google cloud storage to deps update dependency google to deps update dependency google resumable media to deps update dependency googleapis common protos to deps update dependency importlib metadata to deps update dependency jeepney to deps update dependency to deps update dependency keyring to deps update dependency markupsafe to deps update dependency protobuf to deps update dependency pyjwt to deps update dependency requests to deps update dependency typing extensions to deps update dependency zipp to 🔐 create all rate limited prs at once 🔐 ⚠ dependency lookup warnings ⚠ renovate failed to look up the following dependencies failed to look up maven package com google cloud gapic showcase failed to look up maven package com google cloud gapic showcase parent failed to look up maven package com google api grpc proto gapic showcase failed to look up maven package com google api grpc grpc gapic showcase files affected coverage report pom xml showcase gapic showcase pom xml showcase grpc gapic showcase pom xml showcase pom xml showcase proto gapic showcase pom xml open these updates have all been created already click a checkbox below to force a retry rebase of any pull pull pull pull pull pull pull pull com google api grpc grpc google common protos com google api grpc grpc google iam com google api grpc proto google common protos com google api grpc proto google iam com google api api common pull com google protobuf protobuf java com google protobuf protobuf bom click on this checkbox to rebase all open prs at once ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull pull detected dependencies bazel workspace github actions github workflows ci yaml actions checkout actions setup java actions checkout actions setup java actions setup java actions checkout actions setup java actions setup java actions checkout actions setup java actions checkout actions setup java actions checkout actions checkout actions setup java github workflows downstream native image yaml actions checkout stcarolas setup maven ayltai setup graalvm github workflows nightly yaml actions checkout actions setup java actions checkout actions setup java actions setup java github workflows shared dependencies yaml actions checkout stcarolas setup maven actions setup java github workflows sonar yaml actions checkout actions setup java actions cache actions cache gradle wrapper rules java gapic resources gradle gradle wrapper gradle wrapper properties gradle maven api common java pom xml com google api gapic generator java pom parent snapshot com google auto value auto value annotations com google auto value auto value com google code findbugs com google errorprone error prone annotations junit junit com google truth truth com google auto value auto value coverage report pom xml com google cloud gapic showcase snapshot com google api gax snapshot com google api gax grpc snapshot com google api gax httpjson snapshot com google api api common snapshot org jacoco jacoco maven plugin com coveo fmt maven plugin gapic generator java bom pom xml com google api gapic generator java pom parent snapshot com google api api common snapshot com google api gax bom snapshot com google api grpc grpc google common protos snapshot com google api grpc proto google common protos snapshot com google api grpc proto google iam snapshot com google api grpc proto google iam snapshot com google api grpc proto google iam snapshot com google api grpc grpc google iam snapshot com google api grpc grpc google iam snapshot com google api grpc grpc google iam snapshot gapic generator java pom parent pom xml com google guava guava jre javax annotation javax annotation api com google auth google auth library bom com google code gson gson com google guava guava bom jre com google protobuf protobuf bom io grpc grpc bom com google cloud google cloud shared config org apache maven plugins maven surefire plugin org apache maven plugins maven failsafe plugin org jacoco jacoco maven plugin javax annotation javax annotation api com google auth google auth library bom com google guava guava jre io grpc grpc bom com google code gson gson com google protobuf protobuf bom javax annotation javax annotation api com google protobuf protobuf bom io grpc grpc bom javax annotation javax annotation api com google guava guava bom jre com google protobuf protobuf bom io grpc grpc bom javax annotation javax annotation api gapic generator java pom xml com google api gapic generator java pom parent snapshot com google api gapic generator java bom snapshot org sonatype plugins nexus staging maven plugin kr motd maven os maven plugin org apache maven plugins maven compiler plugin com google auto value auto value com coveo fmt maven plugin com googlecode maven download plugin download maven plugin org xolstice maven plugins protobuf maven plugin org apache maven plugins maven surefire plugin org jacoco jacoco maven plugin org apache maven plugins maven shade plugin com google guava guava jre com google auto value auto value annotations org yaml snakeyaml com google googlejavaformat google java format junit junit com google errorprone error prone annotations com google truth truth io github java diff utils java diff utils gax java gax bom pom xml com google cloud google cloud shared config com google api gax snapshot com google api gax snapshot com google api gax snapshot com google api gax grpc snapshot com google api gax grpc snapshot com google api gax grpc snapshot com google api gax httpjson snapshot com google api gax httpjson snapshot com google api gax httpjson snapshot gax java gax grpc pom xml com google api gax parent snapshot gax java gax httpjson pom xml com google api gax parent snapshot gax java gax pom xml com google api gax parent snapshot gax java pom xml com google api gapic generator java pom parent snapshot com google api api common snapshot org threeten threetenbp com google code findbugs io opencensus opencensus api com google auto value auto value annotations com google auto value auto value junit junit org mockito mockito core com google api gax snapshot com google api gax snapshot com google api grpc proto google common protos snapshot com google api grpc grpc google common protos snapshot org graalvm sdk graal sdk com google http client google http client bom com google truth truth com google auto value auto value org sonatype plugins nexus staging maven plugin java common protos grpc google common protos pom xml com google api grpc google common protos parent snapshot java common protos pom xml com google api gapic generator java pom parent snapshot com google api grpc grpc google common protos snapshot com google api grpc proto google common protos snapshot com google guava guava jre junit junit org apache maven plugins maven project info reports plugin org apache maven plugins maven javadoc plugin java common protos proto google common protos pom xml com google api grpc google common protos parent snapshot java core google cloud core bom pom xml com google api gapic generator java pom parent snapshot com google cloud google cloud core snapshot com google cloud google cloud core grpc snapshot com google cloud google cloud core http snapshot java core google cloud core grpc pom xml com google cloud google cloud core parent snapshot java core google cloud core http pom xml com google cloud google cloud core parent snapshot java core google cloud core pom xml com google cloud google cloud core parent snapshot java core pom xml com google api gapic generator java pom parent snapshot com google cloud google cloud shared dependencies snapshot junit junit com google truth truth org easymock easymock org mockito mockito all java iam grpc google iam pom xml com google cloud google iam parent snapshot java iam grpc google iam pom xml com google cloud google iam parent snapshot java iam grpc google iam pom xml com google cloud google iam parent snapshot java iam pom xml com google api gapic generator java pom parent snapshot com google api gax bom snapshot com google api grpc proto google iam snapshot com google api grpc grpc google iam snapshot com google api grpc proto google common protos snapshot com google api grpc proto google iam snapshot com google api grpc grpc google iam snapshot com google api grpc grpc google iam snapshot com google api grpc proto google iam snapshot junit junit org apache maven plugins maven project info reports plugin org apache maven plugins maven javadoc plugin java iam proto google iam pom xml com google cloud google iam parent snapshot java iam proto google iam pom xml com google cloud google iam parent snapshot java iam proto google iam pom xml com google cloud google iam parent snapshot java shared dependencies dependency convergence check pom xml com google guava guava jre com google cloud tools dependencies java shared dependencies first party dependencies pom xml com google cloud google cloud shared config com google api gapic generator java bom snapshot com google cloud grpc gcp com google code gson gson com google cloud google cloud core bom snapshot com google http client google http client bom com google oauth client google oauth client bom com google api client google api client bom com google cloud google cloud core snapshot com google cloud google cloud core snapshot java shared dependencies pom xml com google api gapic generator java pom parent snapshot com google cloud first party dependencies snapshot com google cloud third party dependencies snapshot java shared dependencies third party dependencies pom xml com google cloud google cloud shared config org apache httpcomponents httpcore org apache httpcomponents httpclient org threeten threetenbp javax annotation javax annotation api org codehaus mojo animal sniffer annotations com google code findbugs com google errorprone error prone annotations com fasterxml jackson jackson bom commons codec commons codec io opencensus opencensus api io opencensus opencensus contrib grpc util io opencensus opencensus contrib http util io opencensus opencensus contrib zpages io opencensus opencensus exporter stats stackdriver io opencensus opencensus exporter trace stackdriver io opencensus opencensus impl io opencensus opencensus impl core org checkerframework checker qual io perfmark perfmark api java shared dependencies upper bound check pom xml com google cloud google cloud shared config com google cloud google cloud shared dependencies snapshot pom xml com coveo fmt maven plugin org apache maven plugins maven deploy plugin org sonatype plugins nexus staging maven plugin showcase gapic showcase pom xml com google cloud gapic showcase parent snapshot org codehaus mojo exec maven plugin org codehaus mojo exec maven plugin org jacoco jacoco maven plugin com coveo fmt maven plugin com googlecode maven download plugin download maven plugin com google truth truth showcase grpc gapic showcase pom xml com google cloud gapic showcase parent snapshot org codehaus mojo exec maven plugin org codehaus mojo exec maven plugin showcase pom xml com google api gapic generator java bom snapshot com google api grpc proto gapic showcase snapshot com google api grpc grpc gapic showcase snapshot com google cloud gapic showcase snapshot junit junit org apache maven plugins maven surefire plugin showcase proto gapic showcase pom xml com google cloud gapic showcase parent snapshot org codehaus mojo exec maven plugin org codehaus mojo exec maven plugin pip requirements kokoro requirements txt attrs cachetools certifi cffi charset normalizer click colorlog cryptography gcp docuploader gcp releasetool google api core google auth google cloud core google cloud storage google google resumable media googleapis common protos idna importlib metadata jeepney keyring markupsafe packaging protobuf modules pycparser pyjwt pyparsing pyperclip python dateutil requests rsa secretstorage six typing extensions zipp regex gax java dependencies properties com google api grpc proto google common protos com google api grpc grpc google common protos com google auth google auth library http com google auth google auth library credentials io opencensus opencensus api io opencensus opencensus contrib grpc metrics io opencensus opencensus contrib http util io netty netty tcnative boringssl static final io netty netty handler final io netty netty common final io netty netty codec socks final io netty netty codec final io netty netty codec http final io netty netty codec final io netty netty buffer final io netty netty resolver final io netty netty transport final io netty netty handler proxy final io netty netty transport native unix common final io perfmark perfmark api org apache tomcat annotations api com google code gson gson com google guava guava jre com google guava failureaccess org apache commons commons com google android annotations com google code findbugs com google errorprone error prone annotations com google annotations com google auto value auto value com google auto value auto value annotations com google api api common org threeten threetenbp com google api grpc grpc google iam com google api grpc proto google iam com google http client google http client com google http client google http client gson org codehaus mojo animal sniffer annotations javax annotation javax annotation api org graalvm sdk graal sdk junit junit org mockito mockito core org hamcrest hamcrest core com google truth truth com googlecode java diff utils diffutils net bytebuddy byte buddy org objenesis objenesis gax java dependencies properties com google protobuf protobuf java gax java dependencies properties com google googlejavaformat google java format gax java dependencies properties io grpc grpc core check this box to trigger a request for renovate to run again on this repository
1
82,903
16,060,119,227
IssuesEvent
2021-04-23 11:18:17
towavephone/GatsbyBlog
https://api.github.com/repos/towavephone/GatsbyBlog
opened
微前端框架qiankun源码笔记
/qiankun-code-note/ Gitalk
/qiankun-code-note/定义 是一种类似于微服务的架构,它将微服务的理念应用于浏览器端,即将单页面前端应用由单一的单体应用转变为多个小型前端应用聚合为一的应用。各个前端应用还可以独立开发、独立部署。同时,它们也可以在共享组件的同时进行并行开发——这些组件可以通过 NPM 或者 Git Tag、Git…
1.0
微前端框架qiankun源码笔记 - /qiankun-code-note/定义 是一种类似于微服务的架构,它将微服务的理念应用于浏览器端,即将单页面前端应用由单一的单体应用转变为多个小型前端应用聚合为一的应用。各个前端应用还可以独立开发、独立部署。同时,它们也可以在共享组件的同时进行并行开发——这些组件可以通过 NPM 或者 Git Tag、Git…
non_process
微前端框架qiankun源码笔记 qiankun code note 定义 是一种类似于微服务的架构,它将微服务的理念应用于浏览器端,即将单页面前端应用由单一的单体应用转变为多个小型前端应用聚合为一的应用。各个前端应用还可以独立开发、独立部署。同时,它们也可以在共享组件的同时进行并行开发——这些组件可以通过 npm 或者 git tag、git…
0
13,711
16,469,473,594
IssuesEvent
2021-05-23 05:37:17
pat-rogers/Ada-202x-WG9-Informal-Review
https://api.github.com/repos/pat-rogers/Ada-202x-WG9-Informal-Review
closed
5.6.1(4/5) and others, example not preceded by our usual "Example of ..." lead in
duplicate processed
Originator: Tucker Taft (via ARG list) Paragraph: 5.6.1(4/5), 5.5.3(30/5), 5.5.3(34/5), 6.3(8) The example in section 5.6.1 (Parallel Block Statements) is not preceded by our usual "Example of ..." lead in. In RM 6.3, we have "Example of procedure body:" and "Example of a function body:" in the same set of examples. It would seem we should add "a" to the first lead in. [Editor's notes: Tucker had the wrong subclause reference - he probably got it from the bottom of the page, which shows the *last* subclause on a page. No idea why Word does that. Anyway, this has been true since Ada 9x - it's in the Sept 1993 edition. Must be important. ;-)] In RM 5.5.3 (Procedural Iterators) the lead ins to the examples are not italicized (paragraphs 30/5 and 34/5). In RM 5.5.2 (Generalized Loop Iteration) the example lead in is unlike most others. It is a comment in the same font size as the example. Normally the lead in is larger, italicized, and starts with "Example[s] of ...".
1.0
5.6.1(4/5) and others, example not preceded by our usual "Example of ..." lead in - Originator: Tucker Taft (via ARG list) Paragraph: 5.6.1(4/5), 5.5.3(30/5), 5.5.3(34/5), 6.3(8) The example in section 5.6.1 (Parallel Block Statements) is not preceded by our usual "Example of ..." lead in. In RM 6.3, we have "Example of procedure body:" and "Example of a function body:" in the same set of examples. It would seem we should add "a" to the first lead in. [Editor's notes: Tucker had the wrong subclause reference - he probably got it from the bottom of the page, which shows the *last* subclause on a page. No idea why Word does that. Anyway, this has been true since Ada 9x - it's in the Sept 1993 edition. Must be important. ;-)] In RM 5.5.3 (Procedural Iterators) the lead ins to the examples are not italicized (paragraphs 30/5 and 34/5). In RM 5.5.2 (Generalized Loop Iteration) the example lead in is unlike most others. It is a comment in the same font size as the example. Normally the lead in is larger, italicized, and starts with "Example[s] of ...".
process
and others example not preceded by our usual example of lead in originator tucker taft via arg list paragraph the example in section parallel block statements is not preceded by our usual example of lead in in rm we have example of procedure body and example of a function body in the same set of examples it would seem we should add a to the first lead in editor s notes tucker had the wrong subclause reference he probably got it from the bottom of the page which shows the last subclause on a page no idea why word does that anyway this has been true since ada it s in the sept edition must be important in rm procedural iterators the lead ins to the examples are not italicized paragraphs and in rm generalized loop iteration the example lead in is unlike most others it is a comment in the same font size as the example normally the lead in is larger italicized and starts with example of
1
22,351
31,027,533,588
IssuesEvent
2023-08-10 10:08:57
DxytJuly3/gitalk_blog
https://api.github.com/repos/DxytJuly3/gitalk_blog
opened
[Linux] 进程状态相关概念、Linux实际进程状态、进程优先级 - July.cc Blogs
Gitalk /posts/Linux-Process-States
https://www.julysblog.cn/posts/Linux-Process-States 生成进程之后, 操作系统会给进程分配其所需要的资源。而为了整个操作系统的稳定和效率, 操作系统会对进程进行调度、管理, 在调度管理的过程中进程会存在不同的状态, 这些不同的状态就被称为进程状态。
1.0
[Linux] 进程状态相关概念、Linux实际进程状态、进程优先级 - July.cc Blogs - https://www.julysblog.cn/posts/Linux-Process-States 生成进程之后, 操作系统会给进程分配其所需要的资源。而为了整个操作系统的稳定和效率, 操作系统会对进程进行调度、管理, 在调度管理的过程中进程会存在不同的状态, 这些不同的状态就被称为进程状态。
process
进程状态相关概念、linux实际进程状态、进程优先级 july cc blogs 生成进程之后 操作系统会给进程分配其所需要的资源。而为了整个操作系统的稳定和效率 操作系统会对进程进行调度、管理 在调度管理的过程中进程会存在不同的状态 这些不同的状态就被称为进程状态。
1
7,304
10,443,164,173
IssuesEvent
2019-09-18 14:23:51
threefoldtech/0-core
https://api.github.com/repos/threefoldtech/0-core
closed
add primitives to move a nic into a network namespace
process_wontfix type_feature
it is a primitive for `ip link set $interface netns @namespace`
1.0
add primitives to move a nic into a network namespace - it is a primitive for `ip link set $interface netns @namespace`
process
add primitives to move a nic into a network namespace it is a primitive for ip link set interface netns namespace
1
6,324
9,345,009,826
IssuesEvent
2019-03-30 03:04:23
bow-simulation/virtualbow
https://api.github.com/repos/bow-simulation/virtualbow
closed
Consider vcpkg for dependency management
area: software process type: improvement
In GitLab by **spfeifer** on Dec 5, 2018, 16:32 Anouncement: https://blogs.msdn.microsoft.com/vcblog/2018/04/24/announcing-a-single-c-library-manager-for-linux-macos-and-windows-vcpkg/ Discussions: https://www.reddit.com/r/cpp/comments/8emja5/vcpkg_library_manager_now_also_on_linux_and_macos/ https://www.reddit.com/r/cpp/comments/9m4l0p/conan_vcpkg_or_build2/ Available libraries so far: https://github.com/Microsoft/vcpkg/tree/master/ports Setting specific library versions: https://github.com/Microsoft/vcpkg/issues/1681
1.0
Consider vcpkg for dependency management - In GitLab by **spfeifer** on Dec 5, 2018, 16:32 Anouncement: https://blogs.msdn.microsoft.com/vcblog/2018/04/24/announcing-a-single-c-library-manager-for-linux-macos-and-windows-vcpkg/ Discussions: https://www.reddit.com/r/cpp/comments/8emja5/vcpkg_library_manager_now_also_on_linux_and_macos/ https://www.reddit.com/r/cpp/comments/9m4l0p/conan_vcpkg_or_build2/ Available libraries so far: https://github.com/Microsoft/vcpkg/tree/master/ports Setting specific library versions: https://github.com/Microsoft/vcpkg/issues/1681
process
consider vcpkg for dependency management in gitlab by spfeifer on dec anouncement discussions available libraries so far setting specific library versions
1
4,358
7,260,513,756
IssuesEvent
2018-02-18 10:53:06
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE][Processing] Minimal enclosing circle
Automatic new feature Processing
Original commit: https://github.com/qgis/QGIS/commit/e30f7044c9af22096dda8722d188d452228d420c by nyalldawson Unfortunately this naughty coder did not write a description... :-(
1.0
[FEATURE][Processing] Minimal enclosing circle - Original commit: https://github.com/qgis/QGIS/commit/e30f7044c9af22096dda8722d188d452228d420c by nyalldawson Unfortunately this naughty coder did not write a description... :-(
process
minimal enclosing circle original commit by nyalldawson unfortunately this naughty coder did not write a description
1
78
2,531,050,807
IssuesEvent
2015-01-23 01:12:30
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
closed
Add branch() to GraphTraversal.
enhancement process
This should probably happen after you figure out how to elegantly make JumpStep extend it and thus, solidify the BranchStep API.
1.0
Add branch() to GraphTraversal. - This should probably happen after you figure out how to elegantly make JumpStep extend it and thus, solidify the BranchStep API.
process
add branch to graphtraversal this should probably happen after you figure out how to elegantly make jumpstep extend it and thus solidify the branchstep api
1
742,421
25,853,684,821
IssuesEvent
2022-12-13 12:13:47
pystardust/ani-cli
https://api.github.com/repos/pystardust/ani-cli
closed
Create directory if needed
type: feature request priority 4: wishlist category: ui ux good first issue
**Is your feature request related to a problem? Please describe.** Currently if you want to download to a specified folder (flag `-p`) and the folder doesn't exist, the download fails **Describe the solution you'd like** Detect if folder doesn't exist, and create it (using `mkdir -p`). Or we don't create a folder at all, but exit right on startup with an error message. Or default to current folder (with a verbose warning). **Describe alternatives you've considered** Using `mkdir` without the `-p` flag might be more stable **Additional context** Feature was discussed on [discord](https://discord.com/channels/931186431215435807/932006490703822938/1008361758341222521)
1.0
Create directory if needed - **Is your feature request related to a problem? Please describe.** Currently if you want to download to a specified folder (flag `-p`) and the folder doesn't exist, the download fails **Describe the solution you'd like** Detect if folder doesn't exist, and create it (using `mkdir -p`). Or we don't create a folder at all, but exit right on startup with an error message. Or default to current folder (with a verbose warning). **Describe alternatives you've considered** Using `mkdir` without the `-p` flag might be more stable **Additional context** Feature was discussed on [discord](https://discord.com/channels/931186431215435807/932006490703822938/1008361758341222521)
non_process
create directory if needed is your feature request related to a problem please describe currently if you want to download to a specified folder flag p and the folder doesn t exist the download fails describe the solution you d like detect if folder doesn t exist and create it using mkdir p or we don t create a folder at all but exit right on startup with an error message or default to current folder with a verbose warning describe alternatives you ve considered using mkdir without the p flag might be more stable additional context feature was discussed on
0
27,048
13,168,820,911
IssuesEvent
2020-08-11 12:48:00
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
reopened
FNX3-16565 ⁃ Fenix should run places maintenance periodically
P3 eng:health eng:performance
@grigoryk indicated that it does not, which will result in worse perf, more disk usage, etc.
True
FNX3-16565 ⁃ Fenix should run places maintenance periodically - @grigoryk indicated that it does not, which will result in worse perf, more disk usage, etc.
non_process
⁃ fenix should run places maintenance periodically grigoryk indicated that it does not which will result in worse perf more disk usage etc
0
21,768
30,287,305,883
IssuesEvent
2023-07-08 21:13:08
MikaylaFischler/cc-mek-scada
https://api.github.com/repos/MikaylaFischler/cc-mek-scada
closed
Process Waste Control
supervisor coordinator process control
Automatic (coordinated) waste processing control. Support optionally switching to plutonium production during the night when polonium processing is not possible, which would also apply to switching off of antimatter. - [x] Auto waste control interface - [x] Coordinated waste control (all units assigned to the same target) - [x] Day/night balancing option requiring the below task - [x] Solar neutron activator integration - [x] SPS integration (basic status) Auto control will always control all units assigned to the same output. If the user wants some to do one thing and others to do another, that's what manual assignment is for.
1.0
Process Waste Control - Automatic (coordinated) waste processing control. Support optionally switching to plutonium production during the night when polonium processing is not possible, which would also apply to switching off of antimatter. - [x] Auto waste control interface - [x] Coordinated waste control (all units assigned to the same target) - [x] Day/night balancing option requiring the below task - [x] Solar neutron activator integration - [x] SPS integration (basic status) Auto control will always control all units assigned to the same output. If the user wants some to do one thing and others to do another, that's what manual assignment is for.
process
process waste control automatic coordinated waste processing control support optionally switching to plutonium production during the night when polonium processing is not possible which would also apply to switching off of antimatter auto waste control interface coordinated waste control all units assigned to the same target day night balancing option requiring the below task solar neutron activator integration sps integration basic status auto control will always control all units assigned to the same output if the user wants some to do one thing and others to do another that s what manual assignment is for
1
16,857
22,137,876,638
IssuesEvent
2022-06-03 02:05:17
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
opened
Add artifact build and upload to 0.1 release branch
enhancement P1 process
### Problem `release/0.1` is missing logic to archive and upload the npm package. ### Solution Add logic to archive and upload artifact ### Alternatives _No response_
1.0
Add artifact build and upload to 0.1 release branch - ### Problem `release/0.1` is missing logic to archive and upload the npm package. ### Solution Add logic to archive and upload artifact ### Alternatives _No response_
process
add artifact build and upload to release branch problem release is missing logic to archive and upload the npm package solution add logic to archive and upload artifact alternatives no response
1
93,526
8,432,995,197
IssuesEvent
2018-10-17 05:17:22
hazelcast/hazelcast-jet
https://api.github.com/repos/hazelcast/hazelcast-jet
closed
com.hazelcast.jet.impl.connector.ReadWithPartitionIteratorP_MigrationDetectionTest.when_migration_then_detected_remote
test-failure
https://hazelcast-l337.ci.cloudbees.com/view/Jet/job/jet/com.hazelcast.jet$hazelcast-jet-core/927/testReport/junit/com.hazelcast.jet.impl.connector/ReadWithPartitionIteratorP_MigrationDetectionTest/when_migration_then_detected_remote/ ``` Error Message Expected test to throw (an instance of java.util.concurrent.ExecutionException and exception with message a string containing "migration detected") Stacktrace java.lang.AssertionError: Expected test to throw (an instance of java.util.concurrent.ExecutionException and exception with message a string containing "migration detected") at org.junit.Assert.fail(Assert.java:88) at org.junit.rules.ExpectedException.failDueToMissingException(ExpectedException.java:263) at org.junit.rules.ExpectedException.access$200(ExpectedException.java:106) at org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:245) at com.hazelcast.test.DumpBuildInfoOnFailureRule$1.evaluate(DumpBuildInfoOnFailureRule.java:37) ```
1.0
com.hazelcast.jet.impl.connector.ReadWithPartitionIteratorP_MigrationDetectionTest.when_migration_then_detected_remote - https://hazelcast-l337.ci.cloudbees.com/view/Jet/job/jet/com.hazelcast.jet$hazelcast-jet-core/927/testReport/junit/com.hazelcast.jet.impl.connector/ReadWithPartitionIteratorP_MigrationDetectionTest/when_migration_then_detected_remote/ ``` Error Message Expected test to throw (an instance of java.util.concurrent.ExecutionException and exception with message a string containing "migration detected") Stacktrace java.lang.AssertionError: Expected test to throw (an instance of java.util.concurrent.ExecutionException and exception with message a string containing "migration detected") at org.junit.Assert.fail(Assert.java:88) at org.junit.rules.ExpectedException.failDueToMissingException(ExpectedException.java:263) at org.junit.rules.ExpectedException.access$200(ExpectedException.java:106) at org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:245) at com.hazelcast.test.DumpBuildInfoOnFailureRule$1.evaluate(DumpBuildInfoOnFailureRule.java:37) ```
non_process
com hazelcast jet impl connector readwithpartitioniteratorp migrationdetectiontest when migration then detected remote error message expected test to throw an instance of java util concurrent executionexception and exception with message a string containing migration detected stacktrace java lang assertionerror expected test to throw an instance of java util concurrent executionexception and exception with message a string containing migration detected at org junit assert fail assert java at org junit rules expectedexception failduetomissingexception expectedexception java at org junit rules expectedexception access expectedexception java at org junit rules expectedexception expectedexceptionstatement evaluate expectedexception java at com hazelcast test dumpbuildinfoonfailurerule evaluate dumpbuildinfoonfailurerule java
0
3,312
2,755,737,677
IssuesEvent
2015-04-26 22:17:56
PerlDancer/Dancer2
https://api.github.com/repos/PerlDancer/Dancer2
closed
Make prefix a full spec like a path
Beginner Suitable Documentation Enhancement Testing
Currently the `prefix` can only be a simple string, it cannot be a regexp in and of itself. We should allow it to have the same spec as a regular path. This is mentioned (and, in a way, is part of) #558.
1.0
Make prefix a full spec like a path - Currently the `prefix` can only be a simple string, it cannot be a regexp in and of itself. We should allow it to have the same spec as a regular path. This is mentioned (and, in a way, is part of) #558.
non_process
make prefix a full spec like a path currently the prefix can only be a simple string it cannot be a regexp in and of itself we should allow it to have the same spec as a regular path this is mentioned and in a way is part of
0
15,006
18,720,010,133
IssuesEvent
2021-11-03 10:40:53
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Fix 'modulation of complement activation in other organism' and children
multi-species process
Hello, I think the following terms should be 'in host' rather than 'in other organism': 'modulation of complement activation in other organism' and children: 'modulation of complement activation, alternative pathway in other organism' 'negative regulation of complement activation, alternative pathway in other organism' 'positive regulation of complement activation, alternative pathway in other organism' 'modulation of complement activation, classical pathway in other organism' 'negative regulation of complement activation, classical pathway in other organism' 'positive regulation of complement activation, classical pathway in other organism' 'modulation of complement activation, lectin pathway in other organism' 'negative regulation of complement activation, lectin pathway in other organism' 'positive regulation of complement activation, lectin pathway in other organism' 'modulation of mast cell degranulation in other organism' 'positive regulation of mast cell degranulation in other organism' (note that there are no annotations to any of these terms) OK @mgiglio99 @addiehl @ValWood ? Thanks, Pascale
1.0
Fix 'modulation of complement activation in other organism' and children - Hello, I think the following terms should be 'in host' rather than 'in other organism': 'modulation of complement activation in other organism' and children: 'modulation of complement activation, alternative pathway in other organism' 'negative regulation of complement activation, alternative pathway in other organism' 'positive regulation of complement activation, alternative pathway in other organism' 'modulation of complement activation, classical pathway in other organism' 'negative regulation of complement activation, classical pathway in other organism' 'positive regulation of complement activation, classical pathway in other organism' 'modulation of complement activation, lectin pathway in other organism' 'negative regulation of complement activation, lectin pathway in other organism' 'positive regulation of complement activation, lectin pathway in other organism' 'modulation of mast cell degranulation in other organism' 'positive regulation of mast cell degranulation in other organism' (note that there are no annotations to any of these terms) OK @mgiglio99 @addiehl @ValWood ? Thanks, Pascale
process
fix modulation of complement activation in other organism and children hello i think the following terms should be in host rather than in other organism modulation of complement activation in other organism and children modulation of complement activation alternative pathway in other organism negative regulation of complement activation alternative pathway in other organism positive regulation of complement activation alternative pathway in other organism modulation of complement activation classical pathway in other organism negative regulation of complement activation classical pathway in other organism positive regulation of complement activation classical pathway in other organism modulation of complement activation lectin pathway in other organism negative regulation of complement activation lectin pathway in other organism positive regulation of complement activation lectin pathway in other organism modulation of mast cell degranulation in other organism positive regulation of mast cell degranulation in other organism note that there are no annotations to any of these terms ok addiehl valwood thanks pascale
1
2,709
5,578,337,699
IssuesEvent
2017-03-28 12:08:53
deb-sandeep/JoveNotesWebApp
https://api.github.com/repos/deb-sandeep/JoveNotesWebApp
closed
Introduction of @exercise type
enhancement jove_notes_db jove_notes_grammar jove_notes_processor jove_notes_server jove_notes_ui
## Support for exercise questions ### Background In its current form JoveNotes is primarily focused towards the bulk of study for K1-8 classes - retention and recollection (RnR). However as we move towards classes XI and beyond, there is a marked shift in focus towards applicability and problem solving. This enhancement is an endeavor to stretch the envelop of JoveNotes to foray into the arena of applicability. The current vision is to incorporate the following aspects of applicability class of problems: 1. Capture / Digitization - of exercise questions 2. Presentment 3. Operational tracking 4. Collecting data points 5. Providing insights by analyzing data points 6. Scoring of points - extension to current gamification ### In which way are exercise different than the current note element types? Exercise problems (for example, essays, numericals, problem solving etc) are characteristically different from existing note element types, fundamentally because it addresses a different segment of the education topology - applicability, while the existing note elements address the retention and recollection part. Salient points which differentiates exercises from other notes elements are: - **Presentment** - Exercises should not be mixed with RnR presentment, especially during the flash card sessions for the following reasons: - RnR flash cards sessions are rapid fire sessions with an average turn around time per question around 15-20 seconds. Numericals will not fit into the scheme - gear shift jerks. - RnR flash cards are designed to funnel focus on the screen - Numericals will have to be solved manually, causing a loss of focus on RnR streak. - Numericals require a cluster attempt behavior - read the questions paper, prioritize, review, rework, mark and then submit. This is contrary to the atomic way RnR questions are presented. - **Operational tracking** - - For one, numericals don't require the five step spaced sequenced repetition like those of RnR. - The time span between re-presentment is different as compared to RnR - Lot more data needs to be collected at session level - pre-read time, work, review, rework etc ## Modules impacted 1. JoveNotes grammar - com.sandy.xtext.jovenotes - com.sandy.xtext.jovenotes.ui - com.sandy.xtext.jovenotes.tests 2. JoveNotes processor 3. Database 4. JoveNotesWebApp - Dashboard - New numericals section 5. JoveNotesMaker ## Feature branch `feature/@exercise` ## Example of DSL ### DSL Grammar ``` '@exercise' ( hideFromView = 'hide' )? ( 'marks' '=' marks = INT ) ( question = STRING ) ( 'hints' '{' ( hints += STRING )+ '}' )? ( 'answer' answer = STRING ) ``` ### DSL example ``` java @exercise marks=20 "What is the velocity of an object after 5 seconds which starts from rest and accelerating for 2 seconds at 5 m/s^2?" hints { "If the body starts from rest, initial velocity is zero" "Once the body stops accelerating, velocity remains constant" } answer "10 m/s" ```
1.0
Introduction of @exercise type - ## Support for exercise questions ### Background In its current form JoveNotes is primarily focused towards the bulk of study for K1-8 classes - retention and recollection (RnR). However as we move towards classes XI and beyond, there is a marked shift in focus towards applicability and problem solving. This enhancement is an endeavor to stretch the envelop of JoveNotes to foray into the arena of applicability. The current vision is to incorporate the following aspects of applicability class of problems: 1. Capture / Digitization - of exercise questions 2. Presentment 3. Operational tracking 4. Collecting data points 5. Providing insights by analyzing data points 6. Scoring of points - extension to current gamification ### In which way are exercise different than the current note element types? Exercise problems (for example, essays, numericals, problem solving etc) are characteristically different from existing note element types, fundamentally because it addresses a different segment of the education topology - applicability, while the existing note elements address the retention and recollection part. Salient points which differentiates exercises from other notes elements are: - **Presentment** - Exercises should not be mixed with RnR presentment, especially during the flash card sessions for the following reasons: - RnR flash cards sessions are rapid fire sessions with an average turn around time per question around 15-20 seconds. Numericals will not fit into the scheme - gear shift jerks. - RnR flash cards are designed to funnel focus on the screen - Numericals will have to be solved manually, causing a loss of focus on RnR streak. - Numericals require a cluster attempt behavior - read the questions paper, prioritize, review, rework, mark and then submit. This is contrary to the atomic way RnR questions are presented. - **Operational tracking** - - For one, numericals don't require the five step spaced sequenced repetition like those of RnR. - The time span between re-presentment is different as compared to RnR - Lot more data needs to be collected at session level - pre-read time, work, review, rework etc ## Modules impacted 1. JoveNotes grammar - com.sandy.xtext.jovenotes - com.sandy.xtext.jovenotes.ui - com.sandy.xtext.jovenotes.tests 2. JoveNotes processor 3. Database 4. JoveNotesWebApp - Dashboard - New numericals section 5. JoveNotesMaker ## Feature branch `feature/@exercise` ## Example of DSL ### DSL Grammar ``` '@exercise' ( hideFromView = 'hide' )? ( 'marks' '=' marks = INT ) ( question = STRING ) ( 'hints' '{' ( hints += STRING )+ '}' )? ( 'answer' answer = STRING ) ``` ### DSL example ``` java @exercise marks=20 "What is the velocity of an object after 5 seconds which starts from rest and accelerating for 2 seconds at 5 m/s^2?" hints { "If the body starts from rest, initial velocity is zero" "Once the body stops accelerating, velocity remains constant" } answer "10 m/s" ```
process
introduction of exercise type support for exercise questions background in its current form jovenotes is primarily focused towards the bulk of study for classes retention and recollection rnr however as we move towards classes xi and beyond there is a marked shift in focus towards applicability and problem solving this enhancement is an endeavor to stretch the envelop of jovenotes to foray into the arena of applicability the current vision is to incorporate the following aspects of applicability class of problems capture digitization of exercise questions presentment operational tracking collecting data points providing insights by analyzing data points scoring of points extension to current gamification in which way are exercise different than the current note element types exercise problems for example essays numericals problem solving etc are characteristically different from existing note element types fundamentally because it addresses a different segment of the education topology applicability while the existing note elements address the retention and recollection part salient points which differentiates exercises from other notes elements are presentment exercises should not be mixed with rnr presentment especially during the flash card sessions for the following reasons rnr flash cards sessions are rapid fire sessions with an average turn around time per question around seconds numericals will not fit into the scheme gear shift jerks rnr flash cards are designed to funnel focus on the screen numericals will have to be solved manually causing a loss of focus on rnr streak numericals require a cluster attempt behavior read the questions paper prioritize review rework mark and then submit this is contrary to the atomic way rnr questions are presented operational tracking for one numericals don t require the five step spaced sequenced repetition like those of rnr the time span between re presentment is different as compared to rnr lot more data needs to be collected at session level pre read time work review rework etc modules impacted jovenotes grammar com sandy xtext jovenotes com sandy xtext jovenotes ui com sandy xtext jovenotes tests jovenotes processor database jovenoteswebapp dashboard new numericals section jovenotesmaker feature branch feature exercise example of dsl dsl grammar exercise hidefromview hide marks marks int question string hints hints string answer answer string dsl example java exercise marks what is the velocity of an object after seconds which starts from rest and accelerating for seconds at m s hints if the body starts from rest initial velocity is zero once the body stops accelerating velocity remains constant answer m s
1
262,861
8,272,561,480
IssuesEvent
2018-09-16 21:30:13
javaee/glassfish
https://api.github.com/repos/javaee/glassfish
closed
Monitoring support for @DataSourceDefinition
Component: jdbc ERR: Assignee Priority: Major Type: New Feature
Existing implementation of @DataSourceDefinition is realized internally as a jdbc-resource and jdbc-connection-pool. It would be useful to expose monitoring statistics of @DataSourceDefinition (ie., the jdbc-connection-pool monitoring statistics) to users. Following monitoring hierarchy could be used to display the statistics according to the scope in which the @DataSourceDefinition is defined. server.<application-name>.datasource-definition.java:global/<global-scoped-dsd> server.<application-name>.datasource-definition.java:app/<app-scoped-dsd> server.<application-name>.datasource-definition.<module-name>.java:module/<module-scoped-dsd> server.<application-name>.datasource-definition.<module-name>.<component-name>.java:comp/<component-scoped-dsd> Note : 1) For component scoped @DSD in web bundle descriptors, following convention will be used server.<application-name>.<module-name>.datasource-definition.java:comp/<component-scoped-dsd> 2) For applications that are not .ear, <module-name> will not be applicable. #### Affected Versions [4.0]
1.0
Monitoring support for @DataSourceDefinition - Existing implementation of @DataSourceDefinition is realized internally as a jdbc-resource and jdbc-connection-pool. It would be useful to expose monitoring statistics of @DataSourceDefinition (ie., the jdbc-connection-pool monitoring statistics) to users. Following monitoring hierarchy could be used to display the statistics according to the scope in which the @DataSourceDefinition is defined. server.<application-name>.datasource-definition.java:global/<global-scoped-dsd> server.<application-name>.datasource-definition.java:app/<app-scoped-dsd> server.<application-name>.datasource-definition.<module-name>.java:module/<module-scoped-dsd> server.<application-name>.datasource-definition.<module-name>.<component-name>.java:comp/<component-scoped-dsd> Note : 1) For component scoped @DSD in web bundle descriptors, following convention will be used server.<application-name>.<module-name>.datasource-definition.java:comp/<component-scoped-dsd> 2) For applications that are not .ear, <module-name> will not be applicable. #### Affected Versions [4.0]
non_process
monitoring support for datasourcedefinition existing implementation of datasourcedefinition is realized internally as a jdbc resource and jdbc connection pool it would be useful to expose monitoring statistics of datasourcedefinition ie the jdbc connection pool monitoring statistics to users following monitoring hierarchy could be used to display the statistics according to the scope in which the datasourcedefinition is defined server datasource definition java global server datasource definition java app server datasource definition java module server datasource definition java comp note for component scoped dsd in web bundle descriptors following convention will be used server datasource definition java comp for applications that are not ear will not be applicable affected versions
0
339,568
10,256,216,394
IssuesEvent
2019-08-21 17:07:41
cerner/terra-core
https://api.github.com/repos/cerner/terra-core
closed
Update components which interface with old React context API
Intl Orion Reviewed i18n priority:medium
# Feature Request ## Description In a lot of the terra components that interface with react-intl, [we've been interfacing with the old React context API](https://github.com/cerner/terra-core/blob/master/packages/terra-demographics-banner/src/DemographicsBanner.jsx#L67-L74) directly. We should update the components to use an abstraction like `FormattedMessage` or `injectIntl` to interface with react-intl's Intl context. This should allows us to more easily update to react-intl v3 when it is released. More info on how we recommend to build components with `FormattedMessage` and/or `injectIntl`: https://engineering.cerner.com/terra-ui/#/getting-started/terra-ui/internationalization/building-components-which-include-translations
1.0
Update components which interface with old React context API - # Feature Request ## Description In a lot of the terra components that interface with react-intl, [we've been interfacing with the old React context API](https://github.com/cerner/terra-core/blob/master/packages/terra-demographics-banner/src/DemographicsBanner.jsx#L67-L74) directly. We should update the components to use an abstraction like `FormattedMessage` or `injectIntl` to interface with react-intl's Intl context. This should allows us to more easily update to react-intl v3 when it is released. More info on how we recommend to build components with `FormattedMessage` and/or `injectIntl`: https://engineering.cerner.com/terra-ui/#/getting-started/terra-ui/internationalization/building-components-which-include-translations
non_process
update components which interface with old react context api feature request description in a lot of the terra components that interface with react intl directly we should update the components to use an abstraction like formattedmessage or injectintl to interface with react intl s intl context this should allows us to more easily update to react intl when it is released more info on how we recommend to build components with formattedmessage and or injectintl
0
67,931
7,078,965,018
IssuesEvent
2018-01-10 07:27:50
GTNewHorizons/NewHorizons
https://api.github.com/repos/GTNewHorizons/NewHorizons
closed
Cetane-Boosted Diesel from Diesel not profit
FixedInDev need to be tested
#### Which modpack version are you using? 2.0.1.3 We can craft 500 mb Cetane-Boosted Diesel from 1000 mb Diesel in mixer. But if we count, 500 mb Cetane-Boosted Diesel give us 360000 eu, but 1000 mb Diesel give us 400000 eu. I suggest to change Cetane-Boosted Diesel output to 1000 mb.
1.0
Cetane-Boosted Diesel from Diesel not profit - #### Which modpack version are you using? 2.0.1.3 We can craft 500 mb Cetane-Boosted Diesel from 1000 mb Diesel in mixer. But if we count, 500 mb Cetane-Boosted Diesel give us 360000 eu, but 1000 mb Diesel give us 400000 eu. I suggest to change Cetane-Boosted Diesel output to 1000 mb.
non_process
cetane boosted diesel from diesel not profit which modpack version are you using we can craft mb cetane boosted diesel from mb diesel in mixer but if we count mb cetane boosted diesel give us eu but mb diesel give us eu i suggest to change cetane boosted diesel output to mb
0
7,931
11,114,444,329
IssuesEvent
2019-12-18 08:41:16
ocaml-batteries-team/batteries-included
https://api.github.com/repos/ocaml-batteries-team/batteries-included
opened
make v3 the new master
Easy development process
And have a maintenance branch named v2 or 2.x or whatever, for any upcoming fixes on the v2. WDYT? Have some time to do it quickly if everybody agree.
1.0
make v3 the new master - And have a maintenance branch named v2 or 2.x or whatever, for any upcoming fixes on the v2. WDYT? Have some time to do it quickly if everybody agree.
process
make the new master and have a maintenance branch named or x or whatever for any upcoming fixes on the wdyt have some time to do it quickly if everybody agree
1
19,097
25,148,011,302
IssuesEvent
2022-11-10 07:41:48
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Outputs attributes tables from Processing/SAGA have crippled special chars if input is not a shapefile
Processing Bug
Take a vector with special/accented characters in its attribute table. If you processing this with Processing/SAGA and the input format is not a shapefile then the output will have those chars crippled. This do not happen if the input is a shapefile and also do not happen if running from the CLI the same SAGA command that QGIS runs.
1.0
Outputs attributes tables from Processing/SAGA have crippled special chars if input is not a shapefile - Take a vector with special/accented characters in its attribute table. If you processing this with Processing/SAGA and the input format is not a shapefile then the output will have those chars crippled. This do not happen if the input is a shapefile and also do not happen if running from the CLI the same SAGA command that QGIS runs.
process
outputs attributes tables from processing saga have crippled special chars if input is not a shapefile take a vector with special accented characters in its attribute table if you processing this with processing saga and the input format is not a shapefile then the output will have those chars crippled this do not happen if the input is a shapefile and also do not happen if running from the cli the same saga command that qgis runs
1
125,571
26,688,403,993
IssuesEvent
2023-01-27 00:53:08
WordPress/openverse-catalog
https://api.github.com/repos/WordPress/openverse-catalog
closed
Re-enable Slack load completion notifications for Finnish
🟧 priority: high 🛠 goal: fix 💻 aspect: code
## Problem <!-- Describe a problem solved by this feature; or delete the section entirely. --> Finnish museums has been turned back on after converting it to a dated DAG. There is a backfill of several months worth of daily DagRuns running, which will spam `report_load_completion` messages in Slack at an abnormal rate. For the duration of the backfill, we're silencing those load completion messages for Finnish. This issue is to track: 1. manually reviewing the state of the backfill (errors will still be reported) 2. re-enabling the messages when the backfill completes <!-- If you would like to work on this, please comment below separately. -->
1.0
Re-enable Slack load completion notifications for Finnish - ## Problem <!-- Describe a problem solved by this feature; or delete the section entirely. --> Finnish museums has been turned back on after converting it to a dated DAG. There is a backfill of several months worth of daily DagRuns running, which will spam `report_load_completion` messages in Slack at an abnormal rate. For the duration of the backfill, we're silencing those load completion messages for Finnish. This issue is to track: 1. manually reviewing the state of the backfill (errors will still be reported) 2. re-enabling the messages when the backfill completes <!-- If you would like to work on this, please comment below separately. -->
non_process
re enable slack load completion notifications for finnish problem finnish museums has been turned back on after converting it to a dated dag there is a backfill of several months worth of daily dagruns running which will spam report load completion messages in slack at an abnormal rate for the duration of the backfill we re silencing those load completion messages for finnish this issue is to track manually reviewing the state of the backfill errors will still be reported re enabling the messages when the backfill completes
0
5,078
7,873,796,246
IssuesEvent
2018-06-25 15:11:46
AlexsLemonade/refinebio
https://api.github.com/repos/AlexsLemonade/refinebio
closed
Aggregate and Deliver User-Selected Samples for Download
api aws database email processor review sci review smasher
### Context Once a user has a created a set of samples that they wish to download, they need to tell us to prepare a download for them, so we need to pull them off of EFS/S3, aggregate them, move them to back to S3, then deliver the result to them either via the web or via email. This requires three main parts (in addition to the required front-end work): ### Problem or idea #### API Two new API endpoints are needed: one which accepts a `POST` with a list of sample IDs, a method of aggregation (and any other desired preferences), and an email, which returns a Download ID, and other which returns the status (and final URL) for a given Download ID which can be polled while the download is being prepared. _TODO: Investigate potential for abuse and mitigation strategies._ #### Processing This should be a queue-driven process which fetches the data from EFS/S3, smashes it as desired, packages it with metadata and moves it to S3. I have no idea how much processing power is required for this or how what the time bounds are for these jobs. #### Delivery Assuming the processing step takes more than _n_ seconds, we will need a way to deliver the final asset to the user. This will require constructing an email and ensuring its delivery via AWS SES. We can probably look to Cognoma as a model for this. ### Solution or next step I'd probably start this by figuring out how we want to format the metadata, and figuring how what types of aggregation we want to offer and how to do them. I have some Pandas code [here](https://github.com/Miserlou/science) that may or may not be useful for this. ### New Issue Checklist - [x] The title is short and descriptive. - [x] You have explained the context that led you to write this issue. - [x] You have reported a problem or idea. - [x] You have proposed a solution or next step.
1.0
Aggregate and Deliver User-Selected Samples for Download - ### Context Once a user has a created a set of samples that they wish to download, they need to tell us to prepare a download for them, so we need to pull them off of EFS/S3, aggregate them, move them to back to S3, then deliver the result to them either via the web or via email. This requires three main parts (in addition to the required front-end work): ### Problem or idea #### API Two new API endpoints are needed: one which accepts a `POST` with a list of sample IDs, a method of aggregation (and any other desired preferences), and an email, which returns a Download ID, and other which returns the status (and final URL) for a given Download ID which can be polled while the download is being prepared. _TODO: Investigate potential for abuse and mitigation strategies._ #### Processing This should be a queue-driven process which fetches the data from EFS/S3, smashes it as desired, packages it with metadata and moves it to S3. I have no idea how much processing power is required for this or how what the time bounds are for these jobs. #### Delivery Assuming the processing step takes more than _n_ seconds, we will need a way to deliver the final asset to the user. This will require constructing an email and ensuring its delivery via AWS SES. We can probably look to Cognoma as a model for this. ### Solution or next step I'd probably start this by figuring out how we want to format the metadata, and figuring how what types of aggregation we want to offer and how to do them. I have some Pandas code [here](https://github.com/Miserlou/science) that may or may not be useful for this. ### New Issue Checklist - [x] The title is short and descriptive. - [x] You have explained the context that led you to write this issue. - [x] You have reported a problem or idea. - [x] You have proposed a solution or next step.
process
aggregate and deliver user selected samples for download context once a user has a created a set of samples that they wish to download they need to tell us to prepare a download for them so we need to pull them off of efs aggregate them move them to back to then deliver the result to them either via the web or via email this requires three main parts in addition to the required front end work problem or idea api two new api endpoints are needed one which accepts a post with a list of sample ids a method of aggregation and any other desired preferences and an email which returns a download id and other which returns the status and final url for a given download id which can be polled while the download is being prepared todo investigate potential for abuse and mitigation strategies processing this should be a queue driven process which fetches the data from efs smashes it as desired packages it with metadata and moves it to i have no idea how much processing power is required for this or how what the time bounds are for these jobs delivery assuming the processing step takes more than n seconds we will need a way to deliver the final asset to the user this will require constructing an email and ensuring its delivery via aws ses we can probably look to cognoma as a model for this solution or next step i d probably start this by figuring out how we want to format the metadata and figuring how what types of aggregation we want to offer and how to do them i have some pandas code that may or may not be useful for this new issue checklist the title is short and descriptive you have explained the context that led you to write this issue you have reported a problem or idea you have proposed a solution or next step
1
492,857
14,221,665,757
IssuesEvent
2020-11-17 15:58:05
clastix/capsule
https://api.github.com/repos/clastix/capsule
closed
Add the node pool label as additionalPrinterColumns
enhancement low-priority
# Describe the feature It would be nice to have the node pool label of a Tenant as Additional Printer Column in the output of `kubectl get tenants` # What would the new user story look like? Having a tenant oil with a given node selector ```yaml apiVersion: capsule.clastix.io/v1alpha1 kind: Tenant metadata: name: oil spec: nodeSelector: pool: caas ... ``` you get ``` $ kubectl get tenants NAME NAMESPACE QUOTA NAMESPACE COUNT OWNER NAME OWNER KIND SELECTOR AGE oil 10 3 alice User pool=caas 39d ``` # Expected behavior see above
1.0
Add the node pool label as additionalPrinterColumns - # Describe the feature It would be nice to have the node pool label of a Tenant as Additional Printer Column in the output of `kubectl get tenants` # What would the new user story look like? Having a tenant oil with a given node selector ```yaml apiVersion: capsule.clastix.io/v1alpha1 kind: Tenant metadata: name: oil spec: nodeSelector: pool: caas ... ``` you get ``` $ kubectl get tenants NAME NAMESPACE QUOTA NAMESPACE COUNT OWNER NAME OWNER KIND SELECTOR AGE oil 10 3 alice User pool=caas 39d ``` # Expected behavior see above
non_process
add the node pool label as additionalprintercolumns describe the feature it would be nice to have the node pool label of a tenant as additional printer column in the output of kubectl get tenants what would the new user story look like having a tenant oil with a given node selector yaml apiversion capsule clastix io kind tenant metadata name oil spec nodeselector pool caas you get kubectl get tenants name namespace quota namespace count owner name owner kind selector age oil alice user pool caas expected behavior see above
0
8,430
11,596,180,898
IssuesEvent
2020-02-24 18:26:47
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
opened
Panic for empty env var
process/candidate
When having this schema (models don't really matter) ```prisma datasource db { provider = "sqlite" url = env("DB_URL") } generator client { provider = "prisma-client-js" binaryTargets = ["native"] } model User { id String @default(cuid()) @id email String @unique name String? posts Post[] treeHouseMaterial Tree? } model Post { id String @default(cuid()) @id createdAt DateTime @default(now()) updatedAt DateTime @updatedAt published Boolean title String content String? author User? } enum Tree { OAK DOUGLAS } ``` And the env var `DB_URL=""` set to an empty string, then starting the query engine in http mode, I get the following error, printed in stdout: ``` { timestamp: 'Feb 24 10:22:58.628', level: 'ERROR', target: 'prisma', fields: { message: 'PANIC', reason: 'called `Option::unwrap()` on a `None` value', file: '/rustc/5e1a799842ba6ed4a57e91f7ab9435947482f7d8/src/libcore/macros/mod.rs', line: 15, column: 40 } } ``` It would be great if we can turn this into a known error.
1.0
Panic for empty env var - When having this schema (models don't really matter) ```prisma datasource db { provider = "sqlite" url = env("DB_URL") } generator client { provider = "prisma-client-js" binaryTargets = ["native"] } model User { id String @default(cuid()) @id email String @unique name String? posts Post[] treeHouseMaterial Tree? } model Post { id String @default(cuid()) @id createdAt DateTime @default(now()) updatedAt DateTime @updatedAt published Boolean title String content String? author User? } enum Tree { OAK DOUGLAS } ``` And the env var `DB_URL=""` set to an empty string, then starting the query engine in http mode, I get the following error, printed in stdout: ``` { timestamp: 'Feb 24 10:22:58.628', level: 'ERROR', target: 'prisma', fields: { message: 'PANIC', reason: 'called `Option::unwrap()` on a `None` value', file: '/rustc/5e1a799842ba6ed4a57e91f7ab9435947482f7d8/src/libcore/macros/mod.rs', line: 15, column: 40 } } ``` It would be great if we can turn this into a known error.
process
panic for empty env var when having this schema models don t really matter prisma datasource db provider sqlite url env db url generator client provider prisma client js binarytargets model user id string default cuid id email string unique name string posts post treehousematerial tree model post id string default cuid id createdat datetime default now updatedat datetime updatedat published boolean title string content string author user enum tree oak douglas and the env var db url set to an empty string then starting the query engine in http mode i get the following error printed in stdout timestamp feb level error target prisma fields message panic reason called option unwrap on a none value file rustc src libcore macros mod rs line column it would be great if we can turn this into a known error
1
509,848
14,750,280,664
IssuesEvent
2021-01-08 01:43:25
GentenStudios/Phoenix
https://api.github.com/repos/GentenStudios/Phoenix
closed
Licences for art and other media
enhancement priority-high
## Story Currently not all files have a proper licence attached to them. We may wanna add one.
1.0
Licences for art and other media - ## Story Currently not all files have a proper licence attached to them. We may wanna add one.
non_process
licences for art and other media story currently not all files have a proper licence attached to them we may wanna add one
0
12,669
15,037,635,559
IssuesEvent
2021-02-02 16:33:16
retaildevcrews/ngsa
https://api.github.com/repos/retaildevcrews/ngsa
closed
Engineering Fundamentals Process
EngPrac Process
## Description What: - checklists and pre-work documentation Why: - to follow CSE recommended engineering fundamentals When: - monthly Where: - github, ado, and snowball ## Tasks - [X] Game Plan Document - [X] Game Plan Review - [x] CPRs - [x] Tech Led Engineering Checklist - [x] Engineering Surveys ## Acceptance Criteria - Game Plan review completed & plan uploaded - CPR pre-work completed and uploaded 24hrs in advance - Artifacts created and uploaded into ADO ## Constraints - N/A
1.0
Engineering Fundamentals Process - ## Description What: - checklists and pre-work documentation Why: - to follow CSE recommended engineering fundamentals When: - monthly Where: - github, ado, and snowball ## Tasks - [X] Game Plan Document - [X] Game Plan Review - [x] CPRs - [x] Tech Led Engineering Checklist - [x] Engineering Surveys ## Acceptance Criteria - Game Plan review completed & plan uploaded - CPR pre-work completed and uploaded 24hrs in advance - Artifacts created and uploaded into ADO ## Constraints - N/A
process
engineering fundamentals process description what checklists and pre work documentation why to follow cse recommended engineering fundamentals when monthly where github ado and snowball tasks game plan document game plan review cprs tech led engineering checklist engineering surveys acceptance criteria game plan review completed plan uploaded cpr pre work completed and uploaded in advance artifacts created and uploaded into ado constraints n a
1
15,937
20,160,881,993
IssuesEvent
2022-02-09 21:24:08
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Investigate Unit conversions in PH Flue gas
Process Heating
All the conversions for the inputs are working correctly (including the displayed heating values in the fuel modals) but the Available heat changes when you do a unit conversion. I wonder if there is a conversion that is getting missed before sending to the suite? Maybe check Ambient Air temp first (recently added field). If can't find anything, maybe make a note and move to the backlog/icebox. original ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/b2eb7691-2d92-4e12-81c5-fc5b4fd8b023) converted to metric ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/5e76ded1-dfa6-43a3-9a5c-50186e0e1736)
1.0
Investigate Unit conversions in PH Flue gas - All the conversions for the inputs are working correctly (including the displayed heating values in the fuel modals) but the Available heat changes when you do a unit conversion. I wonder if there is a conversion that is getting missed before sending to the suite? Maybe check Ambient Air temp first (recently added field). If can't find anything, maybe make a note and move to the backlog/icebox. original ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/b2eb7691-2d92-4e12-81c5-fc5b4fd8b023) converted to metric ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/5e76ded1-dfa6-43a3-9a5c-50186e0e1736)
process
investigate unit conversions in ph flue gas all the conversions for the inputs are working correctly including the displayed heating values in the fuel modals but the available heat changes when you do a unit conversion i wonder if there is a conversion that is getting missed before sending to the suite maybe check ambient air temp first recently added field if can t find anything maybe make a note and move to the backlog icebox original converted to metric
1
170,080
13,172,552,647
IssuesEvent
2020-08-11 18:38:46
Azure/azure-sdk-for-js
https://api.github.com/repos/Azure/azure-sdk-for-js
closed
Storage File Datalake Readme Issue
Client Docs Storage Up for grabs test-manual-pass
1. Section [Link](https://github.com/Azure/azure-sdk-for-js/tree/master/sdk/storage/storage-file-datalake#examples) ![image](https://user-images.githubusercontent.com/24564161/88270550-6370f100-cd08-11ea-8d5d-b3fe8c2c9930.png) Suggestion: Add a summary of examples. As follows: ![image](https://user-images.githubusercontent.com/24564161/88270595-75eb2a80-cd08-11ea-965d-3ee128a14ad3.png)
1.0
Storage File Datalake Readme Issue - 1. Section [Link](https://github.com/Azure/azure-sdk-for-js/tree/master/sdk/storage/storage-file-datalake#examples) ![image](https://user-images.githubusercontent.com/24564161/88270550-6370f100-cd08-11ea-8d5d-b3fe8c2c9930.png) Suggestion: Add a summary of examples. As follows: ![image](https://user-images.githubusercontent.com/24564161/88270595-75eb2a80-cd08-11ea-965d-3ee128a14ad3.png)
non_process
storage file datalake readme issue section suggestion add a summary of examples as follows
0
25,309
18,459,424,305
IssuesEvent
2021-10-15 21:35:30
bolsote/isoduration
https://api.github.com/repos/bolsote/isoduration
opened
Pin development dependencies
infrastructure
Right now, we have a fairly relaxed development requirements file. We should really pin all those versions, so we have a consistent environment everywhere. A `pip-tools`-based approach seems appropriate.
1.0
Pin development dependencies - Right now, we have a fairly relaxed development requirements file. We should really pin all those versions, so we have a consistent environment everywhere. A `pip-tools`-based approach seems appropriate.
non_process
pin development dependencies right now we have a fairly relaxed development requirements file we should really pin all those versions so we have a consistent environment everywhere a pip tools based approach seems appropriate
0
22,119
30,654,934,146
IssuesEvent
2023-07-25 11:27:54
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
skypilot-nightly 1.0.0.dev20230725 has 2 GuardDog issues
guarddog exec-base64 silent-process-execution
https://pypi.org/project/skypilot-nightly https://inspector.pypi.io/project/skypilot-nightly ```{ "dependency": "skypilot-nightly", "version": "1.0.0.dev20230725", "result": { "issues": 2, "errors": {}, "results": { "silent-process-execution": [ { "location": "skypilot-nightly-1.0.0.dev20230725/sky/skylet/log_lib.py:219", "code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ], "exec-base64": [ { "location": "skypilot-nightly-1.0.0.dev20230725/sky/cloud_stores.py:113", "code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ] }, "path": "/tmp/tmp_8rxbnqo/skypilot-nightly" } }```
1.0
skypilot-nightly 1.0.0.dev20230725 has 2 GuardDog issues - https://pypi.org/project/skypilot-nightly https://inspector.pypi.io/project/skypilot-nightly ```{ "dependency": "skypilot-nightly", "version": "1.0.0.dev20230725", "result": { "issues": 2, "errors": {}, "results": { "silent-process-execution": [ { "location": "skypilot-nightly-1.0.0.dev20230725/sky/skylet/log_lib.py:219", "code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ], "exec-base64": [ { "location": "skypilot-nightly-1.0.0.dev20230725/sky/cloud_stores.py:113", "code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ] }, "path": "/tmp/tmp_8rxbnqo/skypilot-nightly" } }```
process
skypilot nightly has guarddog issues dependency skypilot nightly version result issues errors results silent process execution location skypilot nightly sky skylet log lib py code subprocess popen n daemon cmd n start new session true n suppress output n stdout subprocess devnull n stderr subprocess devnull n disa message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null exec location skypilot nightly sky cloud stores py code p subprocess run command n stdout subprocess pipe n shell true n check true n executable bin bash message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n path tmp tmp skypilot nightly
1
146,902
5,630,374,239
IssuesEvent
2017-04-05 12:06:05
research-resource/research_resource
https://api.github.com/repos/research-resource/research_resource
closed
Add link to log in on sign up form
priority-2 T1h
Need to add a link underneath / next to (see UI) the sign up module to say "Already signed up? Log in" with the "Log in" linking to the log in page.
1.0
Add link to log in on sign up form - Need to add a link underneath / next to (see UI) the sign up module to say "Already signed up? Log in" with the "Log in" linking to the log in page.
non_process
add link to log in on sign up form need to add a link underneath next to see ui the sign up module to say already signed up log in with the log in linking to the log in page
0
6,110
8,969,444,908
IssuesEvent
2019-01-29 10:49:27
pwittchen/prefser
https://api.github.com/repos/pwittchen/prefser
closed
Release 2.2.2-rx2
release process
**Release notes**: - updated project dependencies - 74de9db550ff941083448c28c48b6daffaed7b6c - migrated to `androidx` - 74de9db550ff941083448c28c48b6daffaed7b6c, fixes #130 **Things to do**: - [x] bump library version - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` after Maven Sync - [x] create new GitHub release
1.0
Release 2.2.2-rx2 - **Release notes**: - updated project dependencies - 74de9db550ff941083448c28c48b6daffaed7b6c - migrated to `androidx` - 74de9db550ff941083448c28c48b6daffaed7b6c, fixes #130 **Things to do**: - [x] bump library version - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` after Maven Sync - [x] create new GitHub release
process
release release notes updated project dependencies migrated to androidx fixes things to do bump library version upload archives to maven central repository close and release artifact on nexus update changelog md after maven sync bump library version in readme md after maven sync create new github release
1
63,882
18,041,999,766
IssuesEvent
2021-09-18 07:35:15
mozilla-lockwise/lockwise-android
https://api.github.com/repos/mozilla-lockwise/lockwise-android
opened
Doesn't open when two Firefox are installed
defect
## Steps to reproduce Install Firefox Beta and Stable in the same device. Log in each with a different Firefox accounts. Use fingerprint to unlock Lovkwise (it might not be necessary). ### Expected behavior Allow choose account to use with Lockwise or show at least the passwords of one account (probably Stable would have sense). ### Actual behavior After authentication with finger print, the vault stays lock and keep asking for fingerprint. ### Device & build information * Device: Motorola "moto g⁸ power (XT2041-3)" * Build version: Android 11, 1st August 2021, kernel 4.14.180-perf+, build RPES31.Q4U-47-35-8 ### Notes Attachments:
1.0
Doesn't open when two Firefox are installed - ## Steps to reproduce Install Firefox Beta and Stable in the same device. Log in each with a different Firefox accounts. Use fingerprint to unlock Lovkwise (it might not be necessary). ### Expected behavior Allow choose account to use with Lockwise or show at least the passwords of one account (probably Stable would have sense). ### Actual behavior After authentication with finger print, the vault stays lock and keep asking for fingerprint. ### Device & build information * Device: Motorola "moto g⁸ power (XT2041-3)" * Build version: Android 11, 1st August 2021, kernel 4.14.180-perf+, build RPES31.Q4U-47-35-8 ### Notes Attachments:
non_process
doesn t open when two firefox are installed steps to reproduce install firefox beta and stable in the same device log in each with a different firefox accounts use fingerprint to unlock lovkwise it might not be necessary expected behavior allow choose account to use with lockwise or show at least the passwords of one account probably stable would have sense actual behavior after authentication with finger print the vault stays lock and keep asking for fingerprint device build information device motorola moto g⁸ power build version android august kernel perf build notes attachments
0
195,883
22,362,799,005
IssuesEvent
2022-06-15 22:38:23
snowflakedb/snowflake-hive-metastore-connector
https://api.github.com/repos/snowflakedb/snowflake-hive-metastore-connector
closed
CVE-2020-24750 (High) detected in jackson-databind-2.6.5.jar
security vulnerability
## CVE-2020-24750 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.5/jackson-databind-2.6.5.jar</p> <p> Dependency Hierarchy: - hive-metastore-2.3.5.jar (Root Library) - hive-serde-2.3.5.jar - hive-common-2.3.5.jar - :x: **jackson-databind-2.6.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowflakedb/snowflake-hive-metastore-connector/commit/37f5b0ac91898ef82cc1bf4610b729970f6eed58">37f5b0ac91898ef82cc1bf4610b729970f6eed58</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration. <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-09-17</p> <p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.6.7.4</p> <p>Direct dependency fix Resolution (org.apache.hive:hive-metastore): 2.3.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.hive","packageName":"hive-metastore","packageVersion":"2.3.5","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.hive:hive-metastore:2.3.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.6","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24750","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-24750 (High) detected in jackson-databind-2.6.5.jar - ## CVE-2020-24750 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.5/jackson-databind-2.6.5.jar</p> <p> Dependency Hierarchy: - hive-metastore-2.3.5.jar (Root Library) - hive-serde-2.3.5.jar - hive-common-2.3.5.jar - :x: **jackson-databind-2.6.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowflakedb/snowflake-hive-metastore-connector/commit/37f5b0ac91898ef82cc1bf4610b729970f6eed58">37f5b0ac91898ef82cc1bf4610b729970f6eed58</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration. <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-09-17</p> <p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.6.7.4</p> <p>Direct dependency fix Resolution (org.apache.hive:hive-metastore): 2.3.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.hive","packageName":"hive-metastore","packageVersion":"2.3.5","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.hive:hive-metastore:2.3.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.6","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24750","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy hive metastore jar root library hive serde jar hive common jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com pastdev httpcomponents configuration jndiconfiguration publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution org apache hive hive metastore rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache hive hive metastore isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com pastdev httpcomponents configuration jndiconfiguration vulnerabilityurl
0
220,034
24,549,694,556
IssuesEvent
2022-10-12 11:37:21
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Security Solution]: Text not aligned with node icons under analyzer.
bug impact:low Team:Threat Hunting Team: SecuritySolution Team:Threat Hunting:Investigations v8.5.0
**Describe the bug:** Text not aligned with node icon under analyzer. **Build Details:** ``` VERSION: 8.5.0-SNAPSHOT BUILD: 55925 COMMIT: dc43193d73c5869335a239c7012528bb1fffd509 ``` **Preconditions** 1. Kibana should be running. **Steps to Reproduce** 1. Navigate to Alerts tab. 2. Now, select an alert where analyzer event is available. 3. Open the analyzer and click on node legend icon. 4. Observe that text not aligned with node icons under analyzer. **Actual Result** Text not aligned with node icons under analyzer. **Expected Result** Text should be aligned with node icons under analyzer. **Screenshot:** ![Screenshot (918)](https://user-images.githubusercontent.com/84007952/187882023-59eff206-8a09-4931-94bf-63067c5fe653.png)
True
[Security Solution]: Text not aligned with node icons under analyzer. - **Describe the bug:** Text not aligned with node icon under analyzer. **Build Details:** ``` VERSION: 8.5.0-SNAPSHOT BUILD: 55925 COMMIT: dc43193d73c5869335a239c7012528bb1fffd509 ``` **Preconditions** 1. Kibana should be running. **Steps to Reproduce** 1. Navigate to Alerts tab. 2. Now, select an alert where analyzer event is available. 3. Open the analyzer and click on node legend icon. 4. Observe that text not aligned with node icons under analyzer. **Actual Result** Text not aligned with node icons under analyzer. **Expected Result** Text should be aligned with node icons under analyzer. **Screenshot:** ![Screenshot (918)](https://user-images.githubusercontent.com/84007952/187882023-59eff206-8a09-4931-94bf-63067c5fe653.png)
non_process
text not aligned with node icons under analyzer describe the bug text not aligned with node icon under analyzer build details version snapshot build commit preconditions kibana should be running steps to reproduce navigate to alerts tab now select an alert where analyzer event is available open the analyzer and click on node legend icon observe that text not aligned with node icons under analyzer actual result text not aligned with node icons under analyzer expected result text should be aligned with node icons under analyzer screenshot
0
17,529
23,341,099,823
IssuesEvent
2022-08-09 14:05:46
VolpeUSDOT/FHWA-DANATool
https://api.github.com/repos/VolpeUSDOT/FHWA-DANATool
closed
Geographic intersection for tier 1 matching is inaccurate
bug Process Improvements Medium Priority For Next Version
There is a deepish dive we could do for the Tier 1 matches to make sure it is working as expected and as desired. It may be matching multiple links at the same time in some cases, and there may be other things that could be improved. The geographic matching of TMAS count stations to NPMRDS TMCs is inaccurate because it tries to use .2 degrees, which is a different profile in different parts of the globe. It is distorted the further you get from the equator. So we should use a different CRS to convert to some number of miles instead, so we get a more accurate matching function.
1.0
Geographic intersection for tier 1 matching is inaccurate - There is a deepish dive we could do for the Tier 1 matches to make sure it is working as expected and as desired. It may be matching multiple links at the same time in some cases, and there may be other things that could be improved. The geographic matching of TMAS count stations to NPMRDS TMCs is inaccurate because it tries to use .2 degrees, which is a different profile in different parts of the globe. It is distorted the further you get from the equator. So we should use a different CRS to convert to some number of miles instead, so we get a more accurate matching function.
process
geographic intersection for tier matching is inaccurate there is a deepish dive we could do for the tier matches to make sure it is working as expected and as desired it may be matching multiple links at the same time in some cases and there may be other things that could be improved the geographic matching of tmas count stations to npmrds tmcs is inaccurate because it tries to use degrees which is a different profile in different parts of the globe it is distorted the further you get from the equator so we should use a different crs to convert to some number of miles instead so we get a more accurate matching function
1
623,703
19,676,611,151
IssuesEvent
2022-01-11 13:01:42
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Same token should not be listed for both Swap from and to
bug priority/P3 QA/Yes release-notes/exclude feature/wallet OS/Android
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!-- Provide a brief description of the issue --> Same token should not be listed for both Swap from and to ## Steps to reproduce <!-- Please add a series of steps to reproduce the issue --> 1. Create wallet 2. Select Swap 3. Able to set same token for both swap from and to ## Actual result <!-- Please add screenshots if needed --> https://user-images.githubusercontent.com/17010094/145383433-4a0981b3-2a92-43d8-8077-bf7faa33de35.mp4 ## Expected result Similar to Desktop and iOS, token that is set in Swap from should not be listed in Swap to list ## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] --> Easy ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current Play Store version? Yes when released - Can you reproduce this issue with the current Play Store Beta version? Yes - Can you reproduce this issue with the current Play Store Nightly version? Yes ## Device details - Install type (ARM, x86): ARM - Device type (Phone, Tablet, Phablet): All - Android version: All ## Brave version 1.33.103 ### Website problems only - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ### Additional information <!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue --> cc: @SergeyZhukovsky
1.0
Same token should not be listed for both Swap from and to - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!-- Provide a brief description of the issue --> Same token should not be listed for both Swap from and to ## Steps to reproduce <!-- Please add a series of steps to reproduce the issue --> 1. Create wallet 2. Select Swap 3. Able to set same token for both swap from and to ## Actual result <!-- Please add screenshots if needed --> https://user-images.githubusercontent.com/17010094/145383433-4a0981b3-2a92-43d8-8077-bf7faa33de35.mp4 ## Expected result Similar to Desktop and iOS, token that is set in Swap from should not be listed in Swap to list ## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] --> Easy ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current Play Store version? Yes when released - Can you reproduce this issue with the current Play Store Beta version? Yes - Can you reproduce this issue with the current Play Store Nightly version? Yes ## Device details - Install type (ARM, x86): ARM - Device type (Phone, Tablet, Phablet): All - Android version: All ## Brave version 1.33.103 ### Website problems only - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ### Additional information <!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue --> cc: @SergeyZhukovsky
non_process
same token should not be listed for both swap from and to have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description same token should not be listed for both swap from and to steps to reproduce create wallet select swap able to set same token for both swap from and to actual result expected result similar to desktop and ios token that is set in swap from should not be listed in swap to list issue reproduces how often easy version channel information can you reproduce this issue with the current play store version yes when released can you reproduce this issue with the current play store beta version yes can you reproduce this issue with the current play store nightly version yes device details install type arm arm device type phone tablet phablet all android version all brave version website problems only does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na additional information cc sergeyzhukovsky
0
18,945
24,904,757,604
IssuesEvent
2022-10-29 05:01:51
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
MIPS EABI32 support
Feature: Processor/MIPS
**Is your feature request related to a problem? Please describe.** Ghidra does not support the EABI32 ABI for MIPS. MIPS has supported a fairly large variety of ABIs over the years [[1]](https://en.wikipedia.org/wiki/MIPS_architecture#Calling_conventions). Ghidra should not necessarily be expected to support every ABI out there. Having said that EABI stands out because it is the only MIPS ABI currently supported by GCC that Ghidra does not support. [[2]](https://sourceware.org/gdb/current/onlinedocs/gdb/MIPS.html) Additionally, the commercial disassembler has support for EABI32. **Describe the solution you'd like** I think that an update to the processor spec with a new ABI would solve the issue. That looks like a new cspec file (and other surrounding code). **Describe alternatives you've considered** I initially wrote up an example that allowed an "__eabicall" calling convention. This isn't a good solution because it's not a calling convention, but an ABI. **Additional context** I set up a [repo](https://github.com/lacraig2/Ghidra_MIPS_EABI32_Example) that has source and an example binary that demonstrate the issue. It contains: - Minimal source code that shows the ABI issue - A makefile that works with `mips-linux-gnu-gcc` (default from ubuntu apt) - An example binary As another example see the below function. It should have five arguments. Instead it shows 4 and one argument "in t0", which should be the 5th argument. <img width="509" alt="Screen Shot 2021-11-15 at 4 26 41 PM" src="https://user-images.githubusercontent.com/7518336/141856096-839495ea-6f00-4a34-a003-011a05e01493.png"> There is, unfortunately, not a lot of documentation on the EABI32 format. You can get context from either the GCC source or this one [mailing list post](https://sourceware.org/legacy-ml/binutils/2003-06/msg00436.html). *UPDATE:* I have a PR for review.
1.0
MIPS EABI32 support - **Is your feature request related to a problem? Please describe.** Ghidra does not support the EABI32 ABI for MIPS. MIPS has supported a fairly large variety of ABIs over the years [[1]](https://en.wikipedia.org/wiki/MIPS_architecture#Calling_conventions). Ghidra should not necessarily be expected to support every ABI out there. Having said that EABI stands out because it is the only MIPS ABI currently supported by GCC that Ghidra does not support. [[2]](https://sourceware.org/gdb/current/onlinedocs/gdb/MIPS.html) Additionally, the commercial disassembler has support for EABI32. **Describe the solution you'd like** I think that an update to the processor spec with a new ABI would solve the issue. That looks like a new cspec file (and other surrounding code). **Describe alternatives you've considered** I initially wrote up an example that allowed an "__eabicall" calling convention. This isn't a good solution because it's not a calling convention, but an ABI. **Additional context** I set up a [repo](https://github.com/lacraig2/Ghidra_MIPS_EABI32_Example) that has source and an example binary that demonstrate the issue. It contains: - Minimal source code that shows the ABI issue - A makefile that works with `mips-linux-gnu-gcc` (default from ubuntu apt) - An example binary As another example see the below function. It should have five arguments. Instead it shows 4 and one argument "in t0", which should be the 5th argument. <img width="509" alt="Screen Shot 2021-11-15 at 4 26 41 PM" src="https://user-images.githubusercontent.com/7518336/141856096-839495ea-6f00-4a34-a003-011a05e01493.png"> There is, unfortunately, not a lot of documentation on the EABI32 format. You can get context from either the GCC source or this one [mailing list post](https://sourceware.org/legacy-ml/binutils/2003-06/msg00436.html). *UPDATE:* I have a PR for review.
process
mips support is your feature request related to a problem please describe ghidra does not support the abi for mips mips has supported a fairly large variety of abis over the years ghidra should not necessarily be expected to support every abi out there having said that eabi stands out because it is the only mips abi currently supported by gcc that ghidra does not support additionally the commercial disassembler has support for describe the solution you d like i think that an update to the processor spec with a new abi would solve the issue that looks like a new cspec file and other surrounding code describe alternatives you ve considered i initially wrote up an example that allowed an eabicall calling convention this isn t a good solution because it s not a calling convention but an abi additional context i set up a that has source and an example binary that demonstrate the issue it contains minimal source code that shows the abi issue a makefile that works with mips linux gnu gcc default from ubuntu apt an example binary as another example see the below function it should have five arguments instead it shows and one argument in which should be the argument img width alt screen shot at pm src there is unfortunately not a lot of documentation on the format you can get context from either the gcc source or this one update i have a pr for review
1
57,331
15,730,654,969
IssuesEvent
2021-03-29 16:09:27
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
Hung task related to vdev_trim_calculate_progress()
Status: Triage Needed Type: Defect
### System information Distribution Name | NixOS Distribution Version | 20.03 Linux Kernel | 5.4.39 Architecture | x86_64 ZFS Version | 0.8.3-1 SPL Version | 0.8.3-1 ### Describe the problem you're observing This happened to me twice in last 12 months. I'm using SSDs that might handle trim requests poorly, but that's speculation. Model: ADATA SP900. "zpool import -fFn <pool>" gets me: ``` [178.250374] VERIFY3(0 == metaslab_load(msp)) failed (0 == 52) [178.250418] PANIC at vdev_trim.c:612:vdev_trim_calculate_progress() ``` The pool *can* be imported with "-o readonly=on". That's how I recovered the last time it happened, about ~8 months ago. Afterwards, I'm getting: ``` INFO: task txg_sync:1992 blocked for more than <N> seconds. ``` ### Describe how to reproduce the problem No clue. Both times it happened while waking up from hibernation, after a scrub was completed. I noticed that now autotrim is enabled on this pool again, even though I disabled it the last time this kind of thing happened. I suspect this might've happened during an OS update. It might be that a trim was scheduled on the same day the scrub was.
1.0
Hung task related to vdev_trim_calculate_progress() - ### System information Distribution Name | NixOS Distribution Version | 20.03 Linux Kernel | 5.4.39 Architecture | x86_64 ZFS Version | 0.8.3-1 SPL Version | 0.8.3-1 ### Describe the problem you're observing This happened to me twice in last 12 months. I'm using SSDs that might handle trim requests poorly, but that's speculation. Model: ADATA SP900. "zpool import -fFn <pool>" gets me: ``` [178.250374] VERIFY3(0 == metaslab_load(msp)) failed (0 == 52) [178.250418] PANIC at vdev_trim.c:612:vdev_trim_calculate_progress() ``` The pool *can* be imported with "-o readonly=on". That's how I recovered the last time it happened, about ~8 months ago. Afterwards, I'm getting: ``` INFO: task txg_sync:1992 blocked for more than <N> seconds. ``` ### Describe how to reproduce the problem No clue. Both times it happened while waking up from hibernation, after a scrub was completed. I noticed that now autotrim is enabled on this pool again, even though I disabled it the last time this kind of thing happened. I suspect this might've happened during an OS update. It might be that a trim was scheduled on the same day the scrub was.
non_process
hung task related to vdev trim calculate progress system information distribution name nixos distribution version linux kernel architecture zfs version spl version describe the problem you re observing this happened to me twice in last months i m using ssds that might handle trim requests poorly but that s speculation model adata zpool import ffn gets me metaslab load msp failed panic at vdev trim c vdev trim calculate progress the pool can be imported with o readonly on that s how i recovered the last time it happened about months ago afterwards i m getting info task txg sync blocked for more than seconds describe how to reproduce the problem no clue both times it happened while waking up from hibernation after a scrub was completed i noticed that now autotrim is enabled on this pool again even though i disabled it the last time this kind of thing happened i suspect this might ve happened during an os update it might be that a trim was scheduled on the same day the scrub was
0
21,697
30,194,945,692
IssuesEvent
2023-07-04 19:40:08
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Add shorter display name like `≠` to display info for filter operators and change existing `:display-name` to `:long-display-name`
.Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
In places like joins we actually show just the operator in the UI rather than a nice long display name for filter clause, e.g. ![image](https://github.com/metabase/metabase/assets/1455846/4b31e57b-838a-494e-bfd8-a298fcff9682) ![image](https://github.com/metabase/metabase/assets/1455846/c65bf2e6-f5cf-4f34-a90c-23c4b2db380e) Note the fancy unicode characters like `≠`. We need to return these in the **display info** for `FilterOperator` maps so the FE can use them when we port stuff to MLv2. MLv1 definitions live here https://github.com/metabase/metabase/blob/287c5204d30dd36d1c61656da792ae1f6dc7d750/frontend/src/metabase/query_builder/components/notebook/steps/JoinStep/JoinStep.jsx#L78 I believe these are currently only used for joins, so we technically only need to add them for the 6 operators we return for joins, but we might as well return them for everything; I think for everything besides `:!=`, `:>=`, and `:<=`, it's the same as the tag itself. What should we call this key? I was going to suggest `:short-display-name`, but it's a bit silly to have the display info for filter operators have `:short-display-name` and `:display-name` while everything else has `:display-name` and `:long-display-name`... maybe what we should do here is change the existing `:display-name` to `:long-display-name` and make `≠` the new `:display-name`.
1.0
[MLv2] Add shorter display name like `≠` to display info for filter operators and change existing `:display-name` to `:long-display-name` - In places like joins we actually show just the operator in the UI rather than a nice long display name for filter clause, e.g. ![image](https://github.com/metabase/metabase/assets/1455846/4b31e57b-838a-494e-bfd8-a298fcff9682) ![image](https://github.com/metabase/metabase/assets/1455846/c65bf2e6-f5cf-4f34-a90c-23c4b2db380e) Note the fancy unicode characters like `≠`. We need to return these in the **display info** for `FilterOperator` maps so the FE can use them when we port stuff to MLv2. MLv1 definitions live here https://github.com/metabase/metabase/blob/287c5204d30dd36d1c61656da792ae1f6dc7d750/frontend/src/metabase/query_builder/components/notebook/steps/JoinStep/JoinStep.jsx#L78 I believe these are currently only used for joins, so we technically only need to add them for the 6 operators we return for joins, but we might as well return them for everything; I think for everything besides `:!=`, `:>=`, and `:<=`, it's the same as the tag itself. What should we call this key? I was going to suggest `:short-display-name`, but it's a bit silly to have the display info for filter operators have `:short-display-name` and `:display-name` while everything else has `:display-name` and `:long-display-name`... maybe what we should do here is change the existing `:display-name` to `:long-display-name` and make `≠` the new `:display-name`.
process
add shorter display name like ≠ to display info for filter operators and change existing display name to long display name in places like joins we actually show just the operator in the ui rather than a nice long display name for filter clause e g note the fancy unicode characters like ≠ we need to return these in the display info for filteroperator maps so the fe can use them when we port stuff to definitions live here i believe these are currently only used for joins so we technically only need to add them for the operators we return for joins but we might as well return them for everything i think for everything besides and it s the same as the tag itself what should we call this key i was going to suggest short display name but it s a bit silly to have the display info for filter operators have short display name and display name while everything else has display name and long display name maybe what we should do here is change the existing display name to long display name and make ≠ the new display name
1
19,394
25,537,736,353
IssuesEvent
2022-11-29 13:15:35
eosnetworkfoundation/devrel
https://api.github.com/repos/eosnetworkfoundation/devrel
closed
Create a staging URL to review current state of docs before publishing
Process
need to run every day need a process doc Note: It's crucial to make sure that `latest` still links to the latest version published on the site, not to the potential new version, if any, which will be tested using the calvary link. @ericpassmore
1.0
Create a staging URL to review current state of docs before publishing - need to run every day need a process doc Note: It's crucial to make sure that `latest` still links to the latest version published on the site, not to the potential new version, if any, which will be tested using the calvary link. @ericpassmore
process
create a staging url to review current state of docs before publishing need to run every day need a process doc note it s crucial to make sure that latest still links to the latest version published on the site not to the potential new version if any which will be tested using the calvary link ericpassmore
1
75,644
14,501,526,300
IssuesEvent
2020-12-11 19:38:32
dragonfruitnetwork/dragon6-api
https://api.github.com/repos/dragonfruitnetwork/dragon6-api
closed
reverse the renaming change
area:code-quality area:requests type:cleanup
Decided to create a new repo for the modern stats as they're so different that they should feel as an optional extension. This should mean that the naming change (#181) should be partially reverted, #185 cherry-picked into a new repo and the dragonfruit.common.data library updated
1.0
reverse the renaming change - Decided to create a new repo for the modern stats as they're so different that they should feel as an optional extension. This should mean that the naming change (#181) should be partially reverted, #185 cherry-picked into a new repo and the dragonfruit.common.data library updated
non_process
reverse the renaming change decided to create a new repo for the modern stats as they re so different that they should feel as an optional extension this should mean that the naming change should be partially reverted cherry picked into a new repo and the dragonfruit common data library updated
0
18,519
24,551,777,105
IssuesEvent
2022-10-12 13:08:53
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] [Offline indicator] There should be a consistency in offline error messages in both enrollment flow and review consent flow
Bug P2 iOS Process: Fixed Process: Tested dev
There should be a consistency in offline error messages in both enrollment flow and review consent flow **Enrollment flow:** ![image](https://user-images.githubusercontent.com/71445210/185437916-e61d5ba6-9f6f-4327-9c2d-82a06afd1c2b.png) **Review updated consent:** ![image](https://user-images.githubusercontent.com/71445210/185438066-b47c5efe-5b84-457b-96b0-4f772ace9cc5.png)
2.0
[iOS] [Offline indicator] There should be a consistency in offline error messages in both enrollment flow and review consent flow - There should be a consistency in offline error messages in both enrollment flow and review consent flow **Enrollment flow:** ![image](https://user-images.githubusercontent.com/71445210/185437916-e61d5ba6-9f6f-4327-9c2d-82a06afd1c2b.png) **Review updated consent:** ![image](https://user-images.githubusercontent.com/71445210/185438066-b47c5efe-5b84-457b-96b0-4f772ace9cc5.png)
process
there should be a consistency in offline error messages in both enrollment flow and review consent flow there should be a consistency in offline error messages in both enrollment flow and review consent flow enrollment flow review updated consent
1
22,390
31,142,286,174
IssuesEvent
2023-08-16 01:44:17
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky test: Too long with no output (exceeded 10m0s): context deadline exceeded
OS: linux process: flaky test topic: flake ❄️ stage: flake stale
### Link to dashboard or CircleCI failure - https://app.circleci.com/pipelines/github/cypress-io/cypress/41757/workflows/edf97bff-eef3-4f3f-8057-4277aef7d046/jobs/1730766 - https://app.circleci.com/pipelines/github/cypress-io/cypress/42240/workflows/b213eedd-524b-4ef3-9170-b15bfc01d1c7/jobs/1753458 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/navigation.cy.js#L495 ### Analysis Tests hang during execution of `navigation.cy.js` file <img width="700" alt="Screen Shot 2022-08-11 at 8 00 17 PM" src="https://user-images.githubusercontent.com/26726429/184276927-80c83a8d-e2f2-4b69-9794-a18255e7060d.png"> We also see this same error as seen in https://github.com/cypress-io/cypress/issues/23326 ``` $ node ../../scripts/cypress run --config experimentalSessionAndOrigin=true,experimentalModifyObstructiveThirdPartyCode=true --record --parallel --group 5x-driver-chrome:beta-experimentalSessionAndOrigin --browser chrome:beta libva error: vaGetDriverNameByIndex() failed with unknown libva error, driver_name = (null) [2032:0811/220636.124358:ERROR:gpu_memory_buffer_support_x11.cc(44)] dri3 extension not supported. ``` ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
1.0
Flaky test: Too long with no output (exceeded 10m0s): context deadline exceeded - ### Link to dashboard or CircleCI failure - https://app.circleci.com/pipelines/github/cypress-io/cypress/41757/workflows/edf97bff-eef3-4f3f-8057-4277aef7d046/jobs/1730766 - https://app.circleci.com/pipelines/github/cypress-io/cypress/42240/workflows/b213eedd-524b-4ef3-9170-b15bfc01d1c7/jobs/1753458 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/navigation.cy.js#L495 ### Analysis Tests hang during execution of `navigation.cy.js` file <img width="700" alt="Screen Shot 2022-08-11 at 8 00 17 PM" src="https://user-images.githubusercontent.com/26726429/184276927-80c83a8d-e2f2-4b69-9794-a18255e7060d.png"> We also see this same error as seen in https://github.com/cypress-io/cypress/issues/23326 ``` $ node ../../scripts/cypress run --config experimentalSessionAndOrigin=true,experimentalModifyObstructiveThirdPartyCode=true --record --parallel --group 5x-driver-chrome:beta-experimentalSessionAndOrigin --browser chrome:beta libva error: vaGetDriverNameByIndex() failed with unknown libva error, driver_name = (null) [2032:0811/220636.124358:ERROR:gpu_memory_buffer_support_x11.cc(44)] dri3 extension not supported. ``` ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
process
flaky test too long with no output exceeded context deadline exceeded link to dashboard or circleci failure link to failing test in github analysis tests hang during execution of navigation cy js file img width alt screen shot at pm src we also see this same error as seen in node scripts cypress run config experimentalsessionandorigin true experimentalmodifyobstructivethirdpartycode true record parallel group driver chrome beta experimentalsessionandorigin browser chrome beta libva error vagetdrivernamebyindex failed with unknown libva error driver name null extension not supported cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
1
543,376
15,880,879,724
IssuesEvent
2021-04-09 14:11:49
ansible/awx
https://api.github.com/repos/ansible/awx
closed
When using AWX on plain HTTP, Social Auth is failing
component:api priority:low state:needs_devel type:bug
##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: --> - API ##### SUMMARY When user AWX without HTTPS (plain HTTP), logging in with the GitHub OAuth button fails. ##### ENVIRONMENT * AWX version: 1.0.1.229 * AWX install method: docker on linux * Ansible version: 2.4.1.0 * Operating System: debian stretch * Web Browser: firefox ##### STEPS TO REPRODUCE * Setup AWX, access through port 80 (default) * Setup GitHub Org auth (but really, it's the same with any Social Auth plugin) * Log in using the social auth ##### EXPECTED RESULTS You're logged-in ##### ACTUAL RESULTS You're not. awx_web has a log saying : ``` 2017-11-28 17:57:48,299 ERROR social Session value state missing. ``` on the `GET /sso/complete/github-org/` call. ##### ADDITIONAL INFORMATION The session to persist the `state` parameter between the initial and the `complete` view of the oauth dance is defined by: https://github.com/ansible/awx/blob/9ed2a0da8f4883a7d6311de2958a01e70605b062/awx/settings/defaults.py#L492 Which uses the default Django session store, which uses a secure cookie due to https://github.com/ansible/awx/blob/9ed2a0da8f4883a7d6311de2958a01e70605b062/awx/settings/defaults.py#L188 So, if we're not using HTTPS, the cookie will not be sent on the second call, so the session will not be retrieved, the `state` value is absent and the oauth dance fails. ##### WORK AROUND I guess either use HTTPS or override the SESSION_COOKIE_SECURE to False.
1.0
When using AWX on plain HTTP, Social Auth is failing - ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: --> - API ##### SUMMARY When user AWX without HTTPS (plain HTTP), logging in with the GitHub OAuth button fails. ##### ENVIRONMENT * AWX version: 1.0.1.229 * AWX install method: docker on linux * Ansible version: 2.4.1.0 * Operating System: debian stretch * Web Browser: firefox ##### STEPS TO REPRODUCE * Setup AWX, access through port 80 (default) * Setup GitHub Org auth (but really, it's the same with any Social Auth plugin) * Log in using the social auth ##### EXPECTED RESULTS You're logged-in ##### ACTUAL RESULTS You're not. awx_web has a log saying : ``` 2017-11-28 17:57:48,299 ERROR social Session value state missing. ``` on the `GET /sso/complete/github-org/` call. ##### ADDITIONAL INFORMATION The session to persist the `state` parameter between the initial and the `complete` view of the oauth dance is defined by: https://github.com/ansible/awx/blob/9ed2a0da8f4883a7d6311de2958a01e70605b062/awx/settings/defaults.py#L492 Which uses the default Django session store, which uses a secure cookie due to https://github.com/ansible/awx/blob/9ed2a0da8f4883a7d6311de2958a01e70605b062/awx/settings/defaults.py#L188 So, if we're not using HTTPS, the cookie will not be sent on the second call, so the session will not be retrieved, the `state` value is absent and the oauth dance fails. ##### WORK AROUND I guess either use HTTPS or override the SESSION_COOKIE_SECURE to False.
non_process
when using awx on plain http social auth is failing issue type bug report component name api summary when user awx without https plain http logging in with the github oauth button fails environment awx version awx install method docker on linux ansible version operating system debian stretch web browser firefox steps to reproduce setup awx access through port default setup github org auth but really it s the same with any social auth plugin log in using the social auth expected results you re logged in actual results you re not awx web has a log saying error social session value state missing on the get sso complete github org call additional information the session to persist the state parameter between the initial and the complete view of the oauth dance is defined by which uses the default django session store which uses a secure cookie due to so if we re not using https the cookie will not be sent on the second call so the session will not be retrieved the state value is absent and the oauth dance fails work around i guess either use https or override the session cookie secure to false
0
224,547
17,192,807,468
IssuesEvent
2021-07-16 13:25:20
MICA-MNI/BrainSpace
https://api.github.com/repos/MICA-MNI/BrainSpace
closed
Subject list for group connectivity matrices
documentation
I was wondering if the list of HCP subjects used for the group functional connectivity matrices with `load_group_fc`/`load_group_mpc` was available. My understanding/assumption is that these subsets of HCP data are based on the selection procedures described in [Vos de Wael et al (2018)](https://www.pnas.org/content/115/40/10154), but I'm not 100% certain in that. I'm looking to compute similar connectivity matrices on the samples used in Vos de Wael et al (2018), but I also think it would be great if the documentation could clarify the source of the group connectivity matrices.
1.0
Subject list for group connectivity matrices - I was wondering if the list of HCP subjects used for the group functional connectivity matrices with `load_group_fc`/`load_group_mpc` was available. My understanding/assumption is that these subsets of HCP data are based on the selection procedures described in [Vos de Wael et al (2018)](https://www.pnas.org/content/115/40/10154), but I'm not 100% certain in that. I'm looking to compute similar connectivity matrices on the samples used in Vos de Wael et al (2018), but I also think it would be great if the documentation could clarify the source of the group connectivity matrices.
non_process
subject list for group connectivity matrices i was wondering if the list of hcp subjects used for the group functional connectivity matrices with load group fc load group mpc was available my understanding assumption is that these subsets of hcp data are based on the selection procedures described in but i m not certain in that i m looking to compute similar connectivity matrices on the samples used in vos de wael et al but i also think it would be great if the documentation could clarify the source of the group connectivity matrices
0
319,674
23,785,043,332
IssuesEvent
2022-09-02 09:17:50
Nicmanthecodeman/CMPG-323-Overview-25102737
https://api.github.com/repos/Nicmanthecodeman/CMPG-323-Overview-25102737
opened
Setup repository
documentation good first issue project high
# Requirements - [ ] GitHub Repository - [ ] Create and Configure GitHub Repository - [ ] Create a repository named \'CMPG 323 Overview - \<add your student number>' <code>Priority: 1</code> - [ ] Create the following Milestones: <code>Priority: 2</code> - [ ] Project 1 Submission: Deadline – 18 August - [ ] Project 2 Submission: Deadline – 8 September - [ ] Project 3 Submission: Deadline – 29 September - [ ] Project 4 Submission: Deadline – 20 October - [ ] Project 5 Submission: Deadline – 3 November - [ ] Exam (POE) Submission: Deadline – 18 November - [ ] Create the following Labels: <code>Priority: 2</code> - [ ] Class - [ ] Project - [ ] Training - [ ] \<any other labels you might find relevant> - [ ] Add a ReadME.md file to the repository - [ ] In the ReadME.md, address which repositories will be created and used for each project <code>Priority: 1</code> - [ ] In the ReadME.md, provide a diagram explaining project and repository context and how they are integrated <code>Priority: 3</code> - [ ] In the ReadME.md, explain the branching strategy to be used within each project <code>Priority: 3</code> - [ ] In the ReadME.md, explain the use of a .gitignore file within each project <code>Priority: 3</code> - [ ] In the ReadME.md, explain the storage of credentials and sensitive information <code>Priority: 3</code> - [ ] GitHub Project - [ ] Create and Configure GitHub project - [ ] Create a GitHub Kanban project <code>Priority: 1</code> - [ ] Link the GitHub repository to the project <code>Priority: 2</code> - [ ] Ensure that the project is named appropriately <code>Priority: 3</code> - [ ] Ensure that the project has a description adequately describing the project <code>Priority: 3</code> - [ ] Add the following columns: <code>Priority: 2</code> - [ ] Linked Assessment (field type = text) - [ ] Due Date (field type = date) - [ ] Sprint (field type = single select; add all 8 of the sprints as options with sprint start and end date) - [ ] Effort (in Hours) (field type = number) - [ ] Create and populate project views - [ ] Create a table view within the project named ‘Tabular View’ <code>Priority: 2</code> - [ ] Add Milestone, Label, Linked Pull Request and Repository as fields to the view - [ ] Hide the Assignees field - [ ] Create a board view within the project, named ‘Status View’<code>Priority: 2</code> - [ ] Group by Status - [ ] Create a board view within the project, named ‘Sprint View’<code>Priority: 2</code> - [ ] Group by Sprint - [ ] Create a table view within the project, named ‘Linked Assessment View’<code>Priority: 3</code> - [ ] Group by Linked Assessment - [ ] Create an additional customised view that provides you with additional insight into your semester progress<code>Priority: 4</code>
1.0
Setup repository - # Requirements - [ ] GitHub Repository - [ ] Create and Configure GitHub Repository - [ ] Create a repository named \'CMPG 323 Overview - \<add your student number>' <code>Priority: 1</code> - [ ] Create the following Milestones: <code>Priority: 2</code> - [ ] Project 1 Submission: Deadline – 18 August - [ ] Project 2 Submission: Deadline – 8 September - [ ] Project 3 Submission: Deadline – 29 September - [ ] Project 4 Submission: Deadline – 20 October - [ ] Project 5 Submission: Deadline – 3 November - [ ] Exam (POE) Submission: Deadline – 18 November - [ ] Create the following Labels: <code>Priority: 2</code> - [ ] Class - [ ] Project - [ ] Training - [ ] \<any other labels you might find relevant> - [ ] Add a ReadME.md file to the repository - [ ] In the ReadME.md, address which repositories will be created and used for each project <code>Priority: 1</code> - [ ] In the ReadME.md, provide a diagram explaining project and repository context and how they are integrated <code>Priority: 3</code> - [ ] In the ReadME.md, explain the branching strategy to be used within each project <code>Priority: 3</code> - [ ] In the ReadME.md, explain the use of a .gitignore file within each project <code>Priority: 3</code> - [ ] In the ReadME.md, explain the storage of credentials and sensitive information <code>Priority: 3</code> - [ ] GitHub Project - [ ] Create and Configure GitHub project - [ ] Create a GitHub Kanban project <code>Priority: 1</code> - [ ] Link the GitHub repository to the project <code>Priority: 2</code> - [ ] Ensure that the project is named appropriately <code>Priority: 3</code> - [ ] Ensure that the project has a description adequately describing the project <code>Priority: 3</code> - [ ] Add the following columns: <code>Priority: 2</code> - [ ] Linked Assessment (field type = text) - [ ] Due Date (field type = date) - [ ] Sprint (field type = single select; add all 8 of the sprints as options with sprint start and end date) - [ ] Effort (in Hours) (field type = number) - [ ] Create and populate project views - [ ] Create a table view within the project named ‘Tabular View’ <code>Priority: 2</code> - [ ] Add Milestone, Label, Linked Pull Request and Repository as fields to the view - [ ] Hide the Assignees field - [ ] Create a board view within the project, named ‘Status View’<code>Priority: 2</code> - [ ] Group by Status - [ ] Create a board view within the project, named ‘Sprint View’<code>Priority: 2</code> - [ ] Group by Sprint - [ ] Create a table view within the project, named ‘Linked Assessment View’<code>Priority: 3</code> - [ ] Group by Linked Assessment - [ ] Create an additional customised view that provides you with additional insight into your semester progress<code>Priority: 4</code>
non_process
setup repository requirements github repository create and configure github repository create a repository named cmpg overview priority create the following milestones priority project submission deadline – august project submission deadline – september project submission deadline – september project submission deadline – october project submission deadline – november exam poe submission deadline – november create the following labels priority class project training add a readme md file to the repository in the readme md address which repositories will be created and used for each project priority in the readme md provide a diagram explaining project and repository context and how they are integrated priority in the readme md explain the branching strategy to be used within each project priority in the readme md explain the use of a gitignore file within each project priority in the readme md explain the storage of credentials and sensitive information priority github project create and configure github project create a github kanban project priority link the github repository to the project priority ensure that the project is named appropriately priority ensure that the project has a description adequately describing the project priority add the following columns priority linked assessment field type text due date field type date sprint field type single select add all of the sprints as options with sprint start and end date effort in hours field type number create and populate project views create a table view within the project named ‘tabular view’ priority add milestone label linked pull request and repository as fields to the view hide the assignees field create a board view within the project named ‘status view’ priority group by status create a board view within the project named ‘sprint view’ priority group by sprint create a table view within the project named ‘linked assessment view’ priority group by linked assessment create an additional customised view that provides you with additional insight into your semester progress priority
0
16,883
22,162,690,253
IssuesEvent
2022-06-04 18:46:43
RobertCraigie/prisma-client-py
https://api.github.com/repos/RobertCraigie/prisma-client-py
closed
Prisma CLI and Client use different base paths for relative SQLite files
bug/2-confirmed kind/bug process/candidate topic: client level/intermediate priority/high
<!-- Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output. See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output. --> ## Bug description <!-- A clear and concise description of what the bug is. --> With this schema as `prisma/schema.prisma`: ```prisma datasource db { provider = "sqlite" url = "file:dev.db" } model User { id String @id @default(cuid()) name String } ``` Running `prisma db push` creates a database file at `prisma/dev.db`. Now running the client will error as the query engine uses the path relative to the working directory instead of the schema directory, e.g. `main.py` ```python import asyncio from prisma import Prisma from prisma.models import User async def bar() -> None: p = Prisma(auto_register=True) await p.connect() await User.prisma().create(data={'name': 'Robert'}) asyncio.run(bar()) ``` Will use `dev.db` instead of `prisma/dev.db`. ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> The CLI and the Client should use the same database path.
1.0
Prisma CLI and Client use different base paths for relative SQLite files - <!-- Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output. See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output. --> ## Bug description <!-- A clear and concise description of what the bug is. --> With this schema as `prisma/schema.prisma`: ```prisma datasource db { provider = "sqlite" url = "file:dev.db" } model User { id String @id @default(cuid()) name String } ``` Running `prisma db push` creates a database file at `prisma/dev.db`. Now running the client will error as the query engine uses the path relative to the working directory instead of the schema directory, e.g. `main.py` ```python import asyncio from prisma import Prisma from prisma.models import User async def bar() -> None: p = Prisma(auto_register=True) await p.connect() await User.prisma().create(data={'name': 'Robert'}) asyncio.run(bar()) ``` Will use `dev.db` instead of `prisma/dev.db`. ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> The CLI and the Client should use the same database path.
process
prisma cli and client use different base paths for relative sqlite files thanks for helping us improve prisma client python 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by enabling additional logging output see for how to enable additional logging output bug description with this schema as prisma schema prisma prisma datasource db provider sqlite url file dev db model user id string id default cuid name string running prisma db push creates a database file at prisma dev db now running the client will error as the query engine uses the path relative to the working directory instead of the schema directory e g main py python import asyncio from prisma import prisma from prisma models import user async def bar none p prisma auto register true await p connect await user prisma create data name robert asyncio run bar will use dev db instead of prisma dev db expected behavior the cli and the client should use the same database path
1
20,131
26,669,538,478
IssuesEvent
2023-01-26 09:02:10
googleapis/google-cloud-php
https://api.github.com/repos/googleapis/google-cloud-php
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'longrunning' invalid in LongRunning/.repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'longrunning' invalid in LongRunning/.repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname longrunning invalid in longrunning repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
17,077
2,974,593,164
IssuesEvent
2015-07-15 02:10:24
Reimashi/jotai
https://api.github.com/repos/Reimashi/jotai
closed
System is set for F° yet when placed in tray it shows C° it does show F° in gadget box
auto-migrated Priority-Medium Type-Defect wontfix
``` Using windows XP SP3 ``` Original issue reported on code.google.com by `LEGALVID...@aol.com` on 7 May 2012 at 1:26
1.0
System is set for F° yet when placed in tray it shows C° it does show F° in gadget box - ``` Using windows XP SP3 ``` Original issue reported on code.google.com by `LEGALVID...@aol.com` on 7 May 2012 at 1:26
non_process
system is set for f° yet when placed in tray it shows c° it does show f° in gadget box using windows xp original issue reported on code google com by legalvid aol com on may at
0
9,307
12,321,395,489
IssuesEvent
2020-05-13 08:39:15
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Step 6 The command is not displayed even if you enter "Connect-AzAccount"
Pri1 automation/svc cxp process-automation/subsvc product-question triaged
The comments below were machine-translated @kurihiro0119 commented on [Wed May 06 2020](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/4528) Follow step 6 and enter the form on the left that says Search Library Items Nothing is displayed when I enter "Connect-AzAccount". Therefore, the corresponding command cannot be set and the tutorial ends halfway. --- #### ドキュメントの詳細 ⚠ *このセクションを編集しないでください。 docs.microsoft.com で必須です ➟ GitHub の問題のリンク。* * ID: 4799ebe4-d654-b9d2-e64f-923d025ad7cf * Version Independent ID: f6519afb-bf55-c886-e3c3-d11174da3aea * Content: [Azure Automation でグラフィカル Runbook を作成する](https://docs.microsoft.com/ja-jp/azure/automation/learn/automation-tutorial-runbook-graphical#feedback) * Content Source: [articles/automation/learn/automation-tutorial-runbook-graphical.md](https://github.com/MicrosoftDocs/azure-docs.ja-jp/blob/master/articles/automation/learn/automation-tutorial-runbook-graphical.md) * Sub-service: **process-automation** * GitHub Login: @olprod --- @srvbpigh commented on [Wed May 06 2020](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/4528#issuecomment-624614712) Hello, @kurihiro0119 Thank you for your feedback. We are actively reviewing your comments and will get back to you soon. Kind regards, Microsoft DOCS International Team
1.0
Step 6 The command is not displayed even if you enter "Connect-AzAccount" - The comments below were machine-translated @kurihiro0119 commented on [Wed May 06 2020](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/4528) Follow step 6 and enter the form on the left that says Search Library Items Nothing is displayed when I enter "Connect-AzAccount". Therefore, the corresponding command cannot be set and the tutorial ends halfway. --- #### ドキュメントの詳細 ⚠ *このセクションを編集しないでください。 docs.microsoft.com で必須です ➟ GitHub の問題のリンク。* * ID: 4799ebe4-d654-b9d2-e64f-923d025ad7cf * Version Independent ID: f6519afb-bf55-c886-e3c3-d11174da3aea * Content: [Azure Automation でグラフィカル Runbook を作成する](https://docs.microsoft.com/ja-jp/azure/automation/learn/automation-tutorial-runbook-graphical#feedback) * Content Source: [articles/automation/learn/automation-tutorial-runbook-graphical.md](https://github.com/MicrosoftDocs/azure-docs.ja-jp/blob/master/articles/automation/learn/automation-tutorial-runbook-graphical.md) * Sub-service: **process-automation** * GitHub Login: @olprod --- @srvbpigh commented on [Wed May 06 2020](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/4528#issuecomment-624614712) Hello, @kurihiro0119 Thank you for your feedback. We are actively reviewing your comments and will get back to you soon. Kind regards, Microsoft DOCS International Team
process
step the command is not displayed even if you enter connect azaccount the comments below were machine translated commented on follow step and enter the form on the left that says search library items nothing is displayed when i enter connect azaccount therefore the corresponding command cannot be set and the tutorial ends halfway ドキュメントの詳細 ⚠ このセクションを編集しないでください。 docs microsoft com で必須です ➟ github の問題のリンク。 id version independent id content content source sub service process automation github login olprod srvbpigh commented on hello thank you for your feedback we are actively reviewing your comments and will get back to you soon kind regards microsoft docs international team
1