Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
12,896
| 15,284,184,903
|
IssuesEvent
|
2021-02-23 11:53:16
|
Bedrohung-der-Bienen/Transformationsfelder-Digitalisierung
|
https://api.github.com/repos/Bedrohung-der-Bienen/Transformationsfelder-Digitalisierung
|
opened
|
Benutzer in der Datenbank anlegen
|
backburner backend login process register process
|
Es muss in der Datenbank die Tabelle Benutzer erstellt werden mit ihren Attributen
|
2.0
|
Benutzer in der Datenbank anlegen - Es muss in der Datenbank die Tabelle Benutzer erstellt werden mit ihren Attributen
|
process
|
benutzer in der datenbank anlegen es muss in der datenbank die tabelle benutzer erstellt werden mit ihren attributen
| 1
|
21,345
| 11,198,425,054
|
IssuesEvent
|
2020-01-03 15:54:59
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
closed
|
File low-hanging fruit of HomeActivity creation
|
P3 eng:performance
|
Let's file actionable bugs to address the low-hanging fruit of so we can 1) prioritize them & ensure we're working on the most important things and 2) ensure our backlog has actionable works for devs.
### Acceptance criteria
- File bugs to address the low-hanging fruit (and some larger diminishing returns) in `HomeActivity` creation
- Notify Slack channel when issues are filed
---
Overall startup time is around 300ms (at this point and on a specific device). `HomeActivity.onCreate` takes 307ms. Layout inflation alone takes 126ms. `setupToolbarAndNavigation` takes 70ms.
@colintheshots @csadilek can you tag appropriately? Thanks!
|
True
|
File low-hanging fruit of HomeActivity creation - Let's file actionable bugs to address the low-hanging fruit of so we can 1) prioritize them & ensure we're working on the most important things and 2) ensure our backlog has actionable works for devs.
### Acceptance criteria
- File bugs to address the low-hanging fruit (and some larger diminishing returns) in `HomeActivity` creation
- Notify Slack channel when issues are filed
---
Overall startup time is around 300ms (at this point and on a specific device). `HomeActivity.onCreate` takes 307ms. Layout inflation alone takes 126ms. `setupToolbarAndNavigation` takes 70ms.
@colintheshots @csadilek can you tag appropriately? Thanks!
|
non_process
|
file low hanging fruit of homeactivity creation let s file actionable bugs to address the low hanging fruit of so we can prioritize them ensure we re working on the most important things and ensure our backlog has actionable works for devs acceptance criteria file bugs to address the low hanging fruit and some larger diminishing returns in homeactivity creation notify slack channel when issues are filed overall startup time is around at this point and on a specific device homeactivity oncreate takes layout inflation alone takes setuptoolbarandnavigation takes colintheshots csadilek can you tag appropriately thanks
| 0
|
209,804
| 23,730,845,641
|
IssuesEvent
|
2022-08-31 01:27:39
|
vipinsun/blockchain-carbon-accounting
|
https://api.github.com/repos/vipinsun/blockchain-carbon-accounting
|
closed
|
CVE-2020-11023 (Medium) detected in multiple libraries - autoclosed
|
security vulnerability
|
## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-3.2.1.min.js</b>, <b>jquery-1.7.2.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-3.2.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.2.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.2.1/jquery.min.js</a></p>
<p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/superagent/docs/tail.html</p>
<p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/superagent/docs/tail.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.2.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/jmespath/index.html</p>
<p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/jmespath/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/dot-qs/test/test-dot-querystring.html</p>
<p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/dot-qs/test/test-dot-querystring.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/vipinsun/blockchain-carbon-accounting/commit/d388e16464e00b9ce84df0d247029f534a429b90">d388e16464e00b9ce84df0d247029f534a429b90</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11023 (Medium) detected in multiple libraries - autoclosed - ## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-3.2.1.min.js</b>, <b>jquery-1.7.2.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-3.2.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.2.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.2.1/jquery.min.js</a></p>
<p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/superagent/docs/tail.html</p>
<p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/superagent/docs/tail.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.2.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/jmespath/index.html</p>
<p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/jmespath/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: /utility-emissions-channel/typescript_app/node_modules/dot-qs/test/test-dot-querystring.html</p>
<p>Path to vulnerable library: /utility-emissions-channel/typescript_app/node_modules/dot-qs/test/test-dot-querystring.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/vipinsun/blockchain-carbon-accounting/commit/d388e16464e00b9ce84df0d247029f534a429b90">d388e16464e00b9ce84df0d247029f534a429b90</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in multiple libraries autoclosed cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file utility emissions channel typescript app node modules superagent docs tail html path to vulnerable library utility emissions channel typescript app node modules superagent docs tail html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file utility emissions channel typescript app node modules jmespath index html path to vulnerable library utility emissions channel typescript app node modules jmespath index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file utility emissions channel typescript app node modules dot qs test test dot querystring html path to vulnerable library utility emissions channel typescript app node modules dot qs test test dot querystring html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails step up your open source security game with mend
| 0
|
12,916
| 15,288,166,606
|
IssuesEvent
|
2021-02-23 16:33:47
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
ARM: VMUL not decompiled properly with VFPv2/VFPv3
|
Feature: Processor/ARM Type: Bug
|
**Describe the bug**
When decompiling any binary built with VFPv2/VFPv3 vector extensions for the 32-bit variant of ARM, these instructions get decompiled as CDP instructions. While not necessarily wrong, this does make the instructions significantly more annoying to read.
This [appears](https://github.com/NationalSecurityAgency/ghidra/blob/master/Ghidra/Processors/ARM/data/languages/ARMneon.sinc#L2992-L3002) to be due to the fact that Sleigh constructors only exist for the NEON A1/T1 variants of VMUL.
There doesn't seem to be constructors for the VFPv2/VFPv3/VFPv4 A2/T2 variants that allow a condition code on the A2 variant.
Just figured that I'd report this in case it was an oversight.
**To Reproduce**
1. Decompile any application variant that makes use of VFPv2/VFPv3 variants of VMUL.
2. Notice that it gets decompiled as a CDP instruction.
3. Despair
**Expected behavior**
The instructions get decompiled as VMUL variants, and not as CDP instructions.
**Environment (please complete the following information):**
- OS: Windows 10
- Java Version: 12.0
- Ghidra Version: 9.1
|
1.0
|
ARM: VMUL not decompiled properly with VFPv2/VFPv3 - **Describe the bug**
When decompiling any binary built with VFPv2/VFPv3 vector extensions for the 32-bit variant of ARM, these instructions get decompiled as CDP instructions. While not necessarily wrong, this does make the instructions significantly more annoying to read.
This [appears](https://github.com/NationalSecurityAgency/ghidra/blob/master/Ghidra/Processors/ARM/data/languages/ARMneon.sinc#L2992-L3002) to be due to the fact that Sleigh constructors only exist for the NEON A1/T1 variants of VMUL.
There doesn't seem to be constructors for the VFPv2/VFPv3/VFPv4 A2/T2 variants that allow a condition code on the A2 variant.
Just figured that I'd report this in case it was an oversight.
**To Reproduce**
1. Decompile any application variant that makes use of VFPv2/VFPv3 variants of VMUL.
2. Notice that it gets decompiled as a CDP instruction.
3. Despair
**Expected behavior**
The instructions get decompiled as VMUL variants, and not as CDP instructions.
**Environment (please complete the following information):**
- OS: Windows 10
- Java Version: 12.0
- Ghidra Version: 9.1
|
process
|
arm vmul not decompiled properly with describe the bug when decompiling any binary built with vector extensions for the bit variant of arm these instructions get decompiled as cdp instructions while not necessarily wrong this does make the instructions significantly more annoying to read this to be due to the fact that sleigh constructors only exist for the neon variants of vmul there doesn t seem to be constructors for the variants that allow a condition code on the variant just figured that i d report this in case it was an oversight to reproduce decompile any application variant that makes use of variants of vmul notice that it gets decompiled as a cdp instruction despair expected behavior the instructions get decompiled as vmul variants and not as cdp instructions environment please complete the following information os windows java version ghidra version
| 1
|
18,961
| 24,921,876,549
|
IssuesEvent
|
2022-10-31 01:30:37
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Child process IPC option for plain text serialization
|
child_process feature request stale
|
### Is your feature request related to a problem? Please describe.
Spawning a child process with IPC serialization currently only has [two serialization modes](https://nodejs.org/api/child_process.html#child_process_child_process_spawn_command_args_options), 'json' and 'advanced'. 'advanced' is specific to node/V8, which works great when the subprocess is another Node.js process, but for non node.js processes it's not an option.
JSON serialization mode currently has no option for a custom serializer, so it's hard to send complex data (bigint, function, etc) without first JSON.serializaing with a custom encoder, JSON.parsing it, then having it serialized back to JSON internally.
### Describe the solution you'd like
This overhead can be reduced by adding a plaintext option (probably the easiest) to let the user do the encoding. Though maybe this introduces issues with buffering? If so maybe some way to add a custom JSON replacer.
### Describe alternatives you've considered
Some other IPC mode? Currently I have to resort to a really bad hack to expose the internal writeUtf8String method in child_process/serialization.
https://github.com/extremeheat/JSPyBridge/blob/master/src/pythonia/IpcPipeCom.js#L15
|
1.0
|
Child process IPC option for plain text serialization - ### Is your feature request related to a problem? Please describe.
Spawning a child process with IPC serialization currently only has [two serialization modes](https://nodejs.org/api/child_process.html#child_process_child_process_spawn_command_args_options), 'json' and 'advanced'. 'advanced' is specific to node/V8, which works great when the subprocess is another Node.js process, but for non node.js processes it's not an option.
JSON serialization mode currently has no option for a custom serializer, so it's hard to send complex data (bigint, function, etc) without first JSON.serializaing with a custom encoder, JSON.parsing it, then having it serialized back to JSON internally.
### Describe the solution you'd like
This overhead can be reduced by adding a plaintext option (probably the easiest) to let the user do the encoding. Though maybe this introduces issues with buffering? If so maybe some way to add a custom JSON replacer.
### Describe alternatives you've considered
Some other IPC mode? Currently I have to resort to a really bad hack to expose the internal writeUtf8String method in child_process/serialization.
https://github.com/extremeheat/JSPyBridge/blob/master/src/pythonia/IpcPipeCom.js#L15
|
process
|
child process ipc option for plain text serialization is your feature request related to a problem please describe spawning a child process with ipc serialization currently only has json and advanced advanced is specific to node which works great when the subprocess is another node js process but for non node js processes it s not an option json serialization mode currently has no option for a custom serializer so it s hard to send complex data bigint function etc without first json serializaing with a custom encoder json parsing it then having it serialized back to json internally describe the solution you d like this overhead can be reduced by adding a plaintext option probably the easiest to let the user do the encoding though maybe this introduces issues with buffering if so maybe some way to add a custom json replacer describe alternatives you ve considered some other ipc mode currently i have to resort to a really bad hack to expose the internal method in child process serialization
| 1
|
5,139
| 7,922,824,471
|
IssuesEvent
|
2018-07-05 12:13:54
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
Prompts are not displayed on a newline after evaluating empty string
|
process:eval
|
With cursor at an empty prompt:
```
> *cursor*
```
Hitting RET several times produces this:
```
> > > *cursor*
```
Instead of
```
>
>
> *cursor*
```
|
1.0
|
Prompts are not displayed on a newline after evaluating empty string - With cursor at an empty prompt:
```
> *cursor*
```
Hitting RET several times produces this:
```
> > > *cursor*
```
Instead of
```
>
>
> *cursor*
```
|
process
|
prompts are not displayed on a newline after evaluating empty string with cursor at an empty prompt cursor hitting ret several times produces this cursor instead of cursor
| 1
|
11,734
| 14,576,707,972
|
IssuesEvent
|
2020-12-18 00:08:14
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
how to parse custom log format that contains a pipe as delimiter
|
bug duplicate log-processing
|
my log format as follow:
**03/Sep/2019:10:04:08 +0800 | 10.10.10.10 | http | 304 | 192.168.2.12 | - | www.hello.com | - | - | GET / HTTP/1.1 | 0 | - | Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.110 Safari/547.36 | - | 0.017 | - | - | -**
Please can you help me with the right command ?
Thanks in advance.
Regards
|
1.0
|
how to parse custom log format that contains a pipe as delimiter -
my log format as follow:
**03/Sep/2019:10:04:08 +0800 | 10.10.10.10 | http | 304 | 192.168.2.12 | - | www.hello.com | - | - | GET / HTTP/1.1 | 0 | - | Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.110 Safari/547.36 | - | 0.017 | - | - | -**
Please can you help me with the right command ?
Thanks in advance.
Regards
|
process
|
how to parse custom log format that contains a pipe as delimiter my log format as follow sep http get http mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari please can you help me with the right command thanks in advance regards
| 1
|
15,647
| 19,846,426,421
|
IssuesEvent
|
2022-01-21 07:06:28
|
ooi-data/RS01SBPD-DP01A-06-DOSTAD104-recovered_wfp-dpc_optode_instrument_recovered
|
https://api.github.com/repos/ooi-data/RS01SBPD-DP01A-06-DOSTAD104-recovered_wfp-dpc_optode_instrument_recovered
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:06:28.296042.
## Details
Flow name: `RS01SBPD-DP01A-06-DOSTAD104-recovered_wfp-dpc_optode_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:06:28.296042.
## Details
Flow name: `RS01SBPD-DP01A-06-DOSTAD104-recovered_wfp-dpc_optode_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered wfp dpc optode instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
17,872
| 23,817,012,201
|
IssuesEvent
|
2022-09-05 07:47:43
|
Tencent/tdesign-miniprogram
|
https://api.github.com/repos/Tencent/tdesign-miniprogram
|
closed
|
[t-tabs] t-tab-panel 里面内容绑定事件没有触发
|
bug good first issue in process
|
### tdesign-miniprogram 版本
0.19.3
### 重现链接
_No response_
### 重现步骤
<t-tabs value="{{typeid}}" placement="left" bind:change="onTabsChange">
<t-tab-panel wx:for="{{list}}" wx:key="id" label="{{item.text}}" value="{{item.id}}">
<view bindtap="onchange">内容这里事件无法触发</view>
</t-tab-panel>
</t-tabs>
### 期望结果
_No response_
### 实际结果
_No response_
### 框架版本
_No response_
### 浏览器版本
_No response_
### 系统版本
_No response_
### Node版本
_No response_
### 补充说明
请尽快修复,完全不能用影响挺大的
|
1.0
|
[t-tabs] t-tab-panel 里面内容绑定事件没有触发 - ### tdesign-miniprogram 版本
0.19.3
### 重现链接
_No response_
### 重现步骤
<t-tabs value="{{typeid}}" placement="left" bind:change="onTabsChange">
<t-tab-panel wx:for="{{list}}" wx:key="id" label="{{item.text}}" value="{{item.id}}">
<view bindtap="onchange">内容这里事件无法触发</view>
</t-tab-panel>
</t-tabs>
### 期望结果
_No response_
### 实际结果
_No response_
### 框架版本
_No response_
### 浏览器版本
_No response_
### 系统版本
_No response_
### Node版本
_No response_
### 补充说明
请尽快修复,完全不能用影响挺大的
|
process
|
t tab panel 里面内容绑定事件没有触发 tdesign miniprogram 版本 重现链接 no response 重现步骤 内容这里事件无法触发 期望结果 no response 实际结果 no response 框架版本 no response 浏览器版本 no response 系统版本 no response node版本 no response 补充说明 请尽快修复,完全不能用影响挺大的
| 1
|
41,399
| 8,970,606,115
|
IssuesEvent
|
2019-01-29 14:04:08
|
surge-synthesizer/surge
|
https://api.github.com/repos/surge-synthesizer/surge
|
opened
|
using namespace in header files
|
code cleanup
|
`using namespace` in header files is dangerous.
Google will tell you why but here's a good stack overflow https://stackoverflow.com/questions/4872373/why-is-including-using-namespace-into-a-header-file-a-bad-idea-in-c
We have these:
```
(p3env) paul:~/dev/music/surge$ grep -r "using namespace" src | grep \.h:
src/vst3/SurgeVst3EditController.h:using namespace Steinberg;
src/vst3/SurgeVst3EditController.h:using namespace Steinberg::Vst;
src/vst3/SurgeVst3Processor.h:using namespace Steinberg;
src/common/SurgeStorage.h:using namespace std;
src/common/SurgeSynthesizer.h:using namespace std;
src/common/globals.h:using namespace std;
src/common/gui/PopupEditorSpawner.h:using namespace std;
src/common/gui/SurgeGUIEditor.h:using namespace std;
```
which we should clean up one day. My view is using std:: as a non-used namespace is now becoming the way most folks write, so I'd just eliminate those and fix compile errors. The ones in vst3 are a bit less worrying - especially since the edit controller isn't used.
|
1.0
|
using namespace in header files - `using namespace` in header files is dangerous.
Google will tell you why but here's a good stack overflow https://stackoverflow.com/questions/4872373/why-is-including-using-namespace-into-a-header-file-a-bad-idea-in-c
We have these:
```
(p3env) paul:~/dev/music/surge$ grep -r "using namespace" src | grep \.h:
src/vst3/SurgeVst3EditController.h:using namespace Steinberg;
src/vst3/SurgeVst3EditController.h:using namespace Steinberg::Vst;
src/vst3/SurgeVst3Processor.h:using namespace Steinberg;
src/common/SurgeStorage.h:using namespace std;
src/common/SurgeSynthesizer.h:using namespace std;
src/common/globals.h:using namespace std;
src/common/gui/PopupEditorSpawner.h:using namespace std;
src/common/gui/SurgeGUIEditor.h:using namespace std;
```
which we should clean up one day. My view is using std:: as a non-used namespace is now becoming the way most folks write, so I'd just eliminate those and fix compile errors. The ones in vst3 are a bit less worrying - especially since the edit controller isn't used.
|
non_process
|
using namespace in header files using namespace in header files is dangerous google will tell you why but here s a good stack overflow we have these paul dev music surge grep r using namespace src grep h src h using namespace steinberg src h using namespace steinberg vst src h using namespace steinberg src common surgestorage h using namespace std src common surgesynthesizer h using namespace std src common globals h using namespace std src common gui popupeditorspawner h using namespace std src common gui surgeguieditor h using namespace std which we should clean up one day my view is using std as a non used namespace is now becoming the way most folks write so i d just eliminate those and fix compile errors the ones in are a bit less worrying especially since the edit controller isn t used
| 0
|
22,204
| 30,758,658,676
|
IssuesEvent
|
2023-07-29 12:02:29
|
DxytJuly3/gitalk_blog
|
https://api.github.com/repos/DxytJuly3/gitalk_blog
|
opened
|
[Linux] 守护进程介绍、服务器的部署、日志文件... - July.cc Blogs
|
Gitalk /posts/Linux-Daemon-Process
|
https://www.julysblog.cn/posts/Linux-Daemon-Process
我们使用的系统中, 一般以服务器的方式工作 对外提供服务的服务器, 都是以守护进程的方式在系统中工作的 比如, 我们使用Linux服务器时, 大多都会使用一些终端软件通过ssh远程连接服务器使用 守护进程, 一旦启动之后. 除非用户手动关闭, 否则不会被关闭 会一直运行
|
1.0
|
[Linux] 守护进程介绍、服务器的部署、日志文件... - July.cc Blogs - https://www.julysblog.cn/posts/Linux-Daemon-Process
我们使用的系统中, 一般以服务器的方式工作 对外提供服务的服务器, 都是以守护进程的方式在系统中工作的 比如, 我们使用Linux服务器时, 大多都会使用一些终端软件通过ssh远程连接服务器使用 守护进程, 一旦启动之后. 除非用户手动关闭, 否则不会被关闭 会一直运行
|
process
|
守护进程介绍、服务器的部署、日志文件 july cc blogs 我们使用的系统中 一般以服务器的方式工作 对外提供服务的服务器 都是以守护进程的方式在系统中工作的 比如 我们使用linux服务器时 大多都会使用一些终端软件通过ssh远程连接服务器使用 守护进程 一旦启动之后 除非用户手动关闭 否则不会被关闭 会一直运行
| 1
|
132,658
| 18,268,821,301
|
IssuesEvent
|
2021-10-04 11:40:28
|
artsking/linux-3.0.35
|
https://api.github.com/repos/artsking/linux-3.0.35
|
opened
|
CVE-2017-13695 (Medium) detected in linux-stable-rtv3.8.6
|
security vulnerability
|
## CVE-2017-13695 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35/commit/5992fa81c6ac1b4e9db13f5408d914525c5b7875">5992fa81c6ac1b4e9db13f5408d914525c5b7875</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The acpi_ns_evaluate() function in drivers/acpi/acpica/nseval.c in the Linux kernel through 4.12.9 does not flush the operand cache and causes a kernel stack dump, which allows local users to obtain sensitive information from kernel memory and bypass the KASLR protection mechanism (in the kernel through 4.9) via a crafted ACPI table.
<p>Publish Date: 2017-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-13695>CVE-2017-13695</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/torvalds/linux/commit/97f3c0a4b0579b646b6b10ae5a3d59f0441cc12c#diff-2a91ba8eef66d4e148b6c2abcd63107b">https://github.com/torvalds/linux/commit/97f3c0a4b0579b646b6b10ae5a3d59f0441cc12c#diff-2a91ba8eef66d4e148b6c2abcd63107b</a></p>
<p>Release Date: 2017-08-25</p>
<p>Fix Resolution: v4.17-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-13695 (Medium) detected in linux-stable-rtv3.8.6 - ## CVE-2017-13695 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35/commit/5992fa81c6ac1b4e9db13f5408d914525c5b7875">5992fa81c6ac1b4e9db13f5408d914525c5b7875</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/acpi/acpica/nseval.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The acpi_ns_evaluate() function in drivers/acpi/acpica/nseval.c in the Linux kernel through 4.12.9 does not flush the operand cache and causes a kernel stack dump, which allows local users to obtain sensitive information from kernel memory and bypass the KASLR protection mechanism (in the kernel through 4.9) via a crafted ACPI table.
<p>Publish Date: 2017-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-13695>CVE-2017-13695</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/torvalds/linux/commit/97f3c0a4b0579b646b6b10ae5a3d59f0441cc12c#diff-2a91ba8eef66d4e148b6c2abcd63107b">https://github.com/torvalds/linux/commit/97f3c0a4b0579b646b6b10ae5a3d59f0441cc12c#diff-2a91ba8eef66d4e148b6c2abcd63107b</a></p>
<p>Release Date: 2017-08-25</p>
<p>Fix Resolution: v4.17-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers acpi acpica nseval c drivers acpi acpica nseval c drivers acpi acpica nseval c vulnerability details the acpi ns evaluate function in drivers acpi acpica nseval c in the linux kernel through does not flush the operand cache and causes a kernel stack dump which allows local users to obtain sensitive information from kernel memory and bypass the kaslr protection mechanism in the kernel through via a crafted acpi table publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
411,692
| 27,828,104,238
|
IssuesEvent
|
2023-03-20 00:16:28
|
miguelgrinberg/python-socketio
|
https://api.github.com/repos/miguelgrinberg/python-socketio
|
closed
|
docs: please make sticky sessions requirement louder
|
documentation
|
**Is your feature request related to a problem? Please describe.**
When run as a cluster, this library requires sticky sessions to be enabled. This wasn't obvious at first and so I deployed a buggy cluster.
Please improve the docs or errors to make this requirement clearer (and I'm happy to submit a docs / error message PR, if helpful).
(Also thank for for providing this library -- I am very happy to be able to run my socketio backend in my existing fastapi stack).
**Describe the solution you'd like**
**In the codebase**:
- remove the sticky sessions requirement if possible (I don't understand the internals of this project or the socketio protocol, but I assume the sticky sessions requirement is in here for a legit reason, and removing it would be a large project if possible)
- explain the situation in the error message, if possible. In my case the error is usually 'Invalid session', presumably from 2 places in [handle_request](https://github.com/miguelgrinberg/python-engineio/blob/3af6bf2b4b28dd1d3534ad21b78366ecb60d5503/src/engineio/server.py#L405)
**In docs**:
In the [using a message queue](https://python-socketio.readthedocs.io/en/latest/server.html#using-a-message-queue) section, can you link to the [scalability notes warning](https://python-socketio.readthedocs.io/en/latest/server.html#scalability-notes) (which says to use sticky sessions), and make it clear your cluster will fail if you don't meet the reqs?
**Describe alternatives you've considered**
The alternative I tried first was having my chat intermittently fail in prod. Because my cluster is small, it usually connected eventually, so for a while I assumed it was an error in the ordering of my authorization logic + socket connection.
**Logs**
n/a, but let me know if you want me to capture something
**Additional context**
Did a readthrough of other issues mentioning this:
- #371 'invalid session' with sanic in multiple workers mode
- #437, #219
|
1.0
|
docs: please make sticky sessions requirement louder - **Is your feature request related to a problem? Please describe.**
When run as a cluster, this library requires sticky sessions to be enabled. This wasn't obvious at first and so I deployed a buggy cluster.
Please improve the docs or errors to make this requirement clearer (and I'm happy to submit a docs / error message PR, if helpful).
(Also thank for for providing this library -- I am very happy to be able to run my socketio backend in my existing fastapi stack).
**Describe the solution you'd like**
**In the codebase**:
- remove the sticky sessions requirement if possible (I don't understand the internals of this project or the socketio protocol, but I assume the sticky sessions requirement is in here for a legit reason, and removing it would be a large project if possible)
- explain the situation in the error message, if possible. In my case the error is usually 'Invalid session', presumably from 2 places in [handle_request](https://github.com/miguelgrinberg/python-engineio/blob/3af6bf2b4b28dd1d3534ad21b78366ecb60d5503/src/engineio/server.py#L405)
**In docs**:
In the [using a message queue](https://python-socketio.readthedocs.io/en/latest/server.html#using-a-message-queue) section, can you link to the [scalability notes warning](https://python-socketio.readthedocs.io/en/latest/server.html#scalability-notes) (which says to use sticky sessions), and make it clear your cluster will fail if you don't meet the reqs?
**Describe alternatives you've considered**
The alternative I tried first was having my chat intermittently fail in prod. Because my cluster is small, it usually connected eventually, so for a while I assumed it was an error in the ordering of my authorization logic + socket connection.
**Logs**
n/a, but let me know if you want me to capture something
**Additional context**
Did a readthrough of other issues mentioning this:
- #371 'invalid session' with sanic in multiple workers mode
- #437, #219
|
non_process
|
docs please make sticky sessions requirement louder is your feature request related to a problem please describe when run as a cluster this library requires sticky sessions to be enabled this wasn t obvious at first and so i deployed a buggy cluster please improve the docs or errors to make this requirement clearer and i m happy to submit a docs error message pr if helpful also thank for for providing this library i am very happy to be able to run my socketio backend in my existing fastapi stack describe the solution you d like in the codebase remove the sticky sessions requirement if possible i don t understand the internals of this project or the socketio protocol but i assume the sticky sessions requirement is in here for a legit reason and removing it would be a large project if possible explain the situation in the error message if possible in my case the error is usually invalid session presumably from places in in docs in the section can you link to the which says to use sticky sessions and make it clear your cluster will fail if you don t meet the reqs describe alternatives you ve considered the alternative i tried first was having my chat intermittently fail in prod because my cluster is small it usually connected eventually so for a while i assumed it was an error in the ordering of my authorization logic socket connection logs n a but let me know if you want me to capture something additional context did a readthrough of other issues mentioning this invalid session with sanic in multiple workers mode
| 0
|
20,163
| 26,716,686,143
|
IssuesEvent
|
2023-01-28 16:05:01
|
LLazyEmail/nomoretogo_email_template
|
https://api.github.com/repos/LLazyEmail/nomoretogo_email_template
|
closed
|
rename to Weekly Menu button and move to config file
|
todo in process
|
https://github.com/LLazyEmail/nomoretogo_email_template/blob/9c1f21518fc1e72bfa95baa91db45fc2d23b822c/src/display/displayContent.js#L175
```javascript
// TODO rename to Weekly Menu button and move to config file
const contentButtonParams = {
id: '04-22-2021',
href: 'https://www.nomoretogo.com/weekly-menu/',
```
|
1.0
|
rename to Weekly Menu button and move to config file - https://github.com/LLazyEmail/nomoretogo_email_template/blob/9c1f21518fc1e72bfa95baa91db45fc2d23b822c/src/display/displayContent.js#L175
```javascript
// TODO rename to Weekly Menu button and move to config file
const contentButtonParams = {
id: '04-22-2021',
href: 'https://www.nomoretogo.com/weekly-menu/',
```
|
process
|
rename to weekly menu button and move to config file javascript todo rename to weekly menu button and move to config file const contentbuttonparams id href
| 1
|
112,038
| 14,212,892,203
|
IssuesEvent
|
2020-11-17 01:10:53
|
phetsims/energy-skate-park
|
https://api.github.com/repos/phetsims/energy-skate-park
|
opened
|
Slider pixel polishes
|
design:polish
|
@KatieWoe reported in https://github.com/phetsims/QA/issues/578#issuecomment-728427972:
>On MacOS 11 Safari, I noticed a bit of overlap with longer strings and the thumbs of sliders.
><img width="254" alt="Screen Shot 2020-11-16 at 4 35 37 PM" src="https://user-images.githubusercontent.com/41024075/99321450-12db5180-282b-11eb-93fa-8b796a1734a6.png">
This issue is not limited to macOS/Safari or i18n. The slider thumb is a bit too big, and the tick mark length is a bit too short. The result is that the thumb touches letters that descend below the baseline. This affects both ESP and ESPB.


Luckily, the slider options are factored out into [EnergySkatParkConstants.js](https://github.com/phetsims/energy-skate-park/blob/master/js/common/EnergySkateParkConstants.js#L39-L43) which made it really easy to test out a few options locally. I reduced the `thumbSize` height to 23 and increased the `majorTickLength` to 11. This removes the overlap between the slider label and thumb, but does not change the overall height of the panel on the Intro Screen (where vertical space is tight).
| Before | After |
| --- | --- |
|  |  |
This is honestly quite minor, so I'm hesitant to do anything so close to the finish line. @ariel-phet what do you think?
|
1.0
|
Slider pixel polishes - @KatieWoe reported in https://github.com/phetsims/QA/issues/578#issuecomment-728427972:
>On MacOS 11 Safari, I noticed a bit of overlap with longer strings and the thumbs of sliders.
><img width="254" alt="Screen Shot 2020-11-16 at 4 35 37 PM" src="https://user-images.githubusercontent.com/41024075/99321450-12db5180-282b-11eb-93fa-8b796a1734a6.png">
This issue is not limited to macOS/Safari or i18n. The slider thumb is a bit too big, and the tick mark length is a bit too short. The result is that the thumb touches letters that descend below the baseline. This affects both ESP and ESPB.


Luckily, the slider options are factored out into [EnergySkatParkConstants.js](https://github.com/phetsims/energy-skate-park/blob/master/js/common/EnergySkateParkConstants.js#L39-L43) which made it really easy to test out a few options locally. I reduced the `thumbSize` height to 23 and increased the `majorTickLength` to 11. This removes the overlap between the slider label and thumb, but does not change the overall height of the panel on the Intro Screen (where vertical space is tight).
| Before | After |
| --- | --- |
|  |  |
This is honestly quite minor, so I'm hesitant to do anything so close to the finish line. @ariel-phet what do you think?
|
non_process
|
slider pixel polishes katiewoe reported in on macos safari i noticed a bit of overlap with longer strings and the thumbs of sliders img width alt screen shot at pm src this issue is not limited to macos safari or the slider thumb is a bit too big and the tick mark length is a bit too short the result is that the thumb touches letters that descend below the baseline this affects both esp and espb luckily the slider options are factored out into which made it really easy to test out a few options locally i reduced the thumbsize height to and increased the majorticklength to this removes the overlap between the slider label and thumb but does not change the overall height of the panel on the intro screen where vertical space is tight before after this is honestly quite minor so i m hesitant to do anything so close to the finish line ariel phet what do you think
| 0
|
385,358
| 26,633,218,537
|
IssuesEvent
|
2023-01-24 19:33:42
|
ethyca/fidesdocs
|
https://api.github.com/repos/ethyca/fidesdocs
|
closed
|
update `FIDES__LOG_PII` reference to FIDES__LOGGING__LOG_PII`
|
documentation
|
### Docs Update Description
We used to control whether PII was masked in our logs via an env variable (and corresponding application config property) that was `FIDES__LOG_PII`. That was changed to `FIDES__LOGGING__LOG_PII` in [this PR](https://github.com/ethyca/fides/pull/2031). We should update our documentation accordingly.
### Additional context
(Internal helm charts have already been updated accordingly: https://github.com/ethyca/fides-helm/pull/13)
|
1.0
|
update `FIDES__LOG_PII` reference to FIDES__LOGGING__LOG_PII` - ### Docs Update Description
We used to control whether PII was masked in our logs via an env variable (and corresponding application config property) that was `FIDES__LOG_PII`. That was changed to `FIDES__LOGGING__LOG_PII` in [this PR](https://github.com/ethyca/fides/pull/2031). We should update our documentation accordingly.
### Additional context
(Internal helm charts have already been updated accordingly: https://github.com/ethyca/fides-helm/pull/13)
|
non_process
|
update fides log pii reference to fides logging log pii docs update description we used to control whether pii was masked in our logs via an env variable and corresponding application config property that was fides log pii that was changed to fides logging log pii in we should update our documentation accordingly additional context internal helm charts have already been updated accordingly
| 0
|
45,894
| 7,208,625,526
|
IssuesEvent
|
2018-02-07 04:15:20
|
GoogleCloudPlatform/google-cloud-java
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-java
|
closed
|
BigQuery Data Transfer: ListDataSourcesRequest not documented / not linked to
|
api: bigquery documentation priority: p2 type: process
|
See: https://googlecloudplatform.github.io/google-cloud-java/latest/apidocs/com/google/cloud/bigquery/datatransfer/v1/DataTransferServiceClient.html#listDataSources-com.google.cloud.bigquery.datatransfer.v1.ListDataSourcesRequest-
I would expect `com.google.cloud.bigquery.datatransfer.v1.ListDataSourcesRequest` to be a hyperlink to the ListDataSourcesRequest class, but it is not. A search within that page for ListDataSourcesRequest does not show any documentation for that class.
|
1.0
|
BigQuery Data Transfer: ListDataSourcesRequest not documented / not linked to - See: https://googlecloudplatform.github.io/google-cloud-java/latest/apidocs/com/google/cloud/bigquery/datatransfer/v1/DataTransferServiceClient.html#listDataSources-com.google.cloud.bigquery.datatransfer.v1.ListDataSourcesRequest-
I would expect `com.google.cloud.bigquery.datatransfer.v1.ListDataSourcesRequest` to be a hyperlink to the ListDataSourcesRequest class, but it is not. A search within that page for ListDataSourcesRequest does not show any documentation for that class.
|
non_process
|
bigquery data transfer listdatasourcesrequest not documented not linked to see i would expect com google cloud bigquery datatransfer listdatasourcesrequest to be a hyperlink to the listdatasourcesrequest class but it is not a search within that page for listdatasourcesrequest does not show any documentation for that class
| 0
|
59,878
| 3,117,581,479
|
IssuesEvent
|
2015-09-04 02:52:40
|
framingeinstein/issues-test
|
https://api.github.com/repos/framingeinstein/issues-test
|
opened
|
SPK-215: Add to Wishlist: Message contains a link that merges with background color
|
priority:normal priority:normal priority:normal priority:normal priority:normal resolution:fixed resolution:fixed
|
"Item has been added to your wishlist. Click here to continue shopping" message contains a link ("here") that has the same color as the background.
Steps:
1. Go to http://qa.speakman.rallyapp.co/shower-and-bath/
2. Add any item to Wishlist
3. Note message that appears on top of the page
Browsers/OS:
Firefox 37 Mac OS X 10.9.5
Safari 7.1 Mac OS X 10.9.5
Chrome 41 Mac OS X 10.9.4
|
5.0
|
SPK-215: Add to Wishlist: Message contains a link that merges with background color - "Item has been added to your wishlist. Click here to continue shopping" message contains a link ("here") that has the same color as the background.
Steps:
1. Go to http://qa.speakman.rallyapp.co/shower-and-bath/
2. Add any item to Wishlist
3. Note message that appears on top of the page
Browsers/OS:
Firefox 37 Mac OS X 10.9.5
Safari 7.1 Mac OS X 10.9.5
Chrome 41 Mac OS X 10.9.4
|
non_process
|
spk add to wishlist message contains a link that merges with background color item has been added to your wishlist click here to continue shopping message contains a link here that has the same color as the background steps go to add any item to wishlist note message that appears on top of the page browsers os firefox mac os x safari mac os x chrome mac os x
| 0
|
3,670
| 6,705,664,517
|
IssuesEvent
|
2017-10-12 01:49:32
|
zero-os/0-Disk
|
https://api.github.com/repos/zero-os/0-Disk
|
closed
|
Add some testable (godoc) examples to 0-Disk packages
|
process_duplicate process_wontfix type_feature
|
A golang feature we haven't made use of yet, is testable examples. It allows us to write examples which will be available in the godoc documentation for the relevant package, while the `go test` tool ensures those examples are actually valid.
More information: https://blog.golang.org/examples
Packages where we should add such examples:
- [x] github.com/zero-os/0-Disk
+ how to create a new hash, how to create a hash from bytes, how would you use the Hash methods afterwards;
+ how to create and use a hasher;
+ how to create and use a keyed hasher;
- [x] ~github.com/zero-os/0-Disk/log~
+ ~how to create a logger and how to use it;~
+ ~how to use and configure the global logger;~
+ ~how to broadcast errors;~
- [x] github.com/zero-os/0-Disk/nbd/ardb/storage
+ how to create a storage for a given vdisk;
+ how to use a (block)storage;
|
2.0
|
Add some testable (godoc) examples to 0-Disk packages - A golang feature we haven't made use of yet, is testable examples. It allows us to write examples which will be available in the godoc documentation for the relevant package, while the `go test` tool ensures those examples are actually valid.
More information: https://blog.golang.org/examples
Packages where we should add such examples:
- [x] github.com/zero-os/0-Disk
+ how to create a new hash, how to create a hash from bytes, how would you use the Hash methods afterwards;
+ how to create and use a hasher;
+ how to create and use a keyed hasher;
- [x] ~github.com/zero-os/0-Disk/log~
+ ~how to create a logger and how to use it;~
+ ~how to use and configure the global logger;~
+ ~how to broadcast errors;~
- [x] github.com/zero-os/0-Disk/nbd/ardb/storage
+ how to create a storage for a given vdisk;
+ how to use a (block)storage;
|
process
|
add some testable godoc examples to disk packages a golang feature we haven t made use of yet is testable examples it allows us to write examples which will be available in the godoc documentation for the relevant package while the go test tool ensures those examples are actually valid more information packages where we should add such examples github com zero os disk how to create a new hash how to create a hash from bytes how would you use the hash methods afterwards how to create and use a hasher how to create and use a keyed hasher github com zero os disk log how to create a logger and how to use it how to use and configure the global logger how to broadcast errors github com zero os disk nbd ardb storage how to create a storage for a given vdisk how to use a block storage
| 1
|
1,743
| 10,678,143,835
|
IssuesEvent
|
2019-10-21 16:41:49
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
closed
|
Set flank version in ui-test/Dockerfile as ENV var
|
eng:automation 🌟 feature request
|
Currently we download the flank jar file once when the Docker image is built. To have more control over this, we'd like to pass in the desired flank version num to the Dockerfile as an env var.
|
1.0
|
Set flank version in ui-test/Dockerfile as ENV var - Currently we download the flank jar file once when the Docker image is built. To have more control over this, we'd like to pass in the desired flank version num to the Dockerfile as an env var.
|
non_process
|
set flank version in ui test dockerfile as env var currently we download the flank jar file once when the docker image is built to have more control over this we d like to pass in the desired flank version num to the dockerfile as an env var
| 0
|
16,173
| 20,618,541,731
|
IssuesEvent
|
2022-03-07 15:21:39
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject test failed
|
area-System.Diagnostics.Process blocking-clean-ci untriaged
|
Configuration: `net7.0-Linux-Release-x64-CoreCLR_checked-Ubuntu.1804.Amd64.Open`
Build: [runtime pipeline of the rolling build](https://dev.azure.com/dnceng/public/_build/results?buildId=1635569&view=ms.vss-test-web.build-test-results-tab&runId=45291136&paneView=dotnet-dnceng.dnceng-build-release-tasks.helix-test-information-tab&resultId=147215), [console log](https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-f713537715f04c66ac/System.Diagnostics.Process.Tests/1/console.34f7ff4c.log?sv=2019-07-07&se=2022-03-20T08%3A53%3A47Z&sr=c&sp=rl&sig=%2FCxSnOmgSEXdBGlx6bMmwpMV0AFOlnBRyg8Qd6IQSU8%3D)
```
System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject(invokeRequired: False) [FAIL]
Assert.Equal() Failure
↓ (pos 105)
Expected: ··· closed error stream
Actual: ··· closed error streamAssert failure(PID 18855 [0x000049a7], Th···
↑ (pos 105)
Stack Trace:
/_/src/libraries/System.Diagnostics.Process/tests/ProcessStreamReadTests.cs(85,0): at System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject(Boolean invokeRequired)
...
Finished: System.Diagnostics.Process.Tests
=== TEST EXECUTION SUMMARY ===
System.Diagnostics.Process.Tests Total: 317, Errors: 0, Failed: 1, Skipped: 2, Time: 42.933s
```
|
1.0
|
System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject test failed - Configuration: `net7.0-Linux-Release-x64-CoreCLR_checked-Ubuntu.1804.Amd64.Open`
Build: [runtime pipeline of the rolling build](https://dev.azure.com/dnceng/public/_build/results?buildId=1635569&view=ms.vss-test-web.build-test-results-tab&runId=45291136&paneView=dotnet-dnceng.dnceng-build-release-tasks.helix-test-information-tab&resultId=147215), [console log](https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-f713537715f04c66ac/System.Diagnostics.Process.Tests/1/console.34f7ff4c.log?sv=2019-07-07&se=2022-03-20T08%3A53%3A47Z&sr=c&sp=rl&sig=%2FCxSnOmgSEXdBGlx6bMmwpMV0AFOlnBRyg8Qd6IQSU8%3D)
```
System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject(invokeRequired: False) [FAIL]
Assert.Equal() Failure
↓ (pos 105)
Expected: ··· closed error stream
Actual: ··· closed error streamAssert failure(PID 18855 [0x000049a7], Th···
↑ (pos 105)
Stack Trace:
/_/src/libraries/System.Diagnostics.Process/tests/ProcessStreamReadTests.cs(85,0): at System.Diagnostics.Tests.ProcessStreamReadTests.TestAsyncErrorStream_SynchronizingObject(Boolean invokeRequired)
...
Finished: System.Diagnostics.Process.Tests
=== TEST EXECUTION SUMMARY ===
System.Diagnostics.Process.Tests Total: 317, Errors: 0, Failed: 1, Skipped: 2, Time: 42.933s
```
|
process
|
system diagnostics tests processstreamreadtests testasyncerrorstream synchronizingobject test failed configuration linux release coreclr checked ubuntu open build system diagnostics tests processstreamreadtests testasyncerrorstream synchronizingobject invokerequired false assert equal failure ↓ pos expected ··· closed error stream actual ··· closed error streamassert failure pid th··· ↑ pos stack trace src libraries system diagnostics process tests processstreamreadtests cs at system diagnostics tests processstreamreadtests testasyncerrorstream synchronizingobject boolean invokerequired finished system diagnostics process tests test execution summary system diagnostics process tests total errors failed skipped time
| 1
|
21,356
| 29,189,101,062
|
IssuesEvent
|
2023-05-19 18:08:01
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] Implement `remove` and `replace`
|
.metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
### Related PRs
- https://github.com/metabase/metabase/pull/29625
- https://github.com/metabase/metabase/pull/29657
- https://github.com/metabase/metabase/pull/29775
- https://github.com/metabase/metabase/pull/29777
- https://github.com/metabase/metabase/pull/29803
- https://github.com/metabase/metabase/pull/29841
- https://github.com/metabase/metabase/pull/30002
- https://github.com/metabase/metabase/pull/30074
- https://github.com/metabase/metabase/pull/30373
- https://github.com/metabase/metabase/pull/30775
### `remove`
takes a clause and removes it. I think this should probably take `stage-number` as an optional parameter, defaulting to `-1` (last stage of the query) if not explicitly specified, the same way literally everything else does. Since we have UUIDs (see below) I guess it would be possible to walk all the stages and find something to remove, but this seems more complicated than we need right now
```clj
(let [my-order-by (first (order-bys query -1))]
(remove query my-order-by))
```
or
```js
const orderBy = query.orderBys()?.[0];
query = query.remove(orderBy);
```
The whole reason we added `:lib/uuid` in the options in the first place was to facilitate this sort of stuff, so we should obviously leverage this info.
I think we'll want to have some sort of underlying `remove*` multimethod that we can implement to handle various MBQL clauses and maps. e.g. we'd have an `:asc` method which would then go look in `:order-by` and find the matching `:asc` with the same `:lib/uuid` and remove that from the query. Some stuff like `:field` clauses would probably have to look in multiple places, e.g. both `:fields` and `:breakout`. On that note, if we're not already doing it, we should enforce a rule against duplicate UUIDs in a query stage (ignoring metadata)
Removal methods that affect the columns that will be in the response should update `:lib/stage-metadata` for the affected and subsequent stages, e.g. if I remove something from `:fields` or `:breakout` the `:lib/stage-metadata` should reflect this.
Open question: should we "clean" the query and remove orphaned references? E.g. if I have a join and add a filter against a joined field, then remove the join, should we do a "cascading delete" and remove the filter as well? I am about 95% sure the answer should be yes here. So let's implement that too.
This will have to be done on a case-by-case basis for different `remove*` impls
### `replace`
This is basically the same as `remove`, but instead of removing it just replaces the clause in that location with a different clause
```clj
(let [my-order-by (first (lib/order-bys query -1))
new-order-by (lib/order-by-clause (lib/field whatever))]
(lib/replace query my-order-by new-order-by))
```
I guess this should probably be implemented with an underlying multimethod as well. This also needs to update metadata and "clean" the query where appropriate.
Question: can we be smart and instead of removing stuff, update it to the equivalent when possible? E.g. if I join table X and have a filter where X.ID = 1, and I replace the join with one against Table Y, which also has ID, can we update the filter to be Y.ID = 1? That would be **smart AF**. I think we should try to do that as a best-effort thing.
|
1.0
|
[MLv2] Implement `remove` and `replace` - ### Related PRs
- https://github.com/metabase/metabase/pull/29625
- https://github.com/metabase/metabase/pull/29657
- https://github.com/metabase/metabase/pull/29775
- https://github.com/metabase/metabase/pull/29777
- https://github.com/metabase/metabase/pull/29803
- https://github.com/metabase/metabase/pull/29841
- https://github.com/metabase/metabase/pull/30002
- https://github.com/metabase/metabase/pull/30074
- https://github.com/metabase/metabase/pull/30373
- https://github.com/metabase/metabase/pull/30775
### `remove`
takes a clause and removes it. I think this should probably take `stage-number` as an optional parameter, defaulting to `-1` (last stage of the query) if not explicitly specified, the same way literally everything else does. Since we have UUIDs (see below) I guess it would be possible to walk all the stages and find something to remove, but this seems more complicated than we need right now
```clj
(let [my-order-by (first (order-bys query -1))]
(remove query my-order-by))
```
or
```js
const orderBy = query.orderBys()?.[0];
query = query.remove(orderBy);
```
The whole reason we added `:lib/uuid` in the options in the first place was to facilitate this sort of stuff, so we should obviously leverage this info.
I think we'll want to have some sort of underlying `remove*` multimethod that we can implement to handle various MBQL clauses and maps. e.g. we'd have an `:asc` method which would then go look in `:order-by` and find the matching `:asc` with the same `:lib/uuid` and remove that from the query. Some stuff like `:field` clauses would probably have to look in multiple places, e.g. both `:fields` and `:breakout`. On that note, if we're not already doing it, we should enforce a rule against duplicate UUIDs in a query stage (ignoring metadata)
Removal methods that affect the columns that will be in the response should update `:lib/stage-metadata` for the affected and subsequent stages, e.g. if I remove something from `:fields` or `:breakout` the `:lib/stage-metadata` should reflect this.
Open question: should we "clean" the query and remove orphaned references? E.g. if I have a join and add a filter against a joined field, then remove the join, should we do a "cascading delete" and remove the filter as well? I am about 95% sure the answer should be yes here. So let's implement that too.
This will have to be done on a case-by-case basis for different `remove*` impls
### `replace`
This is basically the same as `remove`, but instead of removing it just replaces the clause in that location with a different clause
```clj
(let [my-order-by (first (lib/order-bys query -1))
new-order-by (lib/order-by-clause (lib/field whatever))]
(lib/replace query my-order-by new-order-by))
```
I guess this should probably be implemented with an underlying multimethod as well. This also needs to update metadata and "clean" the query where appropriate.
Question: can we be smart and instead of removing stuff, update it to the equivalent when possible? E.g. if I join table X and have a filter where X.ID = 1, and I replace the join with one against Table Y, which also has ID, can we update the filter to be Y.ID = 1? That would be **smart AF**. I think we should try to do that as a best-effort thing.
|
process
|
implement remove and replace related prs remove takes a clause and removes it i think this should probably take stage number as an optional parameter defaulting to last stage of the query if not explicitly specified the same way literally everything else does since we have uuids see below i guess it would be possible to walk all the stages and find something to remove but this seems more complicated than we need right now clj let remove query my order by or js const orderby query orderbys query query remove orderby the whole reason we added lib uuid in the options in the first place was to facilitate this sort of stuff so we should obviously leverage this info i think we ll want to have some sort of underlying remove multimethod that we can implement to handle various mbql clauses and maps e g we d have an asc method which would then go look in order by and find the matching asc with the same lib uuid and remove that from the query some stuff like field clauses would probably have to look in multiple places e g both fields and breakout on that note if we re not already doing it we should enforce a rule against duplicate uuids in a query stage ignoring metadata removal methods that affect the columns that will be in the response should update lib stage metadata for the affected and subsequent stages e g if i remove something from fields or breakout the lib stage metadata should reflect this open question should we clean the query and remove orphaned references e g if i have a join and add a filter against a joined field then remove the join should we do a cascading delete and remove the filter as well i am about sure the answer should be yes here so let s implement that too this will have to be done on a case by case basis for different remove impls replace this is basically the same as remove but instead of removing it just replaces the clause in that location with a different clause clj let my order by first lib order bys query new order by lib order by clause lib field whatever lib replace query my order by new order by i guess this should probably be implemented with an underlying multimethod as well this also needs to update metadata and clean the query where appropriate question can we be smart and instead of removing stuff update it to the equivalent when possible e g if i join table x and have a filter where x id and i replace the join with one against table y which also has id can we update the filter to be y id that would be smart af i think we should try to do that as a best effort thing
| 1
|
21,708
| 30,208,063,370
|
IssuesEvent
|
2023-07-05 10:48:58
|
q191201771/lal
|
https://api.github.com/repos/q191201771/lal
|
closed
|
demo中的pullrtmp2hls无法工作
|
#Bug *In process * help wanted
|
pullrtmp2hls 和 -i rtsp 做连接 直接crash。
###
2021/07/23 23:05:19.665373 [22;36m INFO[0m [RTMPPULL1] < R _result(). - client_session.go:479
2021/07/23 23:05:19.665373 [22;36m INFO[0m [RTMPPULL1] > W play('mystream'). - client_session.go:482
2021/07/23 23:05:19.665891 [22;33m WARN[0m [RTMPPULL1] read user control message, ignore. buf=00000000 00 04 00 00 00 01 |......|
- client_session.go:350
2021/07/23 23:05:19.666407 [22;33m WARN[0m [RTMPPULL1] read user control message, ignore. buf=00000000 00 00 00 00 00 01 |......|
- client_session.go:350
2021/07/23 23:05:19.666925 [22;36m INFO[0m [RTMPPULL1] < R onStatus('NetStream.Play.Start'). - client_session.go:435
panic: runtime error: invalid memory address or nil pointer dereference
[signal 0xc0000005 code=0x0 addr=0x18 pc=0xdb73a8]
###
比较容易重现的。 lalserver做服务,pullrtmp2hls 和lalserver连接。
|
1.0
|
demo中的pullrtmp2hls无法工作 - pullrtmp2hls 和 -i rtsp 做连接 直接crash。
###
2021/07/23 23:05:19.665373 [22;36m INFO[0m [RTMPPULL1] < R _result(). - client_session.go:479
2021/07/23 23:05:19.665373 [22;36m INFO[0m [RTMPPULL1] > W play('mystream'). - client_session.go:482
2021/07/23 23:05:19.665891 [22;33m WARN[0m [RTMPPULL1] read user control message, ignore. buf=00000000 00 04 00 00 00 01 |......|
- client_session.go:350
2021/07/23 23:05:19.666407 [22;33m WARN[0m [RTMPPULL1] read user control message, ignore. buf=00000000 00 00 00 00 00 01 |......|
- client_session.go:350
2021/07/23 23:05:19.666925 [22;36m INFO[0m [RTMPPULL1] < R onStatus('NetStream.Play.Start'). - client_session.go:435
panic: runtime error: invalid memory address or nil pointer dereference
[signal 0xc0000005 code=0x0 addr=0x18 pc=0xdb73a8]
###
比较容易重现的。 lalserver做服务,pullrtmp2hls 和lalserver连接。
|
process
|
和 i rtsp 做连接 直接crash。 r result client session go w play mystream client session go read user control message ignore buf client session go read user control message ignore buf client session go r onstatus netstream play start client session go panic runtime error invalid memory address or nil pointer dereference 比较容易重现的。 lalserver做服务, 和lalserver连接。
| 1
|
20,109
| 26,648,502,260
|
IssuesEvent
|
2023-01-25 11:53:56
|
firebase/firebase-cpp-sdk
|
https://api.github.com/repos/firebase/firebase-cpp-sdk
|
reopened
|
[C++] Nightly Integration Testing Report for Firestore
|
type: process nightly-testing
|
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit 9c494f708c5918d1b82e9624460b7ab4bbaf8431
Last updated: Tue Jan 24 04:03 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3995022157)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 9c494f708c5918d1b82e9624460b7ab4bbaf8431
Last updated: Tue Jan 24 15:17 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3996121210)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit 21f954fa9d0300d3e68ee16d557eac451136da26
Last updated: Wed Jan 25 03:51 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4005239522)**
|
1.0
|
[C++] Nightly Integration Testing Report for Firestore -
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit 9c494f708c5918d1b82e9624460b7ab4bbaf8431
Last updated: Tue Jan 24 04:03 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3995022157)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 9c494f708c5918d1b82e9624460b7ab4bbaf8431
Last updated: Tue Jan 24 15:17 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3996121210)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit 21f954fa9d0300d3e68ee16d557eac451136da26
Last updated: Wed Jan 25 03:51 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4005239522)**
|
process
|
nightly integration testing report for firestore ✅ nbsp integration test succeeded requested by on commit last updated tue jan pst ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated tue jan pst ✅ nbsp integration test succeeded requested by on commit last updated wed jan pst
| 1
|
77,820
| 3,507,278,186
|
IssuesEvent
|
2016-01-08 12:21:14
|
OregonCore/OregonCore
|
https://api.github.com/repos/OregonCore/OregonCore
|
closed
|
Client - Lagging of core (BB #802)
|
Category: Exploit migrated Priority: Medium Type: Bug
|
This issue was migrated from bitbucket.
**Original Reporter:** smoldar
**Original Date:** 29.01.2015 22:32:32 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** closed
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/802
<hr>
Player tryes to login again, AccountId = XXX
Player tryes to login again, AccountId = XXX
Player tryes to login again, AccountId = XXX
Player tryes to login again, AccountId = XXX
100x spammed in server.log / sec i think about packet attack of core, when this spam runned, core was laggy 50-5000ms
I think this problem is here, this may be blocked by different way i think.
Destroying of client session, or something else to avoid laggy, because client can repeatly spam core.
CharacterHandler.cpp
```
#!c++
void WorldSession::HandlePlayerLoginOpcode(WorldPacket & recv_data)
{
if (PlayerLoading() || GetPlayer() != NULL)
{
sLog.outError("Player tryes to login again, AccountId = %d",GetAccountId());
return;
}
```
|
1.0
|
Client - Lagging of core (BB #802) - This issue was migrated from bitbucket.
**Original Reporter:** smoldar
**Original Date:** 29.01.2015 22:32:32 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** closed
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/802
<hr>
Player tryes to login again, AccountId = XXX
Player tryes to login again, AccountId = XXX
Player tryes to login again, AccountId = XXX
Player tryes to login again, AccountId = XXX
100x spammed in server.log / sec i think about packet attack of core, when this spam runned, core was laggy 50-5000ms
I think this problem is here, this may be blocked by different way i think.
Destroying of client session, or something else to avoid laggy, because client can repeatly spam core.
CharacterHandler.cpp
```
#!c++
void WorldSession::HandlePlayerLoginOpcode(WorldPacket & recv_data)
{
if (PlayerLoading() || GetPlayer() != NULL)
{
sLog.outError("Player tryes to login again, AccountId = %d",GetAccountId());
return;
}
```
|
non_process
|
client lagging of core bb this issue was migrated from bitbucket original reporter smoldar original date gmt original priority major original type bug original state closed direct link player tryes to login again accountid xxx player tryes to login again accountid xxx player tryes to login again accountid xxx player tryes to login again accountid xxx spammed in server log sec i think about packet attack of core when this spam runned core was laggy i think this problem is here this may be blocked by different way i think destroying of client session or something else to avoid laggy because client can repeatly spam core characterhandler cpp c void worldsession handleplayerloginopcode worldpacket recv data if playerloading getplayer null slog outerror player tryes to login again accountid d getaccountid return
| 0
|
17,860
| 23,807,284,441
|
IssuesEvent
|
2022-09-04 08:22:10
|
anitsh/til
|
https://api.github.com/repos/anitsh/til
|
opened
|
Kanban Maturity Model and The 4-Step Process to defect management by Nave
|
agile quality qa process
|
Introduce agile techniques like Definition of Done (DoD), code reviews and pair programming to make sure you eliminate defects at an early development stage. Your team will become more self-managed, motivated and engaged. Shorter lead times lead to faster innovation, greater customer satisfaction and substantial profitability.
Focus on making sure the defects do not occur in the first place or, at least, the defects are significantly reduced before making their way into production.
Defect management will help you increase the consistency, predictability and effectiveness of your development efforts, and it will ultimately reduce your delivery times.
The Kanban Maturity Model describes the general practices of reducing defects to help you improve your business outcomes. Here is the 4-step process which will enable you to achieve a more sustainable delivery of quality software.
Step 1: Visualize Defects
Step 2: Track Defects
Step 3: Perform Defect Clustering Analysis
Step 4: Define Relevant Policies
Step 1: Visualize Defects
We cannot manage what we cannot see. Increasing transparency is the first step towards reducing defects and raising awareness
of the disruptive impact they may have.
When you’re testing a work item, and a new defect is discovered, create a separate ticket and mark it as a defect. Add the following essential information to the ticket:
Add a brief description of the defect.
Add a link to the work item the defect is associated with.
Specify the reason that has caused the defect.
Assign the people responsible for fixing the defect.
Assign severity.
Use a different color for defects to make them easily recognizable from other types of work.
Step 2: Track Defects
In the book “The Economics of Software Quality”, defect tracking is used in more than 55% of military and defense software applications, but in less than 15% of IT, commercial, web or embedded applications. Given their prevention effectiveness, it is surprising that defect measurement is not used more often.
The second step of the defect management process is to track and analyze software defects on a regular basis.
The Throughput Run Chart displays the throughput of your team on a daily, weekly or monthly basis and compares its values over time. Use it to keep track of the number of defects you resolve and assess how trends have developed.
The Throughput Breakdown Chart can help you evaluate what type of work is demanding the bulk of the time. It is especially useful when you want to emphasize the high volume of defects you’re handling.
Step 3: Perform Defect Clustering Analysis
The third step of the defect management process is to identify the clusters of causes hindering your delivery times. To perform defect clustering analysis:
Evaluate the causes of defects and their resolution times.
Group the causes by clustering the defects which are similar in nature.
Capture the total cycle time for each group to discover which are the most critical causes of defects.
Report the defect clustering results at the Service Delivery Review, Risk Review, and Operations Review. That way, you’ll keep your attention on the defects with the highest impact on your development process, and you’ll emphasize the urgency to act upon them sooner.
Step 4: Define Relevant Policies
The goal of this step is to bring continuous improvement to your development practices and help you maintain a stable,
predictable workflow in the long run.
Introduce policies on how defects should be handled. A defect ticket should be attached to the work item that it belongs to. The work item should then remain in the same process state where the problem was first discovered.
Create a new ticket for your defect and associate it with the work item that spawned it.
Place the new ticket in the rework queue state (i.e. Development Queue).
Block the work item in the current process state (i.e. Testing).
By adding new items into your workflow, you count them against your WIP limit.
An excessive number of defects will then block the pull of new work until any quality issues are first resolved.
This approach will also help you avoid the common anti-pattern of moving work backward on your Kanban board.
Make sure your activities are not just targeted at finding defects after they have happened.
https://www.kanbanmaturitymodel.com
https://getnave.com/throughput-run-chart
https://www.infoq.com/articles/blockers-defects-process-improvement
|
1.0
|
Kanban Maturity Model and The 4-Step Process to defect management by Nave - Introduce agile techniques like Definition of Done (DoD), code reviews and pair programming to make sure you eliminate defects at an early development stage. Your team will become more self-managed, motivated and engaged. Shorter lead times lead to faster innovation, greater customer satisfaction and substantial profitability.
Focus on making sure the defects do not occur in the first place or, at least, the defects are significantly reduced before making their way into production.
Defect management will help you increase the consistency, predictability and effectiveness of your development efforts, and it will ultimately reduce your delivery times.
The Kanban Maturity Model describes the general practices of reducing defects to help you improve your business outcomes. Here is the 4-step process which will enable you to achieve a more sustainable delivery of quality software.
Step 1: Visualize Defects
Step 2: Track Defects
Step 3: Perform Defect Clustering Analysis
Step 4: Define Relevant Policies
Step 1: Visualize Defects
We cannot manage what we cannot see. Increasing transparency is the first step towards reducing defects and raising awareness
of the disruptive impact they may have.
When you’re testing a work item, and a new defect is discovered, create a separate ticket and mark it as a defect. Add the following essential information to the ticket:
Add a brief description of the defect.
Add a link to the work item the defect is associated with.
Specify the reason that has caused the defect.
Assign the people responsible for fixing the defect.
Assign severity.
Use a different color for defects to make them easily recognizable from other types of work.
Step 2: Track Defects
In the book “The Economics of Software Quality”, defect tracking is used in more than 55% of military and defense software applications, but in less than 15% of IT, commercial, web or embedded applications. Given their prevention effectiveness, it is surprising that defect measurement is not used more often.
The second step of the defect management process is to track and analyze software defects on a regular basis.
The Throughput Run Chart displays the throughput of your team on a daily, weekly or monthly basis and compares its values over time. Use it to keep track of the number of defects you resolve and assess how trends have developed.
The Throughput Breakdown Chart can help you evaluate what type of work is demanding the bulk of the time. It is especially useful when you want to emphasize the high volume of defects you’re handling.
Step 3: Perform Defect Clustering Analysis
The third step of the defect management process is to identify the clusters of causes hindering your delivery times. To perform defect clustering analysis:
Evaluate the causes of defects and their resolution times.
Group the causes by clustering the defects which are similar in nature.
Capture the total cycle time for each group to discover which are the most critical causes of defects.
Report the defect clustering results at the Service Delivery Review, Risk Review, and Operations Review. That way, you’ll keep your attention on the defects with the highest impact on your development process, and you’ll emphasize the urgency to act upon them sooner.
Step 4: Define Relevant Policies
The goal of this step is to bring continuous improvement to your development practices and help you maintain a stable,
predictable workflow in the long run.
Introduce policies on how defects should be handled. A defect ticket should be attached to the work item that it belongs to. The work item should then remain in the same process state where the problem was first discovered.
Create a new ticket for your defect and associate it with the work item that spawned it.
Place the new ticket in the rework queue state (i.e. Development Queue).
Block the work item in the current process state (i.e. Testing).
By adding new items into your workflow, you count them against your WIP limit.
An excessive number of defects will then block the pull of new work until any quality issues are first resolved.
This approach will also help you avoid the common anti-pattern of moving work backward on your Kanban board.
Make sure your activities are not just targeted at finding defects after they have happened.
https://www.kanbanmaturitymodel.com
https://getnave.com/throughput-run-chart
https://www.infoq.com/articles/blockers-defects-process-improvement
|
process
|
kanban maturity model and the step process to defect management by nave introduce agile techniques like definition of done dod code reviews and pair programming to make sure you eliminate defects at an early development stage your team will become more self managed motivated and engaged shorter lead times lead to faster innovation greater customer satisfaction and substantial profitability focus on making sure the defects do not occur in the first place or at least the defects are significantly reduced before making their way into production defect management will help you increase the consistency predictability and effectiveness of your development efforts and it will ultimately reduce your delivery times the kanban maturity model describes the general practices of reducing defects to help you improve your business outcomes here is the step process which will enable you to achieve a more sustainable delivery of quality software step visualize defects step track defects step perform defect clustering analysis step define relevant policies step visualize defects we cannot manage what we cannot see increasing transparency is the first step towards reducing defects and raising awareness of the disruptive impact they may have when you’re testing a work item and a new defect is discovered create a separate ticket and mark it as a defect add the following essential information to the ticket add a brief description of the defect add a link to the work item the defect is associated with specify the reason that has caused the defect assign the people responsible for fixing the defect assign severity use a different color for defects to make them easily recognizable from other types of work step track defects in the book “the economics of software quality” defect tracking is used in more than of military and defense software applications but in less than of it commercial web or embedded applications given their prevention effectiveness it is surprising that defect measurement is not used more often the second step of the defect management process is to track and analyze software defects on a regular basis the throughput run chart displays the throughput of your team on a daily weekly or monthly basis and compares its values over time use it to keep track of the number of defects you resolve and assess how trends have developed the throughput breakdown chart can help you evaluate what type of work is demanding the bulk of the time it is especially useful when you want to emphasize the high volume of defects you’re handling step perform defect clustering analysis the third step of the defect management process is to identify the clusters of causes hindering your delivery times to perform defect clustering analysis evaluate the causes of defects and their resolution times group the causes by clustering the defects which are similar in nature capture the total cycle time for each group to discover which are the most critical causes of defects report the defect clustering results at the service delivery review risk review and operations review that way you’ll keep your attention on the defects with the highest impact on your development process and you’ll emphasize the urgency to act upon them sooner step define relevant policies the goal of this step is to bring continuous improvement to your development practices and help you maintain a stable predictable workflow in the long run introduce policies on how defects should be handled a defect ticket should be attached to the work item that it belongs to the work item should then remain in the same process state where the problem was first discovered create a new ticket for your defect and associate it with the work item that spawned it place the new ticket in the rework queue state i e development queue block the work item in the current process state i e testing by adding new items into your workflow you count them against your wip limit an excessive number of defects will then block the pull of new work until any quality issues are first resolved this approach will also help you avoid the common anti pattern of moving work backward on your kanban board make sure your activities are not just targeted at finding defects after they have happened
| 1
|
19,135
| 25,190,171,054
|
IssuesEvent
|
2022-11-11 23:14:27
|
hashgraph/hedera-json-rpc-relay
|
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
|
closed
|
Release 0.11.0
|
enhancement P1 process
|
### Problem
v0.11.0 is not fully released
### Solution
Deploy 0.11.0
### Alternatives
_No response_
|
1.0
|
Release 0.11.0 - ### Problem
v0.11.0 is not fully released
### Solution
Deploy 0.11.0
### Alternatives
_No response_
|
process
|
release problem is not fully released solution deploy alternatives no response
| 1
|
252,547
| 21,582,953,335
|
IssuesEvent
|
2022-05-02 20:52:03
|
damccorm/test-migration-target
|
https://api.github.com/repos/damccorm/test-migration-target
|
opened
|
ExternalPythonTransformTest.trivialPythonTransform flaky
|
bug test-failures cross-language P2
|
Example run: https://ci-beam.apache.org/job/beam_PreCommit_Java_Phrase/4806/testReport/junit/org.apache.beam.sdk.extensions.python/ExternalPythonTransformTest/trivialPythonTransform/
{noformat}
java.lang.RuntimeException: java.util.concurrent.TimeoutException: Timeout waiting for Python service startup after 16616 seconds.
at org.apache.beam.sdk.extensions.python.ExternalPythonTransform.expand(ExternalPythonTransform.java:107)
at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:548)
at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:482)
at org.apache.beam.sdk.values.PCollection.apply(PCollection.java:363)
at org.apache.beam.sdk.extensions.python.ExternalPythonTransformTest.trivialPythonTransform(ExternalPythonTransformTest.java:41)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
{noformat}
Imported from Jira [BEAM-14148](https://issues.apache.org/jira/browse/BEAM-14148). Original Jira may contain additional context.
Reported by: lcwik.
|
1.0
|
ExternalPythonTransformTest.trivialPythonTransform flaky - Example run: https://ci-beam.apache.org/job/beam_PreCommit_Java_Phrase/4806/testReport/junit/org.apache.beam.sdk.extensions.python/ExternalPythonTransformTest/trivialPythonTransform/
{noformat}
java.lang.RuntimeException: java.util.concurrent.TimeoutException: Timeout waiting for Python service startup after 16616 seconds.
at org.apache.beam.sdk.extensions.python.ExternalPythonTransform.expand(ExternalPythonTransform.java:107)
at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:548)
at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:482)
at org.apache.beam.sdk.values.PCollection.apply(PCollection.java:363)
at org.apache.beam.sdk.extensions.python.ExternalPythonTransformTest.trivialPythonTransform(ExternalPythonTransformTest.java:41)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
{noformat}
Imported from Jira [BEAM-14148](https://issues.apache.org/jira/browse/BEAM-14148). Original Jira may contain additional context.
Reported by: lcwik.
|
non_process
|
externalpythontransformtest trivialpythontransform flaky example run noformat java lang runtimeexception java util concurrent timeoutexception timeout waiting for python service startup after seconds at org apache beam sdk extensions python externalpythontransform expand externalpythontransform java at org apache beam sdk pipeline applyinternal pipeline java at org apache beam sdk pipeline applytransform pipeline java at org apache beam sdk values pcollection apply pcollection java at org apache beam sdk extensions python externalpythontransformtest trivialpythontransform externalpythontransformtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java noformat imported from jira original jira may contain additional context reported by lcwik
| 0
|
210,905
| 7,196,432,181
|
IssuesEvent
|
2018-02-05 02:54:12
|
wso2/message-broker
|
https://api.github.com/repos/wso2/message-broker
|
closed
|
Implement AMQP queue.delete frames
|
Complexity/Moderate Module/broker-amqp Priority/High Severity/Major Type/Improvement
|
**Description:**
<!-- Give a brief description of the issue -->
Implement queue.delete amqp frame to delete queue through an AMQP client
|
1.0
|
Implement AMQP queue.delete frames - **Description:**
<!-- Give a brief description of the issue -->
Implement queue.delete amqp frame to delete queue through an AMQP client
|
non_process
|
implement amqp queue delete frames description implement queue delete amqp frame to delete queue through an amqp client
| 0
|
91,870
| 3,863,516,295
|
IssuesEvent
|
2016-04-08 09:45:33
|
iamxavier/elmah
|
https://api.github.com/repos/iamxavier/elmah
|
closed
|
Add BCC to ELMAH
|
auto-migrated Priority-Medium Type-Enhancement
|
```
What new or enhanced feature are you proposing?
Add the capability to use BCC for emailing errorlogs.
What goal would this enhancement help you achieve?
This would allow us to hide a developers distribution list.
I have downloaded the source ELMAH-1.2-sp2-src and made changes to the attached
file for your review as well.
```
Original issue reported on code.google.com by `michael....@gmail.com` on 8 Aug 2012 at 6:54
Attachments:
* [ErrorMailModule.cs](https://storage.googleapis.com/google-code-attachments/elmah/issue-305/comment-0/ErrorMailModule.cs)
|
1.0
|
Add BCC to ELMAH - ```
What new or enhanced feature are you proposing?
Add the capability to use BCC for emailing errorlogs.
What goal would this enhancement help you achieve?
This would allow us to hide a developers distribution list.
I have downloaded the source ELMAH-1.2-sp2-src and made changes to the attached
file for your review as well.
```
Original issue reported on code.google.com by `michael....@gmail.com` on 8 Aug 2012 at 6:54
Attachments:
* [ErrorMailModule.cs](https://storage.googleapis.com/google-code-attachments/elmah/issue-305/comment-0/ErrorMailModule.cs)
|
non_process
|
add bcc to elmah what new or enhanced feature are you proposing add the capability to use bcc for emailing errorlogs what goal would this enhancement help you achieve this would allow us to hide a developers distribution list i have downloaded the source elmah src and made changes to the attached file for your review as well original issue reported on code google com by michael gmail com on aug at attachments
| 0
|
246,617
| 20,888,045,567
|
IssuesEvent
|
2022-03-23 08:08:01
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
opened
|
Fail to import one .csv file in one SAS attached table with permissions 'Query, Add'
|
🧪 testing :gear: tables :beetle: regression
|
**Storage Explorer Version:** 1.24.0-dev
**Build Number:** 20220322.15
**Branch:** main
**Platform/OS:** Windows 10/Linux Ubuntu 20.04/MacOS Monterey 12.1 (Apple M1 Pro)
**Architecture:** ia32\x64
**How Found:** Ad-hoc testing
**Regression From:** Previous release (1.22.1)
## Steps to Reproduce ##
1. Expand one storage account -> Tables -> Create a table.
2. Right click the table -> Click 'Get Shared Access Signature...'.
3. Select 'Query, Add' permissions -> Create the SAS URL and copy it.
4. Attach the table using the SAS URL.
5. Go to the attached table -> Try to import one .csv file.
6. Check whether succeeds to import the file.
## Expected Experience ##
Succeed to import the file.
## Actual Experience ##
Fail to import the file.

|
1.0
|
Fail to import one .csv file in one SAS attached table with permissions 'Query, Add' - **Storage Explorer Version:** 1.24.0-dev
**Build Number:** 20220322.15
**Branch:** main
**Platform/OS:** Windows 10/Linux Ubuntu 20.04/MacOS Monterey 12.1 (Apple M1 Pro)
**Architecture:** ia32\x64
**How Found:** Ad-hoc testing
**Regression From:** Previous release (1.22.1)
## Steps to Reproduce ##
1. Expand one storage account -> Tables -> Create a table.
2. Right click the table -> Click 'Get Shared Access Signature...'.
3. Select 'Query, Add' permissions -> Create the SAS URL and copy it.
4. Attach the table using the SAS URL.
5. Go to the attached table -> Try to import one .csv file.
6. Check whether succeeds to import the file.
## Expected Experience ##
Succeed to import the file.
## Actual Experience ##
Fail to import the file.

|
non_process
|
fail to import one csv file in one sas attached table with permissions query add storage explorer version dev build number branch main platform os windows linux ubuntu macos monterey apple pro architecture how found ad hoc testing regression from previous release steps to reproduce expand one storage account tables create a table right click the table click get shared access signature select query add permissions create the sas url and copy it attach the table using the sas url go to the attached table try to import one csv file check whether succeeds to import the file expected experience succeed to import the file actual experience fail to import the file
| 0
|
1,941
| 4,769,492,271
|
IssuesEvent
|
2016-10-26 12:46:13
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
closed
|
Increase amount of alba statistics in GUI to make it more transparent
|
process_wontfix type_enhancement
|
Currently you see the preset in use and policy in use.
But this is too little information to keep the system rolling.
A vital part to know if your backend is still ok, is to know:
* How much data has been written in a certain policy
* How healthy is my policy? (e.g. policy is will stop working after 6 disks or 2 nodes stop working)
* ...
|
1.0
|
Increase amount of alba statistics in GUI to make it more transparent - Currently you see the preset in use and policy in use.
But this is too little information to keep the system rolling.
A vital part to know if your backend is still ok, is to know:
* How much data has been written in a certain policy
* How healthy is my policy? (e.g. policy is will stop working after 6 disks or 2 nodes stop working)
* ...
|
process
|
increase amount of alba statistics in gui to make it more transparent currently you see the preset in use and policy in use but this is too little information to keep the system rolling a vital part to know if your backend is still ok is to know how much data has been written in a certain policy how healthy is my policy e g policy is will stop working after disks or nodes stop working
| 1
|
18,770
| 10,224,254,861
|
IssuesEvent
|
2019-08-16 12:12:27
|
ppy/osu
|
https://api.github.com/repos/ppy/osu
|
closed
|
Seeking in editor is laggy
|
editor pending review performance
|
**Describe the bug:**
Map seeking is very laggy in the editor.
**Screenshots or videos showing encountered issue:**
https://streamable.com/0my1e
**osu!lazer version:**
2019.807.0.0
**Logs:**
[runtime.log](https://github.com/ppy/osu/files/3482264/runtime.log)
[performance.log](https://github.com/ppy/osu/files/3482266/performance.log)
[network.log](https://github.com/ppy/osu/files/3482267/network.log)
[database.log](https://github.com/ppy/osu/files/3482268/database.log)
|
True
|
Seeking in editor is laggy - **Describe the bug:**
Map seeking is very laggy in the editor.
**Screenshots or videos showing encountered issue:**
https://streamable.com/0my1e
**osu!lazer version:**
2019.807.0.0
**Logs:**
[runtime.log](https://github.com/ppy/osu/files/3482264/runtime.log)
[performance.log](https://github.com/ppy/osu/files/3482266/performance.log)
[network.log](https://github.com/ppy/osu/files/3482267/network.log)
[database.log](https://github.com/ppy/osu/files/3482268/database.log)
|
non_process
|
seeking in editor is laggy describe the bug map seeking is very laggy in the editor screenshots or videos showing encountered issue osu lazer version logs
| 0
|
69,172
| 22,264,137,647
|
IssuesEvent
|
2022-06-10 05:23:32
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
opened
|
"Mark as read" should not be shown after "Quick reply" from notification
|
T-Defect
|
### Steps to reproduce

1. Receive an incoming message notification
2. Reply from notification
### Outcome
#### What did you expect?
Mark as read should not be offered after reply.
#### What happened instead?
Mark as read is offered after reply.
### Your phone model
OP3T
### Operating system version
Lineage+microG 18.1 on Android 11
### Application version and app store
Element 1.4.16
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No
|
1.0
|
"Mark as read" should not be shown after "Quick reply" from notification - ### Steps to reproduce

1. Receive an incoming message notification
2. Reply from notification
### Outcome
#### What did you expect?
Mark as read should not be offered after reply.
#### What happened instead?
Mark as read is offered after reply.
### Your phone model
OP3T
### Operating system version
Lineage+microG 18.1 on Android 11
### Application version and app store
Element 1.4.16
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No
|
non_process
|
mark as read should not be shown after quick reply from notification steps to reproduce receive an incoming message notification reply from notification outcome what did you expect mark as read should not be offered after reply what happened instead mark as read is offered after reply your phone model operating system version lineage microg on android application version and app store element homeserver no response will you send logs no are you willing to provide a pr no
| 0
|
29,812
| 24,301,302,790
|
IssuesEvent
|
2022-09-29 14:03:25
|
ephios-dev/ephios
|
https://api.github.com/repos/ephios-dev/ephios
|
closed
|
Derive version from git tag
|
[P] nice to have [C] infrastructure
|
To prevent version bump commits it would be great to determine the target version from the current git tag. Poetry doesn't support that natively, but it will be possible with poetry 1.2 once it is released. Someone is building a plugin for that, the progess is tracked in https://github.com/mtkennerly/poetry-dynamic-versioning/issues/39
|
1.0
|
Derive version from git tag - To prevent version bump commits it would be great to determine the target version from the current git tag. Poetry doesn't support that natively, but it will be possible with poetry 1.2 once it is released. Someone is building a plugin for that, the progess is tracked in https://github.com/mtkennerly/poetry-dynamic-versioning/issues/39
|
non_process
|
derive version from git tag to prevent version bump commits it would be great to determine the target version from the current git tag poetry doesn t support that natively but it will be possible with poetry once it is released someone is building a plugin for that the progess is tracked in
| 0
|
430,125
| 12,440,737,397
|
IssuesEvent
|
2020-05-26 12:32:36
|
BgeeDB/bgee_apps
|
https://api.github.com/repos/BgeeDB/bgee_apps
|
closed
|
Provide anatomical homologies
|
priority: medium ui enhancement
|
In GitLab by @marcrr on Oct 15, 2015, 11:48
Provide ancestral_taxa_homology_annotations.tsv file on Bgee homepage.
Provide small file with only multiple-entity annotations, under name such as "non trivial homologies".
|
1.0
|
Provide anatomical homologies - In GitLab by @marcrr on Oct 15, 2015, 11:48
Provide ancestral_taxa_homology_annotations.tsv file on Bgee homepage.
Provide small file with only multiple-entity annotations, under name such as "non trivial homologies".
|
non_process
|
provide anatomical homologies in gitlab by marcrr on oct provide ancestral taxa homology annotations tsv file on bgee homepage provide small file with only multiple entity annotations under name such as non trivial homologies
| 0
|
2,826
| 5,780,515,406
|
IssuesEvent
|
2017-04-29 01:10:51
|
gaocegege/maintainer
|
https://api.github.com/repos/gaocegege/maintainer
|
closed
|
Fix Docker image because of gh-md-toc
|
priority/P1 process/wip type/bug
|
Now maintainer supports `readme toc`, but the related dependencies is not installed in Dockerfile.
|
1.0
|
Fix Docker image because of gh-md-toc - Now maintainer supports `readme toc`, but the related dependencies is not installed in Dockerfile.
|
process
|
fix docker image because of gh md toc now maintainer supports readme toc but the related dependencies is not installed in dockerfile
| 1
|
32,567
| 15,439,661,901
|
IssuesEvent
|
2021-03-08 01:01:03
|
labstack/echo
|
https://api.github.com/repos/labstack/echo
|
closed
|
Performance regression with echo `4.2.0`
|
performance
|
### Issue Description
We have just updated to echo 4.2.0 and have seen a relatively high performance regression with that release. With bisecting
it seems like commit 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4 has introduced a perfomance penalty of about 30-40%.
The following benchstat has been generated with https://github.com/vishr/web-framework-benchmark comparing commit
cf2fcad01b833d4bac6deb916940759cef3fb376 to 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4.
```
name old time/op new time/op delta
EchoStatic-8 29.5µs ± 2% 38.1µs ± 2% +29.26% (p=0.000 n=9+8)
EchoGitHubAPI-8 46.2µs ± 1% 62.8µs ± 1% +35.93% (p=0.000 n=8+9)
EchoGplusAPI-8 2.29µs ± 4% 3.23µs ± 1% +41.25% (p=0.000 n=9+9)
EchoParseAPI-8 3.99µs ± 1% 5.67µs ± 1% +42.00% (p=0.000 n=9+9)
```
### Steps to reproduce
Run https://github.com/vishr/web-framework-benchmark benchmark with commit 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4 and cf2fcad01b833d4bac6deb916940759cef3fb376
### Version/commit
`4.2.0` first commit with regression 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4
|
True
|
Performance regression with echo `4.2.0` - ### Issue Description
We have just updated to echo 4.2.0 and have seen a relatively high performance regression with that release. With bisecting
it seems like commit 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4 has introduced a perfomance penalty of about 30-40%.
The following benchstat has been generated with https://github.com/vishr/web-framework-benchmark comparing commit
cf2fcad01b833d4bac6deb916940759cef3fb376 to 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4.
```
name old time/op new time/op delta
EchoStatic-8 29.5µs ± 2% 38.1µs ± 2% +29.26% (p=0.000 n=9+8)
EchoGitHubAPI-8 46.2µs ± 1% 62.8µs ± 1% +35.93% (p=0.000 n=8+9)
EchoGplusAPI-8 2.29µs ± 4% 3.23µs ± 1% +41.25% (p=0.000 n=9+9)
EchoParseAPI-8 3.99µs ± 1% 5.67µs ± 1% +42.00% (p=0.000 n=9+9)
```
### Steps to reproduce
Run https://github.com/vishr/web-framework-benchmark benchmark with commit 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4 and cf2fcad01b833d4bac6deb916940759cef3fb376
### Version/commit
`4.2.0` first commit with regression 2d79ff3d6c27592d01a820d93d0fd7887a87a4b4
|
non_process
|
performance regression with echo issue description we have just updated to echo and have seen a relatively high performance regression with that release with bisecting it seems like commit has introduced a perfomance penalty of about the following benchstat has been generated with comparing commit to name old time op new time op delta echostatic ± ± p n echogithubapi ± ± p n echogplusapi ± ± p n echoparseapi ± ± p n steps to reproduce run benchmark with commit and version commit first commit with regression
| 0
|
13,642
| 16,331,832,916
|
IssuesEvent
|
2021-05-12 10:10:00
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Merge GO:0052413 metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction & related
|
multi-species process
|
Pectin is only found in plants.
~Plants are not symbionts of other organisms.~ But there are no examples I know of where a plant synthesises or degrades another plant's cell wall.
So those 4 terms should be merged or obsoleted:
* GO:0052413 metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction
* GO:0052341 catabolism by organism of cell wall pectin in other organism involved in symbiotic interaction
* GO:0052412 metabolism by host of symbiont cell wall pectin
* GO:0052357 catabolism by host of symbiont cell wall pectin
Those can remain:
* GO:0052179 metabolism by symbiont of host cell wall pectin
* GO:0052011 catabolism by symbiont of host cell wall pectin
@ValWood @mgiglio99 @dsiegele any preference for merge/obsolete?
There are no annotations (I moved annotations from GO:0052413 to GO:0052011)
Thanks, Pascale
|
1.0
|
Merge GO:0052413 metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction & related - Pectin is only found in plants.
~Plants are not symbionts of other organisms.~ But there are no examples I know of where a plant synthesises or degrades another plant's cell wall.
So those 4 terms should be merged or obsoleted:
* GO:0052413 metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction
* GO:0052341 catabolism by organism of cell wall pectin in other organism involved in symbiotic interaction
* GO:0052412 metabolism by host of symbiont cell wall pectin
* GO:0052357 catabolism by host of symbiont cell wall pectin
Those can remain:
* GO:0052179 metabolism by symbiont of host cell wall pectin
* GO:0052011 catabolism by symbiont of host cell wall pectin
@ValWood @mgiglio99 @dsiegele any preference for merge/obsolete?
There are no annotations (I moved annotations from GO:0052413 to GO:0052011)
Thanks, Pascale
|
process
|
merge go metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction related pectin is only found in plants plants are not symbionts of other organisms but there are no examples i know of where a plant synthesises or degrades another plant s cell wall so those terms should be merged or obsoleted go metabolism by organism of cell wall pectin in other organism involved in symbiotic interaction go catabolism by organism of cell wall pectin in other organism involved in symbiotic interaction go metabolism by host of symbiont cell wall pectin go catabolism by host of symbiont cell wall pectin those can remain go metabolism by symbiont of host cell wall pectin go catabolism by symbiont of host cell wall pectin valwood dsiegele any preference for merge obsolete there are no annotations i moved annotations from go to go thanks pascale
| 1
|
558
| 3,020,669,461
|
IssuesEvent
|
2015-07-31 09:34:28
|
DynareTeam/dynare
|
https://api.github.com/repos/DynareTeam/dynare
|
closed
|
Make use_calibration option compatible with ML
|
estimation preprocessor
|
ML currently requires to specify explicit starting values inside of the ```estimated_params``` block, rendering the ```estimated_params_init``` block useless.
|
1.0
|
Make use_calibration option compatible with ML - ML currently requires to specify explicit starting values inside of the ```estimated_params``` block, rendering the ```estimated_params_init``` block useless.
|
process
|
make use calibration option compatible with ml ml currently requires to specify explicit starting values inside of the estimated params block rendering the estimated params init block useless
| 1
|
21,597
| 30,000,989,022
|
IssuesEvent
|
2023-06-26 09:16:13
|
metallb/metallb
|
https://api.github.com/repos/metallb/metallb
|
reopened
|
Update MetalLB info in CNCF landscape
|
process lifecycle-stale
|
The information about MetalLB in the CNCF [landscape](https://landscape.cncf.io/) is outdated.
https://github.com/cncf/landscape/blob/1873345e220ab784e20435768ad59b7b7fc3659a/landscape.yml#L1909-L1915
We should open a PR to address the following:
- The project's homepage should be metallb.io (blocked by https://github.com/metallb/metallb/issues/691).
- The logo should be updated in case we get a new design from the CNCF.
- The Twitter handle (which is currently Dave's personal handle) should be either removed or updated in case we decide to manage a MetalLB Twitter account.
- The Crunchbase link should be updated to `https://www.crunchbase.com/organization/cloud-native-computing-foundation`.
|
1.0
|
Update MetalLB info in CNCF landscape - The information about MetalLB in the CNCF [landscape](https://landscape.cncf.io/) is outdated.
https://github.com/cncf/landscape/blob/1873345e220ab784e20435768ad59b7b7fc3659a/landscape.yml#L1909-L1915
We should open a PR to address the following:
- The project's homepage should be metallb.io (blocked by https://github.com/metallb/metallb/issues/691).
- The logo should be updated in case we get a new design from the CNCF.
- The Twitter handle (which is currently Dave's personal handle) should be either removed or updated in case we decide to manage a MetalLB Twitter account.
- The Crunchbase link should be updated to `https://www.crunchbase.com/organization/cloud-native-computing-foundation`.
|
process
|
update metallb info in cncf landscape the information about metallb in the cncf is outdated we should open a pr to address the following the project s homepage should be metallb io blocked by the logo should be updated in case we get a new design from the cncf the twitter handle which is currently dave s personal handle should be either removed or updated in case we decide to manage a metallb twitter account the crunchbase link should be updated to
| 1
|
18,950
| 24,910,734,528
|
IssuesEvent
|
2022-10-29 20:40:05
|
TUM-Dev/NavigaTUM
|
https://api.github.com/repos/TUM-Dev/NavigaTUM
|
closed
|
[Bug] WZW Bereich Mitte
|
bug webform delete-after-processing
|
Das Gewächshauslaborzentrum (und auch andere Gebäude) ist falsch verortet. Es liegt rund 2 km außerhalb von Freising.
Ferner ist der Begriff WZW veraltet, denn das Wissenschaftszentrum Weihenstephan ist in der TUM School of Life Sciences aufgegangen.
|
1.0
|
[Bug] WZW Bereich Mitte - Das Gewächshauslaborzentrum (und auch andere Gebäude) ist falsch verortet. Es liegt rund 2 km außerhalb von Freising.
Ferner ist der Begriff WZW veraltet, denn das Wissenschaftszentrum Weihenstephan ist in der TUM School of Life Sciences aufgegangen.
|
process
|
wzw bereich mitte das gewächshauslaborzentrum und auch andere gebäude ist falsch verortet es liegt rund km außerhalb von freising ferner ist der begriff wzw veraltet denn das wissenschaftszentrum weihenstephan ist in der tum school of life sciences aufgegangen
| 1
|
129,347
| 27,447,624,946
|
IssuesEvent
|
2023-03-02 15:16:27
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
closed
|
Sourcegraph for VSCode extension doesn't render search page
|
bug vscode-extension team/code-exploration
|
- **Sourcegraph version:** <!-- the version of Sourcegraph or "Sourcegraph.com" --> VSCE v2.2.14
- **Platform information:** <!-- OS version, cloud provider, web browser version, Docker version, etc., depending on the issue --> VSCode v1.74.2
#### Steps to reproduce:
1. Install the latest VSCE extension (v2.2.14 for the time being).
2. Open it by clicking the Sourcegraph icon in the sidebar.
3. Check out the search page.
#### Expected behavior:
The search page is rendered (v2.2.13).
<img width="1524" alt="Screenshot 2023-01-09 at 15 54 25" src="https://user-images.githubusercontent.com/25318659/211324192-4547d8ba-7340-4676-90e8-e5a440d440d0.png">
#### Actual behavior:
The search page is not rendered (v2.2.14).
<img width="1524" alt="Screenshot 2023-01-09 at 15 53 06" src="https://user-images.githubusercontent.com/25318659/211324254-04c3cd0f-4f88-433b-9d88-0a170f4729d7.png">
The following errors are thrown in the console in dev mode.
<img width="1215" alt="Screenshot 2023-01-09 at 12 12 05" src="https://user-images.githubusercontent.com/25318659/211324352-2fcf53ce-2495-402e-b467-c2f85555fa0b.png">
If you would like immediate help on this, please email support@sourcegraph.com (you can still create the issue, but there are no [SLAs](https://about.sourcegraph.com/support/) on issues like there are for support requests).
/cc @felixfbecker
|
2.0
|
Sourcegraph for VSCode extension doesn't render search page - - **Sourcegraph version:** <!-- the version of Sourcegraph or "Sourcegraph.com" --> VSCE v2.2.14
- **Platform information:** <!-- OS version, cloud provider, web browser version, Docker version, etc., depending on the issue --> VSCode v1.74.2
#### Steps to reproduce:
1. Install the latest VSCE extension (v2.2.14 for the time being).
2. Open it by clicking the Sourcegraph icon in the sidebar.
3. Check out the search page.
#### Expected behavior:
The search page is rendered (v2.2.13).
<img width="1524" alt="Screenshot 2023-01-09 at 15 54 25" src="https://user-images.githubusercontent.com/25318659/211324192-4547d8ba-7340-4676-90e8-e5a440d440d0.png">
#### Actual behavior:
The search page is not rendered (v2.2.14).
<img width="1524" alt="Screenshot 2023-01-09 at 15 53 06" src="https://user-images.githubusercontent.com/25318659/211324254-04c3cd0f-4f88-433b-9d88-0a170f4729d7.png">
The following errors are thrown in the console in dev mode.
<img width="1215" alt="Screenshot 2023-01-09 at 12 12 05" src="https://user-images.githubusercontent.com/25318659/211324352-2fcf53ce-2495-402e-b467-c2f85555fa0b.png">
If you would like immediate help on this, please email support@sourcegraph.com (you can still create the issue, but there are no [SLAs](https://about.sourcegraph.com/support/) on issues like there are for support requests).
/cc @felixfbecker
|
non_process
|
sourcegraph for vscode extension doesn t render search page sourcegraph version vsce platform information vscode steps to reproduce install the latest vsce extension for the time being open it by clicking the sourcegraph icon in the sidebar check out the search page expected behavior the search page is rendered img width alt screenshot at src actual behavior the search page is not rendered img width alt screenshot at src the following errors are thrown in the console in dev mode img width alt screenshot at src if you would like immediate help on this please email support sourcegraph com you can still create the issue but there are no on issues like there are for support requests cc felixfbecker
| 0
|
5,120
| 7,890,513,098
|
IssuesEvent
|
2018-06-28 09:03:06
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Memory usage (external) reports a negative value?
|
memory process
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: v10.5.0
Platform: Linux PZW2 4.9.35+ #1014 Fri Jun 30 14:34:49 BST 2017 armv6l GNU/Linux
Subsystem: **?d**
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: v10.5.0
* **Platform**: Linux PZW2 4.9.35+ #1014 Fri Jun 30 14:34:49 BST 2017 armv6l GNU/Linux
* **Subsystem**: REPL in Node.js
<!-- Enter your issue details below this comment. -->
Running some http server process in node.js in REPL environment (under tmux), the function: process.memoryUsage() function can return a negative value. Is that a bug? The process (server) was run for many days, and I sometimes use process.memoryUsage() to check the values and I found:
```
> process.memoryUsage()
{ rss: 33452032,
heapTotal: 9031680,
heapUsed: 5112064,
external: -503806 }
```
>
The external memory is negative, but it was still postive 2 days ago when I checked. Notes, I run the node.js with flags:
` --expose-gc --max-old-space-size=256`
This is running on a Raspberry Pi ZERO W.
****
|
1.0
|
Memory usage (external) reports a negative value? - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: v10.5.0
Platform: Linux PZW2 4.9.35+ #1014 Fri Jun 30 14:34:49 BST 2017 armv6l GNU/Linux
Subsystem: **?d**
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: v10.5.0
* **Platform**: Linux PZW2 4.9.35+ #1014 Fri Jun 30 14:34:49 BST 2017 armv6l GNU/Linux
* **Subsystem**: REPL in Node.js
<!-- Enter your issue details below this comment. -->
Running some http server process in node.js in REPL environment (under tmux), the function: process.memoryUsage() function can return a negative value. Is that a bug? The process (server) was run for many days, and I sometimes use process.memoryUsage() to check the values and I found:
```
> process.memoryUsage()
{ rss: 33452032,
heapTotal: 9031680,
heapUsed: 5112064,
external: -503806 }
```
>
The external memory is negative, but it was still postive 2 days ago when I checked. Notes, I run the node.js with flags:
` --expose-gc --max-old-space-size=256`
This is running on a Raspberry Pi ZERO W.
****
|
process
|
memory usage external reports a negative value thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version platform linux fri jun bst gnu linux subsystem d if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version platform linux fri jun bst gnu linux subsystem repl in node js running some http server process in node js in repl environment under tmux the function process memoryusage function can return a negative value is that a bug the process server was run for many days and i sometimes use process memoryusage to check the values and i found process memoryusage rss heaptotal heapused external the external memory is negative but it was still postive days ago when i checked notes i run the node js with flags expose gc max old space size this is running on a raspberry pi zero w
| 1
|
13,477
| 15,987,347,271
|
IssuesEvent
|
2021-04-19 00:13:32
|
shirou/gopsutil
|
https://api.github.com/repos/shirou/gopsutil
|
closed
|
Get Windows Processes Name Take Too Much Time
|
os:windows package:process
|
**Describe the bug**
It takes too much time to get the windows process name. In general, my windows has 260 processes, and When I get the names of these processes, it usually takes 8-9 seconds. It seems execute func getFromSnapProcess(pid int32) (int32, int32, string, error) 260*260 times, not 260 times.
**To Reproduce**
```go
processInfos, _ := process.Processes()
for _, processInfo := range processInfos {
name, _ := processInfo.Name()
}
```
**Expected behavior**
It should execute func getFromSnapProcess(pid int32) (int32, int32, string, error) 260 times and takse less than 1s
**Environment (please complete the following information):**
- [ ] Windows: Microsoft Windows [version 10.0.18363.1440]
**Additional context**
command: go build
C:\WINDOWS\system32>go env
set GO111MODULE=on
set GOARCH=amd64
set GOBIN=
set GOCACHE=C:\Users\xiaotao.hu\AppData\Local\go-build
set GOENV=C:\Users\xiaotao.hu\AppData\Roaming\go\env
set GOEXE=.exe
set GOFLAGS=
set GOHOSTARCH=amd64
set GOHOSTOS=windows
set GOINSECURE=
set GOMODCACHE=C:\Users\xiaotao.hu\go\pkg\mod
set GONOPROXY=
set GONOSUMDB=
set GOOS=windows
set GOPATH=C:\Users\xiaotao.hu\go
set GOPRIVATE=
set GOPROXY=https://goproxy.io
set GOROOT=c:\go
set GOSUMDB=sum.golang.org
set GOTMPDIR=
set GOTOOLDIR=c:\go\pkg\tool\windows_amd64
set GCCGO=gccgo
set AR=ar
set CC=x86_64-w64-mingw32-gcc
set CXX=g++
set CGO_ENABLED=1
set GOMOD=NUL
set CGO_CFLAGS=-g -O2
set CGO_CPPFLAGS=
set CGO_CXXFLAGS=-g -O2
set CGO_FFLAGS=-g -O2
set CGO_LDFLAGS=-g -O2
set PKG_CONFIG=pkg-config
set GOGCCFLAGS=-m64 -mthreads -fmessage-length=0 -fdebug-prefix-map=C:\Users\xiaotao.hu\AppData\Local\Temp\go-build909459346=/tmp/go-build -gno-record-gcc-switches
|
1.0
|
Get Windows Processes Name Take Too Much Time - **Describe the bug**
It takes too much time to get the windows process name. In general, my windows has 260 processes, and When I get the names of these processes, it usually takes 8-9 seconds. It seems execute func getFromSnapProcess(pid int32) (int32, int32, string, error) 260*260 times, not 260 times.
**To Reproduce**
```go
processInfos, _ := process.Processes()
for _, processInfo := range processInfos {
name, _ := processInfo.Name()
}
```
**Expected behavior**
It should execute func getFromSnapProcess(pid int32) (int32, int32, string, error) 260 times and takse less than 1s
**Environment (please complete the following information):**
- [ ] Windows: Microsoft Windows [version 10.0.18363.1440]
**Additional context**
command: go build
C:\WINDOWS\system32>go env
set GO111MODULE=on
set GOARCH=amd64
set GOBIN=
set GOCACHE=C:\Users\xiaotao.hu\AppData\Local\go-build
set GOENV=C:\Users\xiaotao.hu\AppData\Roaming\go\env
set GOEXE=.exe
set GOFLAGS=
set GOHOSTARCH=amd64
set GOHOSTOS=windows
set GOINSECURE=
set GOMODCACHE=C:\Users\xiaotao.hu\go\pkg\mod
set GONOPROXY=
set GONOSUMDB=
set GOOS=windows
set GOPATH=C:\Users\xiaotao.hu\go
set GOPRIVATE=
set GOPROXY=https://goproxy.io
set GOROOT=c:\go
set GOSUMDB=sum.golang.org
set GOTMPDIR=
set GOTOOLDIR=c:\go\pkg\tool\windows_amd64
set GCCGO=gccgo
set AR=ar
set CC=x86_64-w64-mingw32-gcc
set CXX=g++
set CGO_ENABLED=1
set GOMOD=NUL
set CGO_CFLAGS=-g -O2
set CGO_CPPFLAGS=
set CGO_CXXFLAGS=-g -O2
set CGO_FFLAGS=-g -O2
set CGO_LDFLAGS=-g -O2
set PKG_CONFIG=pkg-config
set GOGCCFLAGS=-m64 -mthreads -fmessage-length=0 -fdebug-prefix-map=C:\Users\xiaotao.hu\AppData\Local\Temp\go-build909459346=/tmp/go-build -gno-record-gcc-switches
|
process
|
get windows processes name take too much time describe the bug it takes too much time to get the windows process name in general my windows has processes and when i get the names of these processes it usually takes seconds it seems execute func getfromsnapprocess pid string error times not times to reproduce go processinfos process processes for processinfo range processinfos name processinfo name expected behavior it should execute func getfromsnapprocess pid string error times and takse less than environment please complete the following information windows microsoft windows additional context command go build c windows go env set on set goarch set gobin set gocache c users xiaotao hu appdata local go build set goenv c users xiaotao hu appdata roaming go env set goexe exe set goflags set gohostarch set gohostos windows set goinsecure set gomodcache c users xiaotao hu go pkg mod set gonoproxy set gonosumdb set goos windows set gopath c users xiaotao hu go set goprivate set goproxy set goroot c go set gosumdb sum golang org set gotmpdir set gotooldir c go pkg tool windows set gccgo gccgo set ar ar set cc gcc set cxx g set cgo enabled set gomod nul set cgo cflags g set cgo cppflags set cgo cxxflags g set cgo fflags g set cgo ldflags g set pkg config pkg config set gogccflags mthreads fmessage length fdebug prefix map c users xiaotao hu appdata local temp go tmp go build gno record gcc switches
| 1
|
88,572
| 17,610,951,097
|
IssuesEvent
|
2021-08-18 01:05:50
|
jhona-tam/Farmacia
|
https://api.github.com/repos/jhona-tam/Farmacia
|
closed
|
Usando libreria
|
documentation code
|
Usamos librería `animate.css` para una mejor presentacion en los movimientos.
|
1.0
|
Usando libreria - Usamos librería `animate.css` para una mejor presentacion en los movimientos.
|
non_process
|
usando libreria usamos librería animate css para una mejor presentacion en los movimientos
| 0
|
274,696
| 8,564,351,057
|
IssuesEvent
|
2018-11-09 16:30:01
|
mozilla/MozDef
|
https://api.github.com/repos/mozilla/MozDef
|
closed
|
Feature flagging
|
category:enhancement priority:medium
|
#### ISSUE TYPE
- Feature Idea
#### COMPONENT NAME
Meteor/Mozdef
#### CONFIGURATION
Will add capability to do feature flagging to do things like:
- turn off block IP (already exists)
- turn off the ip blocklist
- turn off the fqdn blocklist
- turn off individual vizualizations
#### DESCRIPTION
Currently it's an all or nothing experience, but maybe folks just want to use Mozdef to track investigations/incidents, or maybe they just want alerts, or just visualizations. This will allow folks to turn on and off the bits they want.
This will start with the UI first, then expand to other components if possible.
|
1.0
|
Feature flagging - #### ISSUE TYPE
- Feature Idea
#### COMPONENT NAME
Meteor/Mozdef
#### CONFIGURATION
Will add capability to do feature flagging to do things like:
- turn off block IP (already exists)
- turn off the ip blocklist
- turn off the fqdn blocklist
- turn off individual vizualizations
#### DESCRIPTION
Currently it's an all or nothing experience, but maybe folks just want to use Mozdef to track investigations/incidents, or maybe they just want alerts, or just visualizations. This will allow folks to turn on and off the bits they want.
This will start with the UI first, then expand to other components if possible.
|
non_process
|
feature flagging issue type feature idea component name meteor mozdef configuration will add capability to do feature flagging to do things like turn off block ip already exists turn off the ip blocklist turn off the fqdn blocklist turn off individual vizualizations description currently it s an all or nothing experience but maybe folks just want to use mozdef to track investigations incidents or maybe they just want alerts or just visualizations this will allow folks to turn on and off the bits they want this will start with the ui first then expand to other components if possible
| 0
|
59,958
| 14,679,657,881
|
IssuesEvent
|
2020-12-31 07:38:04
|
google/xls
|
https://api.github.com/repos/google/xls
|
closed
|
Ubuntu 18.04: undeclared inclusion(s) in rule '@com_google_absl//absl/base:log_severity'
|
build
|
Here's the error:
```
❯ bazel build //xls/tools:codegen_main
INFO: Analyzed target //xls/tools:codegen_main (0 packages loaded, 0 targets configured).
INFO: Found 1 target...
ERROR: /home/lachlan/.cache/bazel/_bazel_lachlan/d8365cac006c3b38c2028f7288f6928e/external/com_google_absl/absl/base/BUILD.bazel:54:11: undeclared inclusion(s) in rule '@com_google_absl//absl/base:log_severity':
this rule is missing dependency declarations for the following files included by 'com_google_absl/absl/base/log_severity.cc':
'/usr/lib/gcc/x86_64-linux-gnu/9/include/stddef.h'
'/usr/lib/gcc/x86_64-linux-gnu/9/include/stdarg.h'
'/usr/lib/gcc/x86_64-linux-gnu/9/include/stdint.h'
'/usr/lib/gcc/x86_64-linux-gnu/9/include/limits.h'
'/usr/lib/gcc/x86_64-linux-gnu/9/include/syslimits.h'
Target //xls/tools:codegen_main failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 3.369s, Critical Path: 3.01s
INFO: 13 processes: 13 internal.
FAILED: Build did NOT complete successfully
```
I'm using ubuntu running within wsl.
|
1.0
|
Ubuntu 18.04: undeclared inclusion(s) in rule '@com_google_absl//absl/base:log_severity' - Here's the error:
```
❯ bazel build //xls/tools:codegen_main
INFO: Analyzed target //xls/tools:codegen_main (0 packages loaded, 0 targets configured).
INFO: Found 1 target...
ERROR: /home/lachlan/.cache/bazel/_bazel_lachlan/d8365cac006c3b38c2028f7288f6928e/external/com_google_absl/absl/base/BUILD.bazel:54:11: undeclared inclusion(s) in rule '@com_google_absl//absl/base:log_severity':
this rule is missing dependency declarations for the following files included by 'com_google_absl/absl/base/log_severity.cc':
'/usr/lib/gcc/x86_64-linux-gnu/9/include/stddef.h'
'/usr/lib/gcc/x86_64-linux-gnu/9/include/stdarg.h'
'/usr/lib/gcc/x86_64-linux-gnu/9/include/stdint.h'
'/usr/lib/gcc/x86_64-linux-gnu/9/include/limits.h'
'/usr/lib/gcc/x86_64-linux-gnu/9/include/syslimits.h'
Target //xls/tools:codegen_main failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 3.369s, Critical Path: 3.01s
INFO: 13 processes: 13 internal.
FAILED: Build did NOT complete successfully
```
I'm using ubuntu running within wsl.
|
non_process
|
ubuntu undeclared inclusion s in rule com google absl absl base log severity here s the error ❯ bazel build xls tools codegen main info analyzed target xls tools codegen main packages loaded targets configured info found target error home lachlan cache bazel bazel lachlan external com google absl absl base build bazel undeclared inclusion s in rule com google absl absl base log severity this rule is missing dependency declarations for the following files included by com google absl absl base log severity cc usr lib gcc linux gnu include stddef h usr lib gcc linux gnu include stdarg h usr lib gcc linux gnu include stdint h usr lib gcc linux gnu include limits h usr lib gcc linux gnu include syslimits h target xls tools codegen main failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes internal failed build did not complete successfully i m using ubuntu running within wsl
| 0
|
22,589
| 31,814,482,806
|
IssuesEvent
|
2023-09-13 19:20:52
|
googleapis/sdk-platform-java
|
https://api.github.com/repos/googleapis/sdk-platform-java
|
closed
|
Dependency Dashboard
|
type: process priority: p4
|
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more.
## Rate-Limited
These updates are currently rate-limited. Click on a checkbox below to force their creation now.
- [x] <!-- unlimit-branch=renovate/org.jacoco-jacoco-maven-plugin-0.x -->build(deps): update dependency org.jacoco:jacoco-maven-plugin to v0.8.9
- [x] <!-- unlimit-branch=renovate/netty-dependencies -->deps: update netty dependencies to v4.1.91.final (`io.netty:netty-buffer`, `io.netty:netty-codec`, `io.netty:netty-codec-http`, `io.netty:netty-codec-http2`, `io.netty:netty-codec-socks`, `io.netty:netty-common`, `io.netty:netty-handler`, `io.netty:netty-handler-proxy`, `io.netty:netty-resolver`, `io.netty:netty-transport`, `io.netty:netty-transport-native-unix-common`)
- [ ] <!-- unlimit-branch=renovate/attrs-22.x -->deps: update dependency attrs to v22.2.0
- [ ] <!-- unlimit-branch=renovate/charset-normalizer-2.x -->deps: update dependency charset-normalizer to v2.1.1
- [ ] <!-- unlimit-branch=renovate/click-8.x -->deps: update dependency click to v8.1.3
- [ ] <!-- unlimit-branch=renovate/gcp-releasetool-1.x -->deps: update dependency gcp-releasetool to v1.11.0
- [ ] <!-- unlimit-branch=renovate/google-api-core-2.x -->deps: update dependency google-api-core to v2.11.0
- [ ] <!-- unlimit-branch=renovate/google-auth-2.x -->deps: update dependency google-auth to v2.17.1
- [ ] <!-- unlimit-branch=renovate/google-cloud-storage-2.x -->deps: update dependency google-cloud-storage to v2.8.0
- [ ] <!-- unlimit-branch=renovate/google-crc32c-1.x -->deps: update dependency google-crc32c to v1.5.0
- [ ] <!-- unlimit-branch=renovate/google-resumable-media-2.x -->deps: update dependency google-resumable-media to v2.4.1
- [ ] <!-- unlimit-branch=renovate/googleapis-common-protos-1.x -->deps: update dependency googleapis-common-protos to v1.59.0
- [ ] <!-- unlimit-branch=renovate/importlib-metadata-4.x -->deps: update dependency importlib-metadata to v4.13.0
- [ ] <!-- unlimit-branch=renovate/jeepney-0.x -->deps: update dependency jeepney to v0.8.0
- [ ] <!-- unlimit-branch=renovate/jinja2-3.x -->deps: update dependency jinja2 to v3.1.2
- [ ] <!-- unlimit-branch=renovate/keyring-23.x -->deps: update dependency keyring to v23.13.1
- [ ] <!-- unlimit-branch=renovate/markupsafe-2.x -->deps: update dependency markupsafe to v2.1.2
- [ ] <!-- unlimit-branch=renovate/protobuf-3.x -->deps: update dependency protobuf to v3.20.3
- [ ] <!-- unlimit-branch=renovate/pyjwt-2.x -->deps: update dependency pyjwt to v2.6.0
- [ ] <!-- unlimit-branch=renovate/requests-2.x -->deps: update dependency requests to v2.28.2
- [ ] <!-- unlimit-branch=renovate/typing-extensions-4.x -->deps: update dependency typing-extensions to v4.5.0
- [ ] <!-- unlimit-branch=renovate/zipp-3.x -->deps: update dependency zipp to v3.15.0
- [ ] <!-- create-all-rate-limited-prs -->🔐 **Create all rate-limited PRs at once** 🔐
---
### ⚠ Dependency Lookup Warnings ⚠
- Renovate failed to look up the following dependencies: `Failed to look up maven package com.google.cloud:gapic-showcase`, `Failed to look up maven package com.google.cloud:gapic-showcase-parent`, `Failed to look up maven package com.google.api.grpc:proto-gapic-showcase-v1beta1`, `Failed to look up maven package com.google.api.grpc:grpc-gapic-showcase-v1beta1`.
Files affected: `coverage-report/pom.xml`, `showcase/gapic-showcase/pom.xml`, `showcase/grpc-gapic-showcase-v1beta1/pom.xml`, `showcase/pom.xml`, `showcase/proto-gapic-showcase-v1beta1/pom.xml`
---
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/cryptography-39.x -->[deps: update dependency cryptography to v39.0.2](../pull/1592)
- [ ] <!-- rebase-branch=renovate/gcp-docuploader-0.x -->[deps: update dependency gcp-docuploader to v0.6.5](../pull/1593)
- [ ] <!-- rebase-branch=renovate/google-cloud-core-2.x -->[deps: update dependency google-cloud-core to v2.3.2](../pull/1595)
- [ ] <!-- rebase-branch=renovate/net.bytebuddy-byte-buddy-1.x -->[deps: update dependency net.bytebuddy:byte-buddy to v1.14.3](../pull/1567)
- [ ] <!-- rebase-branch=renovate/org.threeten-threetenbp-1.x -->[deps: update dependency org.threeten:threetenbp to v1.6.8](../pull/1590)
- [ ] <!-- rebase-branch=renovate/threeten.version -->[deps: update dependency org.threeten:threetenbp to v1.6.8](../pull/1591)
- [ ] <!-- rebase-branch=renovate/urllib3-1.x -->[deps: update dependency urllib3 to v1.26.15](../pull/1596)
- [ ] <!-- rebase-branch=renovate/google-api-dependencies -->[deps: update google api dependencies](../pull/1578) (`com.google.api.grpc:grpc-google-common-protos`, `com.google.api.grpc:grpc-google-iam-v1`, `com.google.api.grpc:proto-google-common-protos`, `com.google.api.grpc:proto-google-iam-v1`, `com.google.api:api-common`)
- [ ] <!-- rebase-branch=renovate/protobuf-dependencies -->[deps: update protobuf dependencies to v3.22.2](../pull/1369) (`com.google.protobuf:protobuf-java`, `com.google.protobuf:protobuf-bom`)
- [x] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once**
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/org.apache.maven.plugins-maven-surefire-plugin-3.x -->[build(deps): update dependency org.apache.maven.plugins:maven-surefire-plugin to v3.0.0](../pull/1488)
- [ ] <!-- recreate-branch=renovate/com.coveo-fmt-maven-plugin-2.x -->[build(deps): update dependency com.coveo:fmt-maven-plugin to v2.13](../pull/1363)
- [ ] <!-- recreate-branch=renovate/com.google.googlejavaformat-google-java-format-1.x -->[deps: update dependency com.google.googlejavaformat:google-java-format to v1.16.0](../pull/1422)
## Detected dependencies
<details><summary>bazel</summary>
<blockquote>
<details><summary>WORKSPACE</summary>
</details>
</blockquote>
</details>
<details><summary>github-actions</summary>
<blockquote>
<details><summary>.github/workflows/ci.yaml</summary>
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/checkout v3`
- `actions/setup-java v3`
</details>
<details><summary>.github/workflows/downstream-native-image.yaml</summary>
- `actions/checkout v3`
- `stCarolas/setup-maven v4`
- `ayltai/setup-graalvm v1`
</details>
<details><summary>.github/workflows/nightly.yaml</summary>
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/setup-java v3`
</details>
<details><summary>.github/workflows/shared_dependencies.yaml</summary>
- `actions/checkout v2`
- `stCarolas/setup-maven v4`
- `actions/setup-java v1`
</details>
<details><summary>.github/workflows/sonar.yaml</summary>
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/cache v3`
- `actions/cache v3`
</details>
</blockquote>
</details>
<details><summary>gradle-wrapper</summary>
<blockquote>
<details><summary>rules_java_gapic/resources/gradle/gradle/wrapper/gradle-wrapper.properties</summary>
- `gradle 7.6.1`
</details>
</blockquote>
</details>
<details><summary>maven</summary>
<blockquote>
<details><summary>api-common-java/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.auto.value:auto-value-annotations 1.10.1`
- `com.google.auto.value:auto-value 1.10.1`
- `com.google.code.findbugs:jsr305 3.0.2`
- `com.google.errorprone:error_prone_annotations 2.18.0`
- `junit:junit 4.13.2`
- `com.google.truth:truth 1.1.3`
- `com.google.auto.value:auto-value 1.10.1`
</details>
<details><summary>coverage-report/pom.xml</summary>
- `com.google.cloud:gapic-showcase 0.0.1-SNAPSHOT`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax-grpc 2.24.1-SNAPSHOT`
- `com.google.api:gax-httpjson 0.109.1-SNAPSHOT`
- `com.google.api:api-common 2.7.1-SNAPSHOT`
- `org.jacoco:jacoco-maven-plugin 0.8.8`
- `com.coveo:fmt-maven-plugin 2.9`
</details>
<details><summary>gapic-generator-java-bom/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api:api-common 2.7.1-SNAPSHOT`
- `com.google.api:gax-bom 2.24.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v1 1.10.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v2 1.10.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v2beta 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v1 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v2 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v2beta 1.10.1-SNAPSHOT`
</details>
<details><summary>gapic-generator-java-pom-parent/pom.xml</summary>
- `com.google.guava:guava 31.1-jre`
- `javax.annotation:javax.annotation-api 1.3.2`
- `com.google.auth:google-auth-library-bom 1.16.0`
- `com.google.code.gson:gson 2.10.1`
- `com.google.guava:guava-bom 31.1-jre`
- `com.google.protobuf:protobuf-bom 3.21.12`
- `io.grpc:grpc-bom 1.54.0`
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `org.apache.maven.plugins:maven-surefire-plugin 3.0.0-M8`
- `org.apache.maven.plugins:maven-failsafe-plugin 3.0.0`
- `org.jacoco:jacoco-maven-plugin 0.8.8`
- `javax.annotation:javax.annotation-api 1.3.2`
- `com.google.auth:google-auth-library-bom 1.16.0`
- `com.google.guava:guava 31.1-jre`
- `io.grpc:grpc-bom 1.54.0`
- `com.google.code.gson:gson 2.10.1`
- `com.google.protobuf:protobuf-bom 3.21.12`
- `javax.annotation:javax.annotation-api 1.3.2`
- `com.google.protobuf:protobuf-bom 3.21.12`
- `io.grpc:grpc-bom 1.54.0`
- `javax.annotation:javax.annotation-api 1.3.2`
- `com.google.guava:guava-bom 31.1-jre`
- `com.google.protobuf:protobuf-bom 3.21.12`
- `io.grpc:grpc-bom 1.54.0`
- `javax.annotation:javax.annotation-api 1.3.2`
</details>
<details><summary>gapic-generator-java/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT`
- `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13`
- `kr.motd.maven:os-maven-plugin 1.7.1`
- `org.apache.maven.plugins:maven-compiler-plugin 3.11.0`
- `com.google.auto.value:auto-value 1.10.1`
- `com.coveo:fmt-maven-plugin 2.9.1`
- `com.googlecode.maven-download-plugin:download-maven-plugin 1.6.8`
- `org.xolstice.maven.plugins:protobuf-maven-plugin 0.6.1`
- `org.apache.maven.plugins:maven-surefire-plugin 3.0.0`
- `org.jacoco:jacoco-maven-plugin 0.8.8`
- `org.apache.maven.plugins:maven-shade-plugin 3.4.1`
- `com.google.guava:guava 31.1-jre`
- `com.google.auto.value:auto-value-annotations 1.10.1`
- `org.yaml:snakeyaml 2.0`
- `com.google.googlejavaformat:google-java-format 1.7`
- `junit:junit 4.13.2`
- `com.google.errorprone:error_prone_annotations 2.18.0`
- `com.google.truth:truth 1.1.3`
- `io.github.java-diff-utils:java-diff-utils 4.12`
</details>
<details><summary>gax-java/gax-bom/pom.xml</summary>
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax-grpc 2.24.1-SNAPSHOT`
- `com.google.api:gax-grpc 2.24.1-SNAPSHOT`
- `com.google.api:gax-grpc 2.24.1-SNAPSHOT`
- `com.google.api:gax-httpjson 0.109.1-SNAPSHOT`
- `com.google.api:gax-httpjson 0.109.1-SNAPSHOT`
- `com.google.api:gax-httpjson 0.109.1-SNAPSHOT`
</details>
<details><summary>gax-java/gax-grpc/pom.xml</summary>
- `com.google.api:gax-parent 2.24.1-SNAPSHOT`
</details>
<details><summary>gax-java/gax-httpjson/pom.xml</summary>
- `com.google.api:gax-parent 2.24.1-SNAPSHOT`
</details>
<details><summary>gax-java/gax/pom.xml</summary>
- `com.google.api:gax-parent 2.24.1-SNAPSHOT`
</details>
<details><summary>gax-java/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api:api-common 2.7.1-SNAPSHOT`
- `org.threeten:threetenbp 1.6.7`
- `com.google.code.findbugs:jsr305 3.0.2`
- `io.opencensus:opencensus-api 0.31.1`
- `com.google.auto.value:auto-value-annotations 1.10.1`
- `com.google.auto.value:auto-value 1.10.1`
- `junit:junit 4.13.2`
- `org.mockito:mockito-core 4.11.0`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT`
- `org.graalvm.sdk:graal-sdk 22.3.1`
- `com.google.http-client:google-http-client-bom 1.43.1`
- `com.google.truth:truth 1.1.3`
- `com.google.auto.value:auto-value 1.10.1`
- `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13`
</details>
<details><summary>java-common-protos/grpc-google-common-protos/pom.xml</summary>
- `com.google.api.grpc:google-common-protos-parent 2.15.1-SNAPSHOT`
</details>
<details><summary>java-common-protos/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.guava:guava 31.1-jre`
- `junit:junit 4.13.2`
- `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.2`
- `org.apache.maven.plugins:maven-javadoc-plugin 3.5.0`
</details>
<details><summary>java-common-protos/proto-google-common-protos/pom.xml</summary>
- `com.google.api.grpc:google-common-protos-parent 2.15.1-SNAPSHOT`
</details>
<details><summary>java-core/google-cloud-core-bom/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT`
- `com.google.cloud:google-cloud-core-grpc 2.14.1-SNAPSHOT`
- `com.google.cloud:google-cloud-core-http 2.14.1-SNAPSHOT`
</details>
<details><summary>java-core/google-cloud-core-grpc/pom.xml</summary>
- `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT`
</details>
<details><summary>java-core/google-cloud-core-http/pom.xml</summary>
- `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT`
</details>
<details><summary>java-core/google-cloud-core/pom.xml</summary>
- `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT`
</details>
<details><summary>java-core/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.cloud:google-cloud-shared-dependencies 3.6.1-SNAPSHOT`
- `junit:junit 4.13.2`
- `com.google.truth:truth 1.1.3`
- `org.easymock:easymock 5.1.0`
- `org.mockito:mockito-all 1.10.19`
</details>
<details><summary>java-iam/grpc-google-iam-v1/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/grpc-google-iam-v2/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/grpc-google-iam-v2beta/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api:gax-bom 2.24.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v2 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v2 1.10.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v2beta 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v1 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v2beta 1.10.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v1 1.10.1-SNAPSHOT`
- `junit:junit 4.13.2`
- `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.2`
- `org.apache.maven.plugins:maven-javadoc-plugin 3.5.0`
</details>
<details><summary>java-iam/proto-google-iam-v1/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/proto-google-iam-v2/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/proto-google-iam-v2beta/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-shared-dependencies/dependency-convergence-check/pom.xml</summary>
- `com.google.guava:guava 31.0.1-jre`
- `com.google.cloud.tools:dependencies 1.5.13`
</details>
<details><summary>java-shared-dependencies/first-party-dependencies/pom.xml</summary>
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT`
- `com.google.cloud:grpc-gcp 1.4.1`
- `com.google.code.gson:gson 2.10.1`
- `com.google.cloud:google-cloud-core-bom 2.14.1-SNAPSHOT`
- `com.google.http-client:google-http-client-bom 1.43.1`
- `com.google.oauth-client:google-oauth-client-bom 1.34.1`
- `com.google.api-client:google-api-client-bom 2.2.0`
- `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT`
- `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT`
</details>
<details><summary>java-shared-dependencies/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.cloud:first-party-dependencies 3.6.1-SNAPSHOT`
- `com.google.cloud:third-party-dependencies 3.6.1-SNAPSHOT`
</details>
<details><summary>java-shared-dependencies/third-party-dependencies/pom.xml</summary>
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `org.apache.httpcomponents:httpcore 4.4.16`
- `org.apache.httpcomponents:httpclient 4.5.14`
- `org.threeten:threetenbp 1.6.7`
- `javax.annotation:javax.annotation-api 1.3.2`
- `org.codehaus.mojo:animal-sniffer-annotations 1.23`
- `com.google.code.findbugs:jsr305 3.0.2`
- `com.google.errorprone:error_prone_annotations 2.18.0`
- `com.fasterxml.jackson:jackson-bom 2.14.2`
- `commons-codec:commons-codec 1.15`
- `io.opencensus:opencensus-api 0.31.1`
- `io.opencensus:opencensus-contrib-grpc-util 0.31.1`
- `io.opencensus:opencensus-contrib-http-util 0.31.1`
- `io.opencensus:opencensus-contrib-zpages 0.31.1`
- `io.opencensus:opencensus-exporter-stats-stackdriver 0.31.1`
- `io.opencensus:opencensus-exporter-trace-stackdriver 0.31.1`
- `io.opencensus:opencensus-impl 0.31.1`
- `io.opencensus:opencensus-impl-core 0.31.1`
- `org.checkerframework:checker-qual 3.32.0`
- `io.perfmark:perfmark-api 0.26.0`
</details>
<details><summary>java-shared-dependencies/upper-bound-check/pom.xml</summary>
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `com.google.cloud:google-cloud-shared-dependencies 3.6.1-SNAPSHOT`
</details>
<details><summary>pom.xml</summary>
- `com.coveo:fmt-maven-plugin 2.9`
- `org.apache.maven.plugins:maven-deploy-plugin 3.1.1`
- `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13`
</details>
<details><summary>showcase/gapic-showcase/pom.xml</summary>
- `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
- `org.jacoco:jacoco-maven-plugin 0.8.8`
- `com.coveo:fmt-maven-plugin 2.9`
- `com.googlecode.maven-download-plugin:download-maven-plugin 1.6.8`
- `com.google.truth:truth 1.1.3`
</details>
<details><summary>showcase/grpc-gapic-showcase-v1beta1/pom.xml</summary>
- `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
</details>
<details><summary>showcase/pom.xml</summary>
- `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT`
- `com.google.api.grpc:proto-gapic-showcase-v1beta1 0.0.1-SNAPSHOT`
- `com.google.api.grpc:grpc-gapic-showcase-v1beta1 0.0.1-SNAPSHOT`
- `com.google.cloud:gapic-showcase 0.0.1-SNAPSHOT`
- `junit:junit 4.13.2`
- `org.apache.maven.plugins:maven-surefire-plugin 3.0.0-M8`
</details>
<details><summary>showcase/proto-gapic-showcase-v1beta1/pom.xml</summary>
- `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
</details>
</blockquote>
</details>
<details><summary>pip_requirements</summary>
<blockquote>
<details><summary>.kokoro/requirements.txt</summary>
- `attrs ==22.1.0`
- `cachetools ==4.2.4`
- `certifi ==2022.12.7`
- `cffi ==1.15.1`
- `charset-normalizer ==2.0.12`
- `click ==8.0.4`
- `colorlog ==6.7.0`
- `cryptography ==39.0.1`
- `gcp-docuploader ==0.6.3`
- `gcp-releasetool ==1.10.5`
- `google-api-core ==2.8.2`
- `google-auth ==2.14.1`
- `google-cloud-core ==2.3.1`
- `google-cloud-storage ==2.0.0`
- `google-crc32c ==1.3.0`
- `google-resumable-media ==2.3.3`
- `googleapis-common-protos ==1.56.3`
- `idna ==3.4`
- `importlib-metadata ==4.8.3`
- `jeepney ==0.7.1`
- `jinja2 ==3.0.3`
- `keyring ==23.4.1`
- `markupsafe ==2.0.1`
- `packaging ==21.3`
- `protobuf ==3.19.5`
- `pyasn1 ==0.4.8`
- `pyasn1-modules ==0.2.8`
- `pycparser ==2.21`
- `pyjwt ==2.4.0`
- `pyparsing ==3.0.9`
- `pyperclip ==1.8.2`
- `python-dateutil ==2.8.2`
- `requests ==2.27.1`
- `rsa ==4.9`
- `secretstorage ==3.3.3`
- `six ==1.16.0`
- `typing-extensions ==4.1.1`
- `urllib3 ==1.26.12`
- `zipp ==3.6.0`
</details>
</blockquote>
</details>
<details><summary>regex</summary>
<blockquote>
<details><summary>gax-java/dependencies.properties</summary>
- `com.google.api.grpc:proto-google-common-protos 2.14.3`
- `com.google.api.grpc:grpc-google-common-protos 2.14.3`
- `com.google.auth:google-auth-library-oauth2-http 1.16.0`
- `com.google.auth:google-auth-library-credentials 1.16.0`
- `io.opencensus:opencensus-api 0.31.1`
- `io.opencensus:opencensus-contrib-grpc-metrics 0.31.1`
- `io.opencensus:opencensus-contrib-http-util 0.31.1`
- `io.netty:netty-tcnative-boringssl-static 2.0.59.Final`
- `io.netty:netty-handler 4.1.90.Final`
- `io.netty:netty-common 4.1.90.Final`
- `io.netty:netty-codec-socks 4.1.90.Final`
- `io.netty:netty-codec-http2 4.1.90.Final`
- `io.netty:netty-codec-http 4.1.90.Final`
- `io.netty:netty-codec 4.1.90.Final`
- `io.netty:netty-buffer 4.1.90.Final`
- `io.netty:netty-resolver 4.1.90.Final`
- `io.netty:netty-transport 4.1.90.Final`
- `io.netty:netty-handler-proxy 4.1.90.Final`
- `io.netty:netty-transport-native-unix-common 4.1.90.Final`
- `io.perfmark:perfmark-api 0.26.0`
- `org.apache.tomcat:annotations-api 6.0.53`
- `com.google.code.gson:gson 2.10.1`
- `com.google.guava:guava 31.1-jre`
- `com.google.guava:failureaccess 1.0.1`
- `org.apache.commons:commons-lang3 3.12.0`
- `com.google.android:annotations 4.1.1.4`
- `com.google.code.findbugs:jsr305 3.0.2`
- `com.google.errorprone:error_prone_annotations 2.18.0`
- `com.google.j2objc:j2objc-annotations 1.3`
- `com.google.auto.value:auto-value 1.10.1`
- `com.google.auto.value:auto-value-annotations 1.10.1`
- `com.google.api:api-common 2.6.3`
- `org.threeten:threetenbp 1.6.7`
- `com.google.api.grpc:grpc-google-iam-v1 1.9.3`
- `com.google.api.grpc:proto-google-iam-v1 1.9.3`
- `com.google.http-client:google-http-client 1.43.1`
- `com.google.http-client:google-http-client-gson 1.43.1`
- `org.codehaus.mojo:animal-sniffer-annotations 1.23`
- `javax.annotation:javax.annotation-api 1.3.2`
- `org.graalvm.sdk:graal-sdk 22.3.1`
- `junit:junit 4.13.2`
- `org.mockito:mockito-core 2.28.2`
- `org.hamcrest:hamcrest-core 1.3`
- `com.google.truth:truth 1.1.3`
- `com.googlecode.java-diff-utils:diffutils 1.3.0`
- `net.bytebuddy:byte-buddy 1.14.2`
- `org.objenesis:objenesis 2.6`
</details>
<details><summary>gax-java/dependencies.properties</summary>
- `com.google.protobuf:protobuf-java 3.21.10`
</details>
<details><summary>gax-java/dependencies.properties</summary>
- `com.google.googlejavaformat:google-java-format 1.15.0`
</details>
<details><summary>gax-java/dependencies.properties</summary>
- `io.grpc:grpc-core 1.54.0`
</details>
</blockquote>
</details>
---
- [x] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more.
## Rate-Limited
These updates are currently rate-limited. Click on a checkbox below to force their creation now.
- [x] <!-- unlimit-branch=renovate/org.jacoco-jacoco-maven-plugin-0.x -->build(deps): update dependency org.jacoco:jacoco-maven-plugin to v0.8.9
- [x] <!-- unlimit-branch=renovate/netty-dependencies -->deps: update netty dependencies to v4.1.91.final (`io.netty:netty-buffer`, `io.netty:netty-codec`, `io.netty:netty-codec-http`, `io.netty:netty-codec-http2`, `io.netty:netty-codec-socks`, `io.netty:netty-common`, `io.netty:netty-handler`, `io.netty:netty-handler-proxy`, `io.netty:netty-resolver`, `io.netty:netty-transport`, `io.netty:netty-transport-native-unix-common`)
- [ ] <!-- unlimit-branch=renovate/attrs-22.x -->deps: update dependency attrs to v22.2.0
- [ ] <!-- unlimit-branch=renovate/charset-normalizer-2.x -->deps: update dependency charset-normalizer to v2.1.1
- [ ] <!-- unlimit-branch=renovate/click-8.x -->deps: update dependency click to v8.1.3
- [ ] <!-- unlimit-branch=renovate/gcp-releasetool-1.x -->deps: update dependency gcp-releasetool to v1.11.0
- [ ] <!-- unlimit-branch=renovate/google-api-core-2.x -->deps: update dependency google-api-core to v2.11.0
- [ ] <!-- unlimit-branch=renovate/google-auth-2.x -->deps: update dependency google-auth to v2.17.1
- [ ] <!-- unlimit-branch=renovate/google-cloud-storage-2.x -->deps: update dependency google-cloud-storage to v2.8.0
- [ ] <!-- unlimit-branch=renovate/google-crc32c-1.x -->deps: update dependency google-crc32c to v1.5.0
- [ ] <!-- unlimit-branch=renovate/google-resumable-media-2.x -->deps: update dependency google-resumable-media to v2.4.1
- [ ] <!-- unlimit-branch=renovate/googleapis-common-protos-1.x -->deps: update dependency googleapis-common-protos to v1.59.0
- [ ] <!-- unlimit-branch=renovate/importlib-metadata-4.x -->deps: update dependency importlib-metadata to v4.13.0
- [ ] <!-- unlimit-branch=renovate/jeepney-0.x -->deps: update dependency jeepney to v0.8.0
- [ ] <!-- unlimit-branch=renovate/jinja2-3.x -->deps: update dependency jinja2 to v3.1.2
- [ ] <!-- unlimit-branch=renovate/keyring-23.x -->deps: update dependency keyring to v23.13.1
- [ ] <!-- unlimit-branch=renovate/markupsafe-2.x -->deps: update dependency markupsafe to v2.1.2
- [ ] <!-- unlimit-branch=renovate/protobuf-3.x -->deps: update dependency protobuf to v3.20.3
- [ ] <!-- unlimit-branch=renovate/pyjwt-2.x -->deps: update dependency pyjwt to v2.6.0
- [ ] <!-- unlimit-branch=renovate/requests-2.x -->deps: update dependency requests to v2.28.2
- [ ] <!-- unlimit-branch=renovate/typing-extensions-4.x -->deps: update dependency typing-extensions to v4.5.0
- [ ] <!-- unlimit-branch=renovate/zipp-3.x -->deps: update dependency zipp to v3.15.0
- [ ] <!-- create-all-rate-limited-prs -->🔐 **Create all rate-limited PRs at once** 🔐
---
### ⚠ Dependency Lookup Warnings ⚠
- Renovate failed to look up the following dependencies: `Failed to look up maven package com.google.cloud:gapic-showcase`, `Failed to look up maven package com.google.cloud:gapic-showcase-parent`, `Failed to look up maven package com.google.api.grpc:proto-gapic-showcase-v1beta1`, `Failed to look up maven package com.google.api.grpc:grpc-gapic-showcase-v1beta1`.
Files affected: `coverage-report/pom.xml`, `showcase/gapic-showcase/pom.xml`, `showcase/grpc-gapic-showcase-v1beta1/pom.xml`, `showcase/pom.xml`, `showcase/proto-gapic-showcase-v1beta1/pom.xml`
---
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/cryptography-39.x -->[deps: update dependency cryptography to v39.0.2](../pull/1592)
- [ ] <!-- rebase-branch=renovate/gcp-docuploader-0.x -->[deps: update dependency gcp-docuploader to v0.6.5](../pull/1593)
- [ ] <!-- rebase-branch=renovate/google-cloud-core-2.x -->[deps: update dependency google-cloud-core to v2.3.2](../pull/1595)
- [ ] <!-- rebase-branch=renovate/net.bytebuddy-byte-buddy-1.x -->[deps: update dependency net.bytebuddy:byte-buddy to v1.14.3](../pull/1567)
- [ ] <!-- rebase-branch=renovate/org.threeten-threetenbp-1.x -->[deps: update dependency org.threeten:threetenbp to v1.6.8](../pull/1590)
- [ ] <!-- rebase-branch=renovate/threeten.version -->[deps: update dependency org.threeten:threetenbp to v1.6.8](../pull/1591)
- [ ] <!-- rebase-branch=renovate/urllib3-1.x -->[deps: update dependency urllib3 to v1.26.15](../pull/1596)
- [ ] <!-- rebase-branch=renovate/google-api-dependencies -->[deps: update google api dependencies](../pull/1578) (`com.google.api.grpc:grpc-google-common-protos`, `com.google.api.grpc:grpc-google-iam-v1`, `com.google.api.grpc:proto-google-common-protos`, `com.google.api.grpc:proto-google-iam-v1`, `com.google.api:api-common`)
- [ ] <!-- rebase-branch=renovate/protobuf-dependencies -->[deps: update protobuf dependencies to v3.22.2](../pull/1369) (`com.google.protobuf:protobuf-java`, `com.google.protobuf:protobuf-bom`)
- [x] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once**
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/org.apache.maven.plugins-maven-surefire-plugin-3.x -->[build(deps): update dependency org.apache.maven.plugins:maven-surefire-plugin to v3.0.0](../pull/1488)
- [ ] <!-- recreate-branch=renovate/com.coveo-fmt-maven-plugin-2.x -->[build(deps): update dependency com.coveo:fmt-maven-plugin to v2.13](../pull/1363)
- [ ] <!-- recreate-branch=renovate/com.google.googlejavaformat-google-java-format-1.x -->[deps: update dependency com.google.googlejavaformat:google-java-format to v1.16.0](../pull/1422)
## Detected dependencies
<details><summary>bazel</summary>
<blockquote>
<details><summary>WORKSPACE</summary>
</details>
</blockquote>
</details>
<details><summary>github-actions</summary>
<blockquote>
<details><summary>.github/workflows/ci.yaml</summary>
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/checkout v3`
- `actions/setup-java v3`
</details>
<details><summary>.github/workflows/downstream-native-image.yaml</summary>
- `actions/checkout v3`
- `stCarolas/setup-maven v4`
- `ayltai/setup-graalvm v1`
</details>
<details><summary>.github/workflows/nightly.yaml</summary>
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/setup-java v3`
</details>
<details><summary>.github/workflows/shared_dependencies.yaml</summary>
- `actions/checkout v2`
- `stCarolas/setup-maven v4`
- `actions/setup-java v1`
</details>
<details><summary>.github/workflows/sonar.yaml</summary>
- `actions/checkout v3`
- `actions/setup-java v3`
- `actions/cache v3`
- `actions/cache v3`
</details>
</blockquote>
</details>
<details><summary>gradle-wrapper</summary>
<blockquote>
<details><summary>rules_java_gapic/resources/gradle/gradle/wrapper/gradle-wrapper.properties</summary>
- `gradle 7.6.1`
</details>
</blockquote>
</details>
<details><summary>maven</summary>
<blockquote>
<details><summary>api-common-java/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.auto.value:auto-value-annotations 1.10.1`
- `com.google.auto.value:auto-value 1.10.1`
- `com.google.code.findbugs:jsr305 3.0.2`
- `com.google.errorprone:error_prone_annotations 2.18.0`
- `junit:junit 4.13.2`
- `com.google.truth:truth 1.1.3`
- `com.google.auto.value:auto-value 1.10.1`
</details>
<details><summary>coverage-report/pom.xml</summary>
- `com.google.cloud:gapic-showcase 0.0.1-SNAPSHOT`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax-grpc 2.24.1-SNAPSHOT`
- `com.google.api:gax-httpjson 0.109.1-SNAPSHOT`
- `com.google.api:api-common 2.7.1-SNAPSHOT`
- `org.jacoco:jacoco-maven-plugin 0.8.8`
- `com.coveo:fmt-maven-plugin 2.9`
</details>
<details><summary>gapic-generator-java-bom/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api:api-common 2.7.1-SNAPSHOT`
- `com.google.api:gax-bom 2.24.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v1 1.10.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v2 1.10.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v2beta 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v1 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v2 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v2beta 1.10.1-SNAPSHOT`
</details>
<details><summary>gapic-generator-java-pom-parent/pom.xml</summary>
- `com.google.guava:guava 31.1-jre`
- `javax.annotation:javax.annotation-api 1.3.2`
- `com.google.auth:google-auth-library-bom 1.16.0`
- `com.google.code.gson:gson 2.10.1`
- `com.google.guava:guava-bom 31.1-jre`
- `com.google.protobuf:protobuf-bom 3.21.12`
- `io.grpc:grpc-bom 1.54.0`
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `org.apache.maven.plugins:maven-surefire-plugin 3.0.0-M8`
- `org.apache.maven.plugins:maven-failsafe-plugin 3.0.0`
- `org.jacoco:jacoco-maven-plugin 0.8.8`
- `javax.annotation:javax.annotation-api 1.3.2`
- `com.google.auth:google-auth-library-bom 1.16.0`
- `com.google.guava:guava 31.1-jre`
- `io.grpc:grpc-bom 1.54.0`
- `com.google.code.gson:gson 2.10.1`
- `com.google.protobuf:protobuf-bom 3.21.12`
- `javax.annotation:javax.annotation-api 1.3.2`
- `com.google.protobuf:protobuf-bom 3.21.12`
- `io.grpc:grpc-bom 1.54.0`
- `javax.annotation:javax.annotation-api 1.3.2`
- `com.google.guava:guava-bom 31.1-jre`
- `com.google.protobuf:protobuf-bom 3.21.12`
- `io.grpc:grpc-bom 1.54.0`
- `javax.annotation:javax.annotation-api 1.3.2`
</details>
<details><summary>gapic-generator-java/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT`
- `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13`
- `kr.motd.maven:os-maven-plugin 1.7.1`
- `org.apache.maven.plugins:maven-compiler-plugin 3.11.0`
- `com.google.auto.value:auto-value 1.10.1`
- `com.coveo:fmt-maven-plugin 2.9.1`
- `com.googlecode.maven-download-plugin:download-maven-plugin 1.6.8`
- `org.xolstice.maven.plugins:protobuf-maven-plugin 0.6.1`
- `org.apache.maven.plugins:maven-surefire-plugin 3.0.0`
- `org.jacoco:jacoco-maven-plugin 0.8.8`
- `org.apache.maven.plugins:maven-shade-plugin 3.4.1`
- `com.google.guava:guava 31.1-jre`
- `com.google.auto.value:auto-value-annotations 1.10.1`
- `org.yaml:snakeyaml 2.0`
- `com.google.googlejavaformat:google-java-format 1.7`
- `junit:junit 4.13.2`
- `com.google.errorprone:error_prone_annotations 2.18.0`
- `com.google.truth:truth 1.1.3`
- `io.github.java-diff-utils:java-diff-utils 4.12`
</details>
<details><summary>gax-java/gax-bom/pom.xml</summary>
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax-grpc 2.24.1-SNAPSHOT`
- `com.google.api:gax-grpc 2.24.1-SNAPSHOT`
- `com.google.api:gax-grpc 2.24.1-SNAPSHOT`
- `com.google.api:gax-httpjson 0.109.1-SNAPSHOT`
- `com.google.api:gax-httpjson 0.109.1-SNAPSHOT`
- `com.google.api:gax-httpjson 0.109.1-SNAPSHOT`
</details>
<details><summary>gax-java/gax-grpc/pom.xml</summary>
- `com.google.api:gax-parent 2.24.1-SNAPSHOT`
</details>
<details><summary>gax-java/gax-httpjson/pom.xml</summary>
- `com.google.api:gax-parent 2.24.1-SNAPSHOT`
</details>
<details><summary>gax-java/gax/pom.xml</summary>
- `com.google.api:gax-parent 2.24.1-SNAPSHOT`
</details>
<details><summary>gax-java/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api:api-common 2.7.1-SNAPSHOT`
- `org.threeten:threetenbp 1.6.7`
- `com.google.code.findbugs:jsr305 3.0.2`
- `io.opencensus:opencensus-api 0.31.1`
- `com.google.auto.value:auto-value-annotations 1.10.1`
- `com.google.auto.value:auto-value 1.10.1`
- `junit:junit 4.13.2`
- `org.mockito:mockito-core 4.11.0`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api:gax 2.24.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT`
- `org.graalvm.sdk:graal-sdk 22.3.1`
- `com.google.http-client:google-http-client-bom 1.43.1`
- `com.google.truth:truth 1.1.3`
- `com.google.auto.value:auto-value 1.10.1`
- `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13`
</details>
<details><summary>java-common-protos/grpc-google-common-protos/pom.xml</summary>
- `com.google.api.grpc:google-common-protos-parent 2.15.1-SNAPSHOT`
</details>
<details><summary>java-common-protos/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.guava:guava 31.1-jre`
- `junit:junit 4.13.2`
- `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.2`
- `org.apache.maven.plugins:maven-javadoc-plugin 3.5.0`
</details>
<details><summary>java-common-protos/proto-google-common-protos/pom.xml</summary>
- `com.google.api.grpc:google-common-protos-parent 2.15.1-SNAPSHOT`
</details>
<details><summary>java-core/google-cloud-core-bom/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT`
- `com.google.cloud:google-cloud-core-grpc 2.14.1-SNAPSHOT`
- `com.google.cloud:google-cloud-core-http 2.14.1-SNAPSHOT`
</details>
<details><summary>java-core/google-cloud-core-grpc/pom.xml</summary>
- `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT`
</details>
<details><summary>java-core/google-cloud-core-http/pom.xml</summary>
- `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT`
</details>
<details><summary>java-core/google-cloud-core/pom.xml</summary>
- `com.google.cloud:google-cloud-core-parent 2.14.1-SNAPSHOT`
</details>
<details><summary>java-core/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.cloud:google-cloud-shared-dependencies 3.6.1-SNAPSHOT`
- `junit:junit 4.13.2`
- `com.google.truth:truth 1.1.3`
- `org.easymock:easymock 5.1.0`
- `org.mockito:mockito-all 1.10.19`
</details>
<details><summary>java-iam/grpc-google-iam-v1/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/grpc-google-iam-v2/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/grpc-google-iam-v2beta/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.api:gax-bom 2.24.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v2 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v2 1.10.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-common-protos 2.15.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v2beta 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v1 1.10.1-SNAPSHOT`
- `com.google.api.grpc:grpc-google-iam-v2beta 1.10.1-SNAPSHOT`
- `com.google.api.grpc:proto-google-iam-v1 1.10.1-SNAPSHOT`
- `junit:junit 4.13.2`
- `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.2`
- `org.apache.maven.plugins:maven-javadoc-plugin 3.5.0`
</details>
<details><summary>java-iam/proto-google-iam-v1/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/proto-google-iam-v2/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-iam/proto-google-iam-v2beta/pom.xml</summary>
- `com.google.cloud:google-iam-parent 1.10.1-SNAPSHOT`
</details>
<details><summary>java-shared-dependencies/dependency-convergence-check/pom.xml</summary>
- `com.google.guava:guava 31.0.1-jre`
- `com.google.cloud.tools:dependencies 1.5.13`
</details>
<details><summary>java-shared-dependencies/first-party-dependencies/pom.xml</summary>
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT`
- `com.google.cloud:grpc-gcp 1.4.1`
- `com.google.code.gson:gson 2.10.1`
- `com.google.cloud:google-cloud-core-bom 2.14.1-SNAPSHOT`
- `com.google.http-client:google-http-client-bom 1.43.1`
- `com.google.oauth-client:google-oauth-client-bom 1.34.1`
- `com.google.api-client:google-api-client-bom 2.2.0`
- `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT`
- `com.google.cloud:google-cloud-core 2.14.1-SNAPSHOT`
</details>
<details><summary>java-shared-dependencies/pom.xml</summary>
- `com.google.api:gapic-generator-java-pom-parent 2.16.1-SNAPSHOT`
- `com.google.cloud:first-party-dependencies 3.6.1-SNAPSHOT`
- `com.google.cloud:third-party-dependencies 3.6.1-SNAPSHOT`
</details>
<details><summary>java-shared-dependencies/third-party-dependencies/pom.xml</summary>
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `org.apache.httpcomponents:httpcore 4.4.16`
- `org.apache.httpcomponents:httpclient 4.5.14`
- `org.threeten:threetenbp 1.6.7`
- `javax.annotation:javax.annotation-api 1.3.2`
- `org.codehaus.mojo:animal-sniffer-annotations 1.23`
- `com.google.code.findbugs:jsr305 3.0.2`
- `com.google.errorprone:error_prone_annotations 2.18.0`
- `com.fasterxml.jackson:jackson-bom 2.14.2`
- `commons-codec:commons-codec 1.15`
- `io.opencensus:opencensus-api 0.31.1`
- `io.opencensus:opencensus-contrib-grpc-util 0.31.1`
- `io.opencensus:opencensus-contrib-http-util 0.31.1`
- `io.opencensus:opencensus-contrib-zpages 0.31.1`
- `io.opencensus:opencensus-exporter-stats-stackdriver 0.31.1`
- `io.opencensus:opencensus-exporter-trace-stackdriver 0.31.1`
- `io.opencensus:opencensus-impl 0.31.1`
- `io.opencensus:opencensus-impl-core 0.31.1`
- `org.checkerframework:checker-qual 3.32.0`
- `io.perfmark:perfmark-api 0.26.0`
</details>
<details><summary>java-shared-dependencies/upper-bound-check/pom.xml</summary>
- `com.google.cloud:google-cloud-shared-config 1.5.5`
- `com.google.cloud:google-cloud-shared-dependencies 3.6.1-SNAPSHOT`
</details>
<details><summary>pom.xml</summary>
- `com.coveo:fmt-maven-plugin 2.9`
- `org.apache.maven.plugins:maven-deploy-plugin 3.1.1`
- `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13`
</details>
<details><summary>showcase/gapic-showcase/pom.xml</summary>
- `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
- `org.jacoco:jacoco-maven-plugin 0.8.8`
- `com.coveo:fmt-maven-plugin 2.9`
- `com.googlecode.maven-download-plugin:download-maven-plugin 1.6.8`
- `com.google.truth:truth 1.1.3`
</details>
<details><summary>showcase/grpc-gapic-showcase-v1beta1/pom.xml</summary>
- `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
</details>
<details><summary>showcase/pom.xml</summary>
- `com.google.api:gapic-generator-java-bom 2.16.1-SNAPSHOT`
- `com.google.api.grpc:proto-gapic-showcase-v1beta1 0.0.1-SNAPSHOT`
- `com.google.api.grpc:grpc-gapic-showcase-v1beta1 0.0.1-SNAPSHOT`
- `com.google.cloud:gapic-showcase 0.0.1-SNAPSHOT`
- `junit:junit 4.13.2`
- `org.apache.maven.plugins:maven-surefire-plugin 3.0.0-M8`
</details>
<details><summary>showcase/proto-gapic-showcase-v1beta1/pom.xml</summary>
- `com.google.cloud:gapic-showcase-parent 0.0.1-SNAPSHOT`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
- `org.codehaus.mojo:exec-maven-plugin 3.1.0`
</details>
</blockquote>
</details>
<details><summary>pip_requirements</summary>
<blockquote>
<details><summary>.kokoro/requirements.txt</summary>
- `attrs ==22.1.0`
- `cachetools ==4.2.4`
- `certifi ==2022.12.7`
- `cffi ==1.15.1`
- `charset-normalizer ==2.0.12`
- `click ==8.0.4`
- `colorlog ==6.7.0`
- `cryptography ==39.0.1`
- `gcp-docuploader ==0.6.3`
- `gcp-releasetool ==1.10.5`
- `google-api-core ==2.8.2`
- `google-auth ==2.14.1`
- `google-cloud-core ==2.3.1`
- `google-cloud-storage ==2.0.0`
- `google-crc32c ==1.3.0`
- `google-resumable-media ==2.3.3`
- `googleapis-common-protos ==1.56.3`
- `idna ==3.4`
- `importlib-metadata ==4.8.3`
- `jeepney ==0.7.1`
- `jinja2 ==3.0.3`
- `keyring ==23.4.1`
- `markupsafe ==2.0.1`
- `packaging ==21.3`
- `protobuf ==3.19.5`
- `pyasn1 ==0.4.8`
- `pyasn1-modules ==0.2.8`
- `pycparser ==2.21`
- `pyjwt ==2.4.0`
- `pyparsing ==3.0.9`
- `pyperclip ==1.8.2`
- `python-dateutil ==2.8.2`
- `requests ==2.27.1`
- `rsa ==4.9`
- `secretstorage ==3.3.3`
- `six ==1.16.0`
- `typing-extensions ==4.1.1`
- `urllib3 ==1.26.12`
- `zipp ==3.6.0`
</details>
</blockquote>
</details>
<details><summary>regex</summary>
<blockquote>
<details><summary>gax-java/dependencies.properties</summary>
- `com.google.api.grpc:proto-google-common-protos 2.14.3`
- `com.google.api.grpc:grpc-google-common-protos 2.14.3`
- `com.google.auth:google-auth-library-oauth2-http 1.16.0`
- `com.google.auth:google-auth-library-credentials 1.16.0`
- `io.opencensus:opencensus-api 0.31.1`
- `io.opencensus:opencensus-contrib-grpc-metrics 0.31.1`
- `io.opencensus:opencensus-contrib-http-util 0.31.1`
- `io.netty:netty-tcnative-boringssl-static 2.0.59.Final`
- `io.netty:netty-handler 4.1.90.Final`
- `io.netty:netty-common 4.1.90.Final`
- `io.netty:netty-codec-socks 4.1.90.Final`
- `io.netty:netty-codec-http2 4.1.90.Final`
- `io.netty:netty-codec-http 4.1.90.Final`
- `io.netty:netty-codec 4.1.90.Final`
- `io.netty:netty-buffer 4.1.90.Final`
- `io.netty:netty-resolver 4.1.90.Final`
- `io.netty:netty-transport 4.1.90.Final`
- `io.netty:netty-handler-proxy 4.1.90.Final`
- `io.netty:netty-transport-native-unix-common 4.1.90.Final`
- `io.perfmark:perfmark-api 0.26.0`
- `org.apache.tomcat:annotations-api 6.0.53`
- `com.google.code.gson:gson 2.10.1`
- `com.google.guava:guava 31.1-jre`
- `com.google.guava:failureaccess 1.0.1`
- `org.apache.commons:commons-lang3 3.12.0`
- `com.google.android:annotations 4.1.1.4`
- `com.google.code.findbugs:jsr305 3.0.2`
- `com.google.errorprone:error_prone_annotations 2.18.0`
- `com.google.j2objc:j2objc-annotations 1.3`
- `com.google.auto.value:auto-value 1.10.1`
- `com.google.auto.value:auto-value-annotations 1.10.1`
- `com.google.api:api-common 2.6.3`
- `org.threeten:threetenbp 1.6.7`
- `com.google.api.grpc:grpc-google-iam-v1 1.9.3`
- `com.google.api.grpc:proto-google-iam-v1 1.9.3`
- `com.google.http-client:google-http-client 1.43.1`
- `com.google.http-client:google-http-client-gson 1.43.1`
- `org.codehaus.mojo:animal-sniffer-annotations 1.23`
- `javax.annotation:javax.annotation-api 1.3.2`
- `org.graalvm.sdk:graal-sdk 22.3.1`
- `junit:junit 4.13.2`
- `org.mockito:mockito-core 2.28.2`
- `org.hamcrest:hamcrest-core 1.3`
- `com.google.truth:truth 1.1.3`
- `com.googlecode.java-diff-utils:diffutils 1.3.0`
- `net.bytebuddy:byte-buddy 1.14.2`
- `org.objenesis:objenesis 2.6`
</details>
<details><summary>gax-java/dependencies.properties</summary>
- `com.google.protobuf:protobuf-java 3.21.10`
</details>
<details><summary>gax-java/dependencies.properties</summary>
- `com.google.googlejavaformat:google-java-format 1.15.0`
</details>
<details><summary>gax-java/dependencies.properties</summary>
- `io.grpc:grpc-core 1.54.0`
</details>
</blockquote>
</details>
---
- [x] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more rate limited these updates are currently rate limited click on a checkbox below to force their creation now build deps update dependency org jacoco jacoco maven plugin to deps update netty dependencies to final io netty netty buffer io netty netty codec io netty netty codec http io netty netty codec io netty netty codec socks io netty netty common io netty netty handler io netty netty handler proxy io netty netty resolver io netty netty transport io netty netty transport native unix common deps update dependency attrs to deps update dependency charset normalizer to deps update dependency click to deps update dependency gcp releasetool to deps update dependency google api core to deps update dependency google auth to deps update dependency google cloud storage to deps update dependency google to deps update dependency google resumable media to deps update dependency googleapis common protos to deps update dependency importlib metadata to deps update dependency jeepney to deps update dependency to deps update dependency keyring to deps update dependency markupsafe to deps update dependency protobuf to deps update dependency pyjwt to deps update dependency requests to deps update dependency typing extensions to deps update dependency zipp to 🔐 create all rate limited prs at once 🔐 ⚠ dependency lookup warnings ⚠ renovate failed to look up the following dependencies failed to look up maven package com google cloud gapic showcase failed to look up maven package com google cloud gapic showcase parent failed to look up maven package com google api grpc proto gapic showcase failed to look up maven package com google api grpc grpc gapic showcase files affected coverage report pom xml showcase gapic showcase pom xml showcase grpc gapic showcase pom xml showcase pom xml showcase proto gapic showcase pom xml open these updates have all been created already click a checkbox below to force a retry rebase of any pull pull pull pull pull pull pull pull com google api grpc grpc google common protos com google api grpc grpc google iam com google api grpc proto google common protos com google api grpc proto google iam com google api api common pull com google protobuf protobuf java com google protobuf protobuf bom click on this checkbox to rebase all open prs at once ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull pull detected dependencies bazel workspace github actions github workflows ci yaml actions checkout actions setup java actions checkout actions setup java actions setup java actions checkout actions setup java actions setup java actions checkout actions setup java actions checkout actions setup java actions checkout actions checkout actions setup java github workflows downstream native image yaml actions checkout stcarolas setup maven ayltai setup graalvm github workflows nightly yaml actions checkout actions setup java actions checkout actions setup java actions setup java github workflows shared dependencies yaml actions checkout stcarolas setup maven actions setup java github workflows sonar yaml actions checkout actions setup java actions cache actions cache gradle wrapper rules java gapic resources gradle gradle wrapper gradle wrapper properties gradle maven api common java pom xml com google api gapic generator java pom parent snapshot com google auto value auto value annotations com google auto value auto value com google code findbugs com google errorprone error prone annotations junit junit com google truth truth com google auto value auto value coverage report pom xml com google cloud gapic showcase snapshot com google api gax snapshot com google api gax grpc snapshot com google api gax httpjson snapshot com google api api common snapshot org jacoco jacoco maven plugin com coveo fmt maven plugin gapic generator java bom pom xml com google api gapic generator java pom parent snapshot com google api api common snapshot com google api gax bom snapshot com google api grpc grpc google common protos snapshot com google api grpc proto google common protos snapshot com google api grpc proto google iam snapshot com google api grpc proto google iam snapshot com google api grpc proto google iam snapshot com google api grpc grpc google iam snapshot com google api grpc grpc google iam snapshot com google api grpc grpc google iam snapshot gapic generator java pom parent pom xml com google guava guava jre javax annotation javax annotation api com google auth google auth library bom com google code gson gson com google guava guava bom jre com google protobuf protobuf bom io grpc grpc bom com google cloud google cloud shared config org apache maven plugins maven surefire plugin org apache maven plugins maven failsafe plugin org jacoco jacoco maven plugin javax annotation javax annotation api com google auth google auth library bom com google guava guava jre io grpc grpc bom com google code gson gson com google protobuf protobuf bom javax annotation javax annotation api com google protobuf protobuf bom io grpc grpc bom javax annotation javax annotation api com google guava guava bom jre com google protobuf protobuf bom io grpc grpc bom javax annotation javax annotation api gapic generator java pom xml com google api gapic generator java pom parent snapshot com google api gapic generator java bom snapshot org sonatype plugins nexus staging maven plugin kr motd maven os maven plugin org apache maven plugins maven compiler plugin com google auto value auto value com coveo fmt maven plugin com googlecode maven download plugin download maven plugin org xolstice maven plugins protobuf maven plugin org apache maven plugins maven surefire plugin org jacoco jacoco maven plugin org apache maven plugins maven shade plugin com google guava guava jre com google auto value auto value annotations org yaml snakeyaml com google googlejavaformat google java format junit junit com google errorprone error prone annotations com google truth truth io github java diff utils java diff utils gax java gax bom pom xml com google cloud google cloud shared config com google api gax snapshot com google api gax snapshot com google api gax snapshot com google api gax grpc snapshot com google api gax grpc snapshot com google api gax grpc snapshot com google api gax httpjson snapshot com google api gax httpjson snapshot com google api gax httpjson snapshot gax java gax grpc pom xml com google api gax parent snapshot gax java gax httpjson pom xml com google api gax parent snapshot gax java gax pom xml com google api gax parent snapshot gax java pom xml com google api gapic generator java pom parent snapshot com google api api common snapshot org threeten threetenbp com google code findbugs io opencensus opencensus api com google auto value auto value annotations com google auto value auto value junit junit org mockito mockito core com google api gax snapshot com google api gax snapshot com google api grpc proto google common protos snapshot com google api grpc grpc google common protos snapshot org graalvm sdk graal sdk com google http client google http client bom com google truth truth com google auto value auto value org sonatype plugins nexus staging maven plugin java common protos grpc google common protos pom xml com google api grpc google common protos parent snapshot java common protos pom xml com google api gapic generator java pom parent snapshot com google api grpc grpc google common protos snapshot com google api grpc proto google common protos snapshot com google guava guava jre junit junit org apache maven plugins maven project info reports plugin org apache maven plugins maven javadoc plugin java common protos proto google common protos pom xml com google api grpc google common protos parent snapshot java core google cloud core bom pom xml com google api gapic generator java pom parent snapshot com google cloud google cloud core snapshot com google cloud google cloud core grpc snapshot com google cloud google cloud core http snapshot java core google cloud core grpc pom xml com google cloud google cloud core parent snapshot java core google cloud core http pom xml com google cloud google cloud core parent snapshot java core google cloud core pom xml com google cloud google cloud core parent snapshot java core pom xml com google api gapic generator java pom parent snapshot com google cloud google cloud shared dependencies snapshot junit junit com google truth truth org easymock easymock org mockito mockito all java iam grpc google iam pom xml com google cloud google iam parent snapshot java iam grpc google iam pom xml com google cloud google iam parent snapshot java iam grpc google iam pom xml com google cloud google iam parent snapshot java iam pom xml com google api gapic generator java pom parent snapshot com google api gax bom snapshot com google api grpc proto google iam snapshot com google api grpc grpc google iam snapshot com google api grpc proto google common protos snapshot com google api grpc proto google iam snapshot com google api grpc grpc google iam snapshot com google api grpc grpc google iam snapshot com google api grpc proto google iam snapshot junit junit org apache maven plugins maven project info reports plugin org apache maven plugins maven javadoc plugin java iam proto google iam pom xml com google cloud google iam parent snapshot java iam proto google iam pom xml com google cloud google iam parent snapshot java iam proto google iam pom xml com google cloud google iam parent snapshot java shared dependencies dependency convergence check pom xml com google guava guava jre com google cloud tools dependencies java shared dependencies first party dependencies pom xml com google cloud google cloud shared config com google api gapic generator java bom snapshot com google cloud grpc gcp com google code gson gson com google cloud google cloud core bom snapshot com google http client google http client bom com google oauth client google oauth client bom com google api client google api client bom com google cloud google cloud core snapshot com google cloud google cloud core snapshot java shared dependencies pom xml com google api gapic generator java pom parent snapshot com google cloud first party dependencies snapshot com google cloud third party dependencies snapshot java shared dependencies third party dependencies pom xml com google cloud google cloud shared config org apache httpcomponents httpcore org apache httpcomponents httpclient org threeten threetenbp javax annotation javax annotation api org codehaus mojo animal sniffer annotations com google code findbugs com google errorprone error prone annotations com fasterxml jackson jackson bom commons codec commons codec io opencensus opencensus api io opencensus opencensus contrib grpc util io opencensus opencensus contrib http util io opencensus opencensus contrib zpages io opencensus opencensus exporter stats stackdriver io opencensus opencensus exporter trace stackdriver io opencensus opencensus impl io opencensus opencensus impl core org checkerframework checker qual io perfmark perfmark api java shared dependencies upper bound check pom xml com google cloud google cloud shared config com google cloud google cloud shared dependencies snapshot pom xml com coveo fmt maven plugin org apache maven plugins maven deploy plugin org sonatype plugins nexus staging maven plugin showcase gapic showcase pom xml com google cloud gapic showcase parent snapshot org codehaus mojo exec maven plugin org codehaus mojo exec maven plugin org jacoco jacoco maven plugin com coveo fmt maven plugin com googlecode maven download plugin download maven plugin com google truth truth showcase grpc gapic showcase pom xml com google cloud gapic showcase parent snapshot org codehaus mojo exec maven plugin org codehaus mojo exec maven plugin showcase pom xml com google api gapic generator java bom snapshot com google api grpc proto gapic showcase snapshot com google api grpc grpc gapic showcase snapshot com google cloud gapic showcase snapshot junit junit org apache maven plugins maven surefire plugin showcase proto gapic showcase pom xml com google cloud gapic showcase parent snapshot org codehaus mojo exec maven plugin org codehaus mojo exec maven plugin pip requirements kokoro requirements txt attrs cachetools certifi cffi charset normalizer click colorlog cryptography gcp docuploader gcp releasetool google api core google auth google cloud core google cloud storage google google resumable media googleapis common protos idna importlib metadata jeepney keyring markupsafe packaging protobuf modules pycparser pyjwt pyparsing pyperclip python dateutil requests rsa secretstorage six typing extensions zipp regex gax java dependencies properties com google api grpc proto google common protos com google api grpc grpc google common protos com google auth google auth library http com google auth google auth library credentials io opencensus opencensus api io opencensus opencensus contrib grpc metrics io opencensus opencensus contrib http util io netty netty tcnative boringssl static final io netty netty handler final io netty netty common final io netty netty codec socks final io netty netty codec final io netty netty codec http final io netty netty codec final io netty netty buffer final io netty netty resolver final io netty netty transport final io netty netty handler proxy final io netty netty transport native unix common final io perfmark perfmark api org apache tomcat annotations api com google code gson gson com google guava guava jre com google guava failureaccess org apache commons commons com google android annotations com google code findbugs com google errorprone error prone annotations com google annotations com google auto value auto value com google auto value auto value annotations com google api api common org threeten threetenbp com google api grpc grpc google iam com google api grpc proto google iam com google http client google http client com google http client google http client gson org codehaus mojo animal sniffer annotations javax annotation javax annotation api org graalvm sdk graal sdk junit junit org mockito mockito core org hamcrest hamcrest core com google truth truth com googlecode java diff utils diffutils net bytebuddy byte buddy org objenesis objenesis gax java dependencies properties com google protobuf protobuf java gax java dependencies properties com google googlejavaformat google java format gax java dependencies properties io grpc grpc core check this box to trigger a request for renovate to run again on this repository
| 1
|
82,903
| 16,060,119,227
|
IssuesEvent
|
2021-04-23 11:18:17
|
towavephone/GatsbyBlog
|
https://api.github.com/repos/towavephone/GatsbyBlog
|
opened
|
微前端框架qiankun源码笔记
|
/qiankun-code-note/ Gitalk
|
/qiankun-code-note/定义 是一种类似于微服务的架构,它将微服务的理念应用于浏览器端,即将单页面前端应用由单一的单体应用转变为多个小型前端应用聚合为一的应用。各个前端应用还可以独立开发、独立部署。同时,它们也可以在共享组件的同时进行并行开发——这些组件可以通过 NPM 或者 Git Tag、Git…
|
1.0
|
微前端框架qiankun源码笔记 - /qiankun-code-note/定义 是一种类似于微服务的架构,它将微服务的理念应用于浏览器端,即将单页面前端应用由单一的单体应用转变为多个小型前端应用聚合为一的应用。各个前端应用还可以独立开发、独立部署。同时,它们也可以在共享组件的同时进行并行开发——这些组件可以通过 NPM 或者 Git Tag、Git…
|
non_process
|
微前端框架qiankun源码笔记 qiankun code note 定义 是一种类似于微服务的架构,它将微服务的理念应用于浏览器端,即将单页面前端应用由单一的单体应用转变为多个小型前端应用聚合为一的应用。各个前端应用还可以独立开发、独立部署。同时,它们也可以在共享组件的同时进行并行开发——这些组件可以通过 npm 或者 git tag、git…
| 0
|
13,711
| 16,469,473,594
|
IssuesEvent
|
2021-05-23 05:37:17
|
pat-rogers/Ada-202x-WG9-Informal-Review
|
https://api.github.com/repos/pat-rogers/Ada-202x-WG9-Informal-Review
|
closed
|
5.6.1(4/5) and others, example not preceded by our usual "Example of ..." lead in
|
duplicate processed
|
Originator: Tucker Taft (via ARG list)
Paragraph: 5.6.1(4/5), 5.5.3(30/5), 5.5.3(34/5), 6.3(8)
The example in section 5.6.1 (Parallel Block Statements) is not preceded by
our usual "Example of ..." lead in.
In RM 6.3, we have "Example of procedure body:" and "Example of a function
body:" in the same set of examples. It would seem we should add "a" to the
first lead in. [Editor's notes: Tucker had the wrong subclause reference -
he probably got it from the bottom of the page, which shows the *last*
subclause on a page. No idea why Word does that. Anyway, this has been true
since Ada 9x - it's in the Sept 1993 edition. Must be important. ;-)]
In RM 5.5.3 (Procedural Iterators) the lead ins to the examples are not
italicized (paragraphs 30/5 and 34/5).
In RM 5.5.2 (Generalized Loop Iteration) the example lead in is unlike most
others. It is a comment in the same font size as the example. Normally the
lead in is larger, italicized, and starts with "Example[s] of ...".
|
1.0
|
5.6.1(4/5) and others, example not preceded by our usual "Example of ..." lead in - Originator: Tucker Taft (via ARG list)
Paragraph: 5.6.1(4/5), 5.5.3(30/5), 5.5.3(34/5), 6.3(8)
The example in section 5.6.1 (Parallel Block Statements) is not preceded by
our usual "Example of ..." lead in.
In RM 6.3, we have "Example of procedure body:" and "Example of a function
body:" in the same set of examples. It would seem we should add "a" to the
first lead in. [Editor's notes: Tucker had the wrong subclause reference -
he probably got it from the bottom of the page, which shows the *last*
subclause on a page. No idea why Word does that. Anyway, this has been true
since Ada 9x - it's in the Sept 1993 edition. Must be important. ;-)]
In RM 5.5.3 (Procedural Iterators) the lead ins to the examples are not
italicized (paragraphs 30/5 and 34/5).
In RM 5.5.2 (Generalized Loop Iteration) the example lead in is unlike most
others. It is a comment in the same font size as the example. Normally the
lead in is larger, italicized, and starts with "Example[s] of ...".
|
process
|
and others example not preceded by our usual example of lead in originator tucker taft via arg list paragraph the example in section parallel block statements is not preceded by our usual example of lead in in rm we have example of procedure body and example of a function body in the same set of examples it would seem we should add a to the first lead in editor s notes tucker had the wrong subclause reference he probably got it from the bottom of the page which shows the last subclause on a page no idea why word does that anyway this has been true since ada it s in the sept edition must be important in rm procedural iterators the lead ins to the examples are not italicized paragraphs and in rm generalized loop iteration the example lead in is unlike most others it is a comment in the same font size as the example normally the lead in is larger italicized and starts with example of
| 1
|
22,351
| 31,027,533,588
|
IssuesEvent
|
2023-08-10 10:08:57
|
DxytJuly3/gitalk_blog
|
https://api.github.com/repos/DxytJuly3/gitalk_blog
|
opened
|
[Linux] 进程状态相关概念、Linux实际进程状态、进程优先级 - July.cc Blogs
|
Gitalk /posts/Linux-Process-States
|
https://www.julysblog.cn/posts/Linux-Process-States
生成进程之后, 操作系统会给进程分配其所需要的资源。而为了整个操作系统的稳定和效率, 操作系统会对进程进行调度、管理, 在调度管理的过程中进程会存在不同的状态, 这些不同的状态就被称为进程状态。
|
1.0
|
[Linux] 进程状态相关概念、Linux实际进程状态、进程优先级 - July.cc Blogs - https://www.julysblog.cn/posts/Linux-Process-States
生成进程之后, 操作系统会给进程分配其所需要的资源。而为了整个操作系统的稳定和效率, 操作系统会对进程进行调度、管理, 在调度管理的过程中进程会存在不同的状态, 这些不同的状态就被称为进程状态。
|
process
|
进程状态相关概念、linux实际进程状态、进程优先级 july cc blogs 生成进程之后 操作系统会给进程分配其所需要的资源。而为了整个操作系统的稳定和效率 操作系统会对进程进行调度、管理 在调度管理的过程中进程会存在不同的状态 这些不同的状态就被称为进程状态。
| 1
|
7,304
| 10,443,164,173
|
IssuesEvent
|
2019-09-18 14:23:51
|
threefoldtech/0-core
|
https://api.github.com/repos/threefoldtech/0-core
|
closed
|
add primitives to move a nic into a network namespace
|
process_wontfix type_feature
|
it is a primitive for `ip link set $interface netns @namespace`
|
1.0
|
add primitives to move a nic into a network namespace - it is a primitive for `ip link set $interface netns @namespace`
|
process
|
add primitives to move a nic into a network namespace it is a primitive for ip link set interface netns namespace
| 1
|
6,324
| 9,345,009,826
|
IssuesEvent
|
2019-03-30 03:04:23
|
bow-simulation/virtualbow
|
https://api.github.com/repos/bow-simulation/virtualbow
|
closed
|
Consider vcpkg for dependency management
|
area: software process type: improvement
|
In GitLab by **spfeifer** on Dec 5, 2018, 16:32
Anouncement:
https://blogs.msdn.microsoft.com/vcblog/2018/04/24/announcing-a-single-c-library-manager-for-linux-macos-and-windows-vcpkg/
Discussions:
https://www.reddit.com/r/cpp/comments/8emja5/vcpkg_library_manager_now_also_on_linux_and_macos/
https://www.reddit.com/r/cpp/comments/9m4l0p/conan_vcpkg_or_build2/
Available libraries so far:
https://github.com/Microsoft/vcpkg/tree/master/ports
Setting specific library versions:
https://github.com/Microsoft/vcpkg/issues/1681
|
1.0
|
Consider vcpkg for dependency management - In GitLab by **spfeifer** on Dec 5, 2018, 16:32
Anouncement:
https://blogs.msdn.microsoft.com/vcblog/2018/04/24/announcing-a-single-c-library-manager-for-linux-macos-and-windows-vcpkg/
Discussions:
https://www.reddit.com/r/cpp/comments/8emja5/vcpkg_library_manager_now_also_on_linux_and_macos/
https://www.reddit.com/r/cpp/comments/9m4l0p/conan_vcpkg_or_build2/
Available libraries so far:
https://github.com/Microsoft/vcpkg/tree/master/ports
Setting specific library versions:
https://github.com/Microsoft/vcpkg/issues/1681
|
process
|
consider vcpkg for dependency management in gitlab by spfeifer on dec anouncement discussions available libraries so far setting specific library versions
| 1
|
4,358
| 7,260,513,756
|
IssuesEvent
|
2018-02-18 10:53:06
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE][Processing] Minimal enclosing circle
|
Automatic new feature Processing
|
Original commit: https://github.com/qgis/QGIS/commit/e30f7044c9af22096dda8722d188d452228d420c by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
1.0
|
[FEATURE][Processing] Minimal enclosing circle - Original commit: https://github.com/qgis/QGIS/commit/e30f7044c9af22096dda8722d188d452228d420c by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
process
|
minimal enclosing circle original commit by nyalldawson unfortunately this naughty coder did not write a description
| 1
|
78
| 2,531,050,807
|
IssuesEvent
|
2015-01-23 01:12:30
|
tinkerpop/tinkerpop3
|
https://api.github.com/repos/tinkerpop/tinkerpop3
|
closed
|
Add branch() to GraphTraversal.
|
enhancement process
|
This should probably happen after you figure out how to elegantly make JumpStep extend it and thus, solidify the BranchStep API.
|
1.0
|
Add branch() to GraphTraversal. - This should probably happen after you figure out how to elegantly make JumpStep extend it and thus, solidify the BranchStep API.
|
process
|
add branch to graphtraversal this should probably happen after you figure out how to elegantly make jumpstep extend it and thus solidify the branchstep api
| 1
|
742,421
| 25,853,684,821
|
IssuesEvent
|
2022-12-13 12:13:47
|
pystardust/ani-cli
|
https://api.github.com/repos/pystardust/ani-cli
|
closed
|
Create directory if needed
|
type: feature request priority 4: wishlist category: ui ux good first issue
|
**Is your feature request related to a problem? Please describe.**
Currently if you want to download to a specified folder (flag `-p`) and the folder doesn't exist, the download fails
**Describe the solution you'd like**
Detect if folder doesn't exist, and create it (using `mkdir -p`). Or we don't create a folder at all, but exit right on startup with an error message. Or default to current folder (with a verbose warning).
**Describe alternatives you've considered**
Using `mkdir` without the `-p` flag might be more stable
**Additional context**
Feature was discussed on [discord](https://discord.com/channels/931186431215435807/932006490703822938/1008361758341222521)
|
1.0
|
Create directory if needed - **Is your feature request related to a problem? Please describe.**
Currently if you want to download to a specified folder (flag `-p`) and the folder doesn't exist, the download fails
**Describe the solution you'd like**
Detect if folder doesn't exist, and create it (using `mkdir -p`). Or we don't create a folder at all, but exit right on startup with an error message. Or default to current folder (with a verbose warning).
**Describe alternatives you've considered**
Using `mkdir` without the `-p` flag might be more stable
**Additional context**
Feature was discussed on [discord](https://discord.com/channels/931186431215435807/932006490703822938/1008361758341222521)
|
non_process
|
create directory if needed is your feature request related to a problem please describe currently if you want to download to a specified folder flag p and the folder doesn t exist the download fails describe the solution you d like detect if folder doesn t exist and create it using mkdir p or we don t create a folder at all but exit right on startup with an error message or default to current folder with a verbose warning describe alternatives you ve considered using mkdir without the p flag might be more stable additional context feature was discussed on
| 0
|
27,048
| 13,168,820,911
|
IssuesEvent
|
2020-08-11 12:48:00
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
reopened
|
FNX3-16565 ⁃ Fenix should run places maintenance periodically
|
P3 eng:health eng:performance
|
@grigoryk indicated that it does not, which will result in worse perf, more disk usage, etc.
|
True
|
FNX3-16565 ⁃ Fenix should run places maintenance periodically - @grigoryk indicated that it does not, which will result in worse perf, more disk usage, etc.
|
non_process
|
⁃ fenix should run places maintenance periodically grigoryk indicated that it does not which will result in worse perf more disk usage etc
| 0
|
21,768
| 30,287,305,883
|
IssuesEvent
|
2023-07-08 21:13:08
|
MikaylaFischler/cc-mek-scada
|
https://api.github.com/repos/MikaylaFischler/cc-mek-scada
|
closed
|
Process Waste Control
|
supervisor coordinator process control
|
Automatic (coordinated) waste processing control.
Support optionally switching to plutonium production during the night when polonium processing is not possible, which would also apply to switching off of antimatter.
- [x] Auto waste control interface
- [x] Coordinated waste control (all units assigned to the same target)
- [x] Day/night balancing option requiring the below task
- [x] Solar neutron activator integration
- [x] SPS integration (basic status)
Auto control will always control all units assigned to the same output. If the user wants some to do one thing and others to do another, that's what manual assignment is for.
|
1.0
|
Process Waste Control - Automatic (coordinated) waste processing control.
Support optionally switching to plutonium production during the night when polonium processing is not possible, which would also apply to switching off of antimatter.
- [x] Auto waste control interface
- [x] Coordinated waste control (all units assigned to the same target)
- [x] Day/night balancing option requiring the below task
- [x] Solar neutron activator integration
- [x] SPS integration (basic status)
Auto control will always control all units assigned to the same output. If the user wants some to do one thing and others to do another, that's what manual assignment is for.
|
process
|
process waste control automatic coordinated waste processing control support optionally switching to plutonium production during the night when polonium processing is not possible which would also apply to switching off of antimatter auto waste control interface coordinated waste control all units assigned to the same target day night balancing option requiring the below task solar neutron activator integration sps integration basic status auto control will always control all units assigned to the same output if the user wants some to do one thing and others to do another that s what manual assignment is for
| 1
|
16,857
| 22,137,876,638
|
IssuesEvent
|
2022-06-03 02:05:17
|
hashgraph/hedera-json-rpc-relay
|
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
|
opened
|
Add artifact build and upload to 0.1 release branch
|
enhancement P1 process
|
### Problem
`release/0.1` is missing logic to archive and upload the npm package.
### Solution
Add logic to archive and upload artifact
### Alternatives
_No response_
|
1.0
|
Add artifact build and upload to 0.1 release branch - ### Problem
`release/0.1` is missing logic to archive and upload the npm package.
### Solution
Add logic to archive and upload artifact
### Alternatives
_No response_
|
process
|
add artifact build and upload to release branch problem release is missing logic to archive and upload the npm package solution add logic to archive and upload artifact alternatives no response
| 1
|
93,526
| 8,432,995,197
|
IssuesEvent
|
2018-10-17 05:17:22
|
hazelcast/hazelcast-jet
|
https://api.github.com/repos/hazelcast/hazelcast-jet
|
closed
|
com.hazelcast.jet.impl.connector.ReadWithPartitionIteratorP_MigrationDetectionTest.when_migration_then_detected_remote
|
test-failure
|
https://hazelcast-l337.ci.cloudbees.com/view/Jet/job/jet/com.hazelcast.jet$hazelcast-jet-core/927/testReport/junit/com.hazelcast.jet.impl.connector/ReadWithPartitionIteratorP_MigrationDetectionTest/when_migration_then_detected_remote/
```
Error Message
Expected test to throw (an instance of java.util.concurrent.ExecutionException and exception with message a string containing "migration detected")
Stacktrace
java.lang.AssertionError: Expected test to throw (an instance of java.util.concurrent.ExecutionException and exception with message a string containing "migration detected")
at org.junit.Assert.fail(Assert.java:88)
at org.junit.rules.ExpectedException.failDueToMissingException(ExpectedException.java:263)
at org.junit.rules.ExpectedException.access$200(ExpectedException.java:106)
at org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:245)
at com.hazelcast.test.DumpBuildInfoOnFailureRule$1.evaluate(DumpBuildInfoOnFailureRule.java:37)
```
|
1.0
|
com.hazelcast.jet.impl.connector.ReadWithPartitionIteratorP_MigrationDetectionTest.when_migration_then_detected_remote - https://hazelcast-l337.ci.cloudbees.com/view/Jet/job/jet/com.hazelcast.jet$hazelcast-jet-core/927/testReport/junit/com.hazelcast.jet.impl.connector/ReadWithPartitionIteratorP_MigrationDetectionTest/when_migration_then_detected_remote/
```
Error Message
Expected test to throw (an instance of java.util.concurrent.ExecutionException and exception with message a string containing "migration detected")
Stacktrace
java.lang.AssertionError: Expected test to throw (an instance of java.util.concurrent.ExecutionException and exception with message a string containing "migration detected")
at org.junit.Assert.fail(Assert.java:88)
at org.junit.rules.ExpectedException.failDueToMissingException(ExpectedException.java:263)
at org.junit.rules.ExpectedException.access$200(ExpectedException.java:106)
at org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:245)
at com.hazelcast.test.DumpBuildInfoOnFailureRule$1.evaluate(DumpBuildInfoOnFailureRule.java:37)
```
|
non_process
|
com hazelcast jet impl connector readwithpartitioniteratorp migrationdetectiontest when migration then detected remote error message expected test to throw an instance of java util concurrent executionexception and exception with message a string containing migration detected stacktrace java lang assertionerror expected test to throw an instance of java util concurrent executionexception and exception with message a string containing migration detected at org junit assert fail assert java at org junit rules expectedexception failduetomissingexception expectedexception java at org junit rules expectedexception access expectedexception java at org junit rules expectedexception expectedexceptionstatement evaluate expectedexception java at com hazelcast test dumpbuildinfoonfailurerule evaluate dumpbuildinfoonfailurerule java
| 0
|
3,312
| 2,755,737,677
|
IssuesEvent
|
2015-04-26 22:17:56
|
PerlDancer/Dancer2
|
https://api.github.com/repos/PerlDancer/Dancer2
|
closed
|
Make prefix a full spec like a path
|
Beginner Suitable Documentation Enhancement Testing
|
Currently the `prefix` can only be a simple string, it cannot be a regexp in and of itself. We should allow it to have the same spec as a regular path.
This is mentioned (and, in a way, is part of) #558.
|
1.0
|
Make prefix a full spec like a path - Currently the `prefix` can only be a simple string, it cannot be a regexp in and of itself. We should allow it to have the same spec as a regular path.
This is mentioned (and, in a way, is part of) #558.
|
non_process
|
make prefix a full spec like a path currently the prefix can only be a simple string it cannot be a regexp in and of itself we should allow it to have the same spec as a regular path this is mentioned and in a way is part of
| 0
|
15,006
| 18,720,010,133
|
IssuesEvent
|
2021-11-03 10:40:53
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Fix 'modulation of complement activation in other organism' and children
|
multi-species process
|
Hello,
I think the following terms should be 'in host' rather than 'in other organism':
'modulation of complement activation in other organism'
and children:
'modulation of complement activation, alternative pathway in other organism'
'negative regulation of complement activation, alternative pathway in other organism'
'positive regulation of complement activation, alternative pathway in other organism'
'modulation of complement activation, classical pathway in other organism'
'negative regulation of complement activation, classical pathway in other organism'
'positive regulation of complement activation, classical pathway in other organism'
'modulation of complement activation, lectin pathway in other organism'
'negative regulation of complement activation, lectin pathway in other organism'
'positive regulation of complement activation, lectin pathway in other organism'
'modulation of mast cell degranulation in other organism'
'positive regulation of mast cell degranulation in other organism'
(note that there are no annotations to any of these terms)
OK @mgiglio99 @addiehl @ValWood ?
Thanks, Pascale
|
1.0
|
Fix 'modulation of complement activation in other organism' and children - Hello,
I think the following terms should be 'in host' rather than 'in other organism':
'modulation of complement activation in other organism'
and children:
'modulation of complement activation, alternative pathway in other organism'
'negative regulation of complement activation, alternative pathway in other organism'
'positive regulation of complement activation, alternative pathway in other organism'
'modulation of complement activation, classical pathway in other organism'
'negative regulation of complement activation, classical pathway in other organism'
'positive regulation of complement activation, classical pathway in other organism'
'modulation of complement activation, lectin pathway in other organism'
'negative regulation of complement activation, lectin pathway in other organism'
'positive regulation of complement activation, lectin pathway in other organism'
'modulation of mast cell degranulation in other organism'
'positive regulation of mast cell degranulation in other organism'
(note that there are no annotations to any of these terms)
OK @mgiglio99 @addiehl @ValWood ?
Thanks, Pascale
|
process
|
fix modulation of complement activation in other organism and children hello i think the following terms should be in host rather than in other organism modulation of complement activation in other organism and children modulation of complement activation alternative pathway in other organism negative regulation of complement activation alternative pathway in other organism positive regulation of complement activation alternative pathway in other organism modulation of complement activation classical pathway in other organism negative regulation of complement activation classical pathway in other organism positive regulation of complement activation classical pathway in other organism modulation of complement activation lectin pathway in other organism negative regulation of complement activation lectin pathway in other organism positive regulation of complement activation lectin pathway in other organism modulation of mast cell degranulation in other organism positive regulation of mast cell degranulation in other organism note that there are no annotations to any of these terms ok addiehl valwood thanks pascale
| 1
|
2,709
| 5,578,337,699
|
IssuesEvent
|
2017-03-28 12:08:53
|
deb-sandeep/JoveNotesWebApp
|
https://api.github.com/repos/deb-sandeep/JoveNotesWebApp
|
closed
|
Introduction of @exercise type
|
enhancement jove_notes_db jove_notes_grammar jove_notes_processor jove_notes_server jove_notes_ui
|
## Support for exercise questions
### Background
In its current form JoveNotes is primarily focused towards the bulk of study for K1-8 classes - retention and recollection (RnR). However as we move towards classes XI and beyond, there is a marked shift in focus towards applicability and problem solving.
This enhancement is an endeavor to stretch the envelop of JoveNotes to foray into the arena of applicability.
The current vision is to incorporate the following aspects of applicability class of problems:
1. Capture / Digitization - of exercise questions
2. Presentment
3. Operational tracking
4. Collecting data points
5. Providing insights by analyzing data points
6. Scoring of points - extension to current gamification
### In which way are exercise different than the current note element types?
Exercise problems (for example, essays, numericals, problem solving etc) are characteristically different from existing note element types, fundamentally because it addresses a different segment of the education topology - applicability, while the existing note elements address the retention and recollection part.
Salient points which differentiates exercises from other notes elements are:
- **Presentment** - Exercises should not be mixed with RnR presentment, especially during the flash card sessions for the following reasons:
- RnR flash cards sessions are rapid fire sessions with an average turn around time per question around 15-20 seconds. Numericals will not fit into the scheme - gear shift jerks.
- RnR flash cards are designed to funnel focus on the screen - Numericals will have to be solved manually, causing a loss of focus on RnR streak.
- Numericals require a cluster attempt behavior - read the questions paper, prioritize, review, rework, mark and then submit. This is contrary to the atomic way RnR questions are presented.
- **Operational tracking** -
- For one, numericals don't require the five step spaced sequenced repetition like those of RnR.
- The time span between re-presentment is different as compared to RnR
- Lot more data needs to be collected at session level - pre-read time, work, review, rework etc
## Modules impacted
1. JoveNotes grammar
- com.sandy.xtext.jovenotes
- com.sandy.xtext.jovenotes.ui
- com.sandy.xtext.jovenotes.tests
2. JoveNotes processor
3. Database
4. JoveNotesWebApp
- Dashboard
- New numericals section
5. JoveNotesMaker
## Feature branch
`feature/@exercise`
## Example of DSL
### DSL Grammar
```
'@exercise' ( hideFromView = 'hide' )?
( 'marks' '=' marks = INT )
( question = STRING )
( 'hints' '{' ( hints += STRING )+ '}' )?
( 'answer' answer = STRING )
```
### DSL example
``` java
@exercise marks=20
"What is the velocity of an object after 5 seconds which starts from rest
and accelerating for 2 seconds at 5 m/s^2?"
hints {
"If the body starts from rest, initial velocity is zero"
"Once the body stops accelerating, velocity remains constant"
}
answer "10 m/s"
```
|
1.0
|
Introduction of @exercise type - ## Support for exercise questions
### Background
In its current form JoveNotes is primarily focused towards the bulk of study for K1-8 classes - retention and recollection (RnR). However as we move towards classes XI and beyond, there is a marked shift in focus towards applicability and problem solving.
This enhancement is an endeavor to stretch the envelop of JoveNotes to foray into the arena of applicability.
The current vision is to incorporate the following aspects of applicability class of problems:
1. Capture / Digitization - of exercise questions
2. Presentment
3. Operational tracking
4. Collecting data points
5. Providing insights by analyzing data points
6. Scoring of points - extension to current gamification
### In which way are exercise different than the current note element types?
Exercise problems (for example, essays, numericals, problem solving etc) are characteristically different from existing note element types, fundamentally because it addresses a different segment of the education topology - applicability, while the existing note elements address the retention and recollection part.
Salient points which differentiates exercises from other notes elements are:
- **Presentment** - Exercises should not be mixed with RnR presentment, especially during the flash card sessions for the following reasons:
- RnR flash cards sessions are rapid fire sessions with an average turn around time per question around 15-20 seconds. Numericals will not fit into the scheme - gear shift jerks.
- RnR flash cards are designed to funnel focus on the screen - Numericals will have to be solved manually, causing a loss of focus on RnR streak.
- Numericals require a cluster attempt behavior - read the questions paper, prioritize, review, rework, mark and then submit. This is contrary to the atomic way RnR questions are presented.
- **Operational tracking** -
- For one, numericals don't require the five step spaced sequenced repetition like those of RnR.
- The time span between re-presentment is different as compared to RnR
- Lot more data needs to be collected at session level - pre-read time, work, review, rework etc
## Modules impacted
1. JoveNotes grammar
- com.sandy.xtext.jovenotes
- com.sandy.xtext.jovenotes.ui
- com.sandy.xtext.jovenotes.tests
2. JoveNotes processor
3. Database
4. JoveNotesWebApp
- Dashboard
- New numericals section
5. JoveNotesMaker
## Feature branch
`feature/@exercise`
## Example of DSL
### DSL Grammar
```
'@exercise' ( hideFromView = 'hide' )?
( 'marks' '=' marks = INT )
( question = STRING )
( 'hints' '{' ( hints += STRING )+ '}' )?
( 'answer' answer = STRING )
```
### DSL example
``` java
@exercise marks=20
"What is the velocity of an object after 5 seconds which starts from rest
and accelerating for 2 seconds at 5 m/s^2?"
hints {
"If the body starts from rest, initial velocity is zero"
"Once the body stops accelerating, velocity remains constant"
}
answer "10 m/s"
```
|
process
|
introduction of exercise type support for exercise questions background in its current form jovenotes is primarily focused towards the bulk of study for classes retention and recollection rnr however as we move towards classes xi and beyond there is a marked shift in focus towards applicability and problem solving this enhancement is an endeavor to stretch the envelop of jovenotes to foray into the arena of applicability the current vision is to incorporate the following aspects of applicability class of problems capture digitization of exercise questions presentment operational tracking collecting data points providing insights by analyzing data points scoring of points extension to current gamification in which way are exercise different than the current note element types exercise problems for example essays numericals problem solving etc are characteristically different from existing note element types fundamentally because it addresses a different segment of the education topology applicability while the existing note elements address the retention and recollection part salient points which differentiates exercises from other notes elements are presentment exercises should not be mixed with rnr presentment especially during the flash card sessions for the following reasons rnr flash cards sessions are rapid fire sessions with an average turn around time per question around seconds numericals will not fit into the scheme gear shift jerks rnr flash cards are designed to funnel focus on the screen numericals will have to be solved manually causing a loss of focus on rnr streak numericals require a cluster attempt behavior read the questions paper prioritize review rework mark and then submit this is contrary to the atomic way rnr questions are presented operational tracking for one numericals don t require the five step spaced sequenced repetition like those of rnr the time span between re presentment is different as compared to rnr lot more data needs to be collected at session level pre read time work review rework etc modules impacted jovenotes grammar com sandy xtext jovenotes com sandy xtext jovenotes ui com sandy xtext jovenotes tests jovenotes processor database jovenoteswebapp dashboard new numericals section jovenotesmaker feature branch feature exercise example of dsl dsl grammar exercise hidefromview hide marks marks int question string hints hints string answer answer string dsl example java exercise marks what is the velocity of an object after seconds which starts from rest and accelerating for seconds at m s hints if the body starts from rest initial velocity is zero once the body stops accelerating velocity remains constant answer m s
| 1
|
262,861
| 8,272,561,480
|
IssuesEvent
|
2018-09-16 21:30:13
|
javaee/glassfish
|
https://api.github.com/repos/javaee/glassfish
|
closed
|
Monitoring support for @DataSourceDefinition
|
Component: jdbc ERR: Assignee Priority: Major Type: New Feature
|
Existing implementation of @DataSourceDefinition is realized internally as a jdbc-resource and jdbc-connection-pool. It would be useful to expose monitoring statistics of @DataSourceDefinition (ie., the jdbc-connection-pool monitoring statistics) to users. Following monitoring hierarchy could be used to display the statistics according to the scope in which the @DataSourceDefinition is defined.
server.<application-name>.datasource-definition.java:global/<global-scoped-dsd>
server.<application-name>.datasource-definition.java:app/<app-scoped-dsd>
server.<application-name>.datasource-definition.<module-name>.java:module/<module-scoped-dsd>
server.<application-name>.datasource-definition.<module-name>.<component-name>.java:comp/<component-scoped-dsd>
Note :
1) For component scoped @DSD in web bundle descriptors, following convention will be used
server.<application-name>.<module-name>.datasource-definition.java:comp/<component-scoped-dsd>
2) For applications that are not .ear, <module-name> will not be applicable.
#### Affected Versions
[4.0]
|
1.0
|
Monitoring support for @DataSourceDefinition - Existing implementation of @DataSourceDefinition is realized internally as a jdbc-resource and jdbc-connection-pool. It would be useful to expose monitoring statistics of @DataSourceDefinition (ie., the jdbc-connection-pool monitoring statistics) to users. Following monitoring hierarchy could be used to display the statistics according to the scope in which the @DataSourceDefinition is defined.
server.<application-name>.datasource-definition.java:global/<global-scoped-dsd>
server.<application-name>.datasource-definition.java:app/<app-scoped-dsd>
server.<application-name>.datasource-definition.<module-name>.java:module/<module-scoped-dsd>
server.<application-name>.datasource-definition.<module-name>.<component-name>.java:comp/<component-scoped-dsd>
Note :
1) For component scoped @DSD in web bundle descriptors, following convention will be used
server.<application-name>.<module-name>.datasource-definition.java:comp/<component-scoped-dsd>
2) For applications that are not .ear, <module-name> will not be applicable.
#### Affected Versions
[4.0]
|
non_process
|
monitoring support for datasourcedefinition existing implementation of datasourcedefinition is realized internally as a jdbc resource and jdbc connection pool it would be useful to expose monitoring statistics of datasourcedefinition ie the jdbc connection pool monitoring statistics to users following monitoring hierarchy could be used to display the statistics according to the scope in which the datasourcedefinition is defined server datasource definition java global server datasource definition java app server datasource definition java module server datasource definition java comp note for component scoped dsd in web bundle descriptors following convention will be used server datasource definition java comp for applications that are not ear will not be applicable affected versions
| 0
|
339,568
| 10,256,216,394
|
IssuesEvent
|
2019-08-21 17:07:41
|
cerner/terra-core
|
https://api.github.com/repos/cerner/terra-core
|
closed
|
Update components which interface with old React context API
|
Intl Orion Reviewed i18n priority:medium
|
# Feature Request
## Description
In a lot of the terra components that interface with react-intl, [we've been interfacing with the old React context API](https://github.com/cerner/terra-core/blob/master/packages/terra-demographics-banner/src/DemographicsBanner.jsx#L67-L74) directly. We should update the components to use an abstraction like `FormattedMessage` or `injectIntl` to interface with react-intl's Intl context. This should allows us to more easily update to react-intl v3 when it is released.
More info on how we recommend to build components with `FormattedMessage` and/or `injectIntl`: https://engineering.cerner.com/terra-ui/#/getting-started/terra-ui/internationalization/building-components-which-include-translations
|
1.0
|
Update components which interface with old React context API - # Feature Request
## Description
In a lot of the terra components that interface with react-intl, [we've been interfacing with the old React context API](https://github.com/cerner/terra-core/blob/master/packages/terra-demographics-banner/src/DemographicsBanner.jsx#L67-L74) directly. We should update the components to use an abstraction like `FormattedMessage` or `injectIntl` to interface with react-intl's Intl context. This should allows us to more easily update to react-intl v3 when it is released.
More info on how we recommend to build components with `FormattedMessage` and/or `injectIntl`: https://engineering.cerner.com/terra-ui/#/getting-started/terra-ui/internationalization/building-components-which-include-translations
|
non_process
|
update components which interface with old react context api feature request description in a lot of the terra components that interface with react intl directly we should update the components to use an abstraction like formattedmessage or injectintl to interface with react intl s intl context this should allows us to more easily update to react intl when it is released more info on how we recommend to build components with formattedmessage and or injectintl
| 0
|
67,931
| 7,078,965,018
|
IssuesEvent
|
2018-01-10 07:27:50
|
GTNewHorizons/NewHorizons
|
https://api.github.com/repos/GTNewHorizons/NewHorizons
|
closed
|
Cetane-Boosted Diesel from Diesel not profit
|
FixedInDev need to be tested
|
#### Which modpack version are you using?
2.0.1.3
We can craft 500 mb Cetane-Boosted Diesel from 1000 mb Diesel in mixer. But if we count, 500 mb Cetane-Boosted Diesel give us 360000 eu, but 1000 mb Diesel give us 400000 eu.
I suggest to change Cetane-Boosted Diesel output to 1000 mb.
|
1.0
|
Cetane-Boosted Diesel from Diesel not profit - #### Which modpack version are you using?
2.0.1.3
We can craft 500 mb Cetane-Boosted Diesel from 1000 mb Diesel in mixer. But if we count, 500 mb Cetane-Boosted Diesel give us 360000 eu, but 1000 mb Diesel give us 400000 eu.
I suggest to change Cetane-Boosted Diesel output to 1000 mb.
|
non_process
|
cetane boosted diesel from diesel not profit which modpack version are you using we can craft mb cetane boosted diesel from mb diesel in mixer but if we count mb cetane boosted diesel give us eu but mb diesel give us eu i suggest to change cetane boosted diesel output to mb
| 0
|
7,931
| 11,114,444,329
|
IssuesEvent
|
2019-12-18 08:41:16
|
ocaml-batteries-team/batteries-included
|
https://api.github.com/repos/ocaml-batteries-team/batteries-included
|
opened
|
make v3 the new master
|
Easy development process
|
And have a maintenance branch named v2 or 2.x or whatever, for any upcoming fixes on the v2.
WDYT?
Have some time to do it quickly if everybody agree.
|
1.0
|
make v3 the new master - And have a maintenance branch named v2 or 2.x or whatever, for any upcoming fixes on the v2.
WDYT?
Have some time to do it quickly if everybody agree.
|
process
|
make the new master and have a maintenance branch named or x or whatever for any upcoming fixes on the wdyt have some time to do it quickly if everybody agree
| 1
|
19,097
| 25,148,011,302
|
IssuesEvent
|
2022-11-10 07:41:48
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Outputs attributes tables from Processing/SAGA have crippled special chars if input is not a shapefile
|
Processing Bug
|
Take a vector with special/accented characters in its attribute table.
If you processing this with Processing/SAGA and the input format is not a shapefile then the output will have those chars crippled.
This do not happen if the input is a shapefile and also do not happen if running from the CLI the same SAGA command that QGIS runs.
|
1.0
|
Outputs attributes tables from Processing/SAGA have crippled special chars if input is not a shapefile - Take a vector with special/accented characters in its attribute table.
If you processing this with Processing/SAGA and the input format is not a shapefile then the output will have those chars crippled.
This do not happen if the input is a shapefile and also do not happen if running from the CLI the same SAGA command that QGIS runs.
|
process
|
outputs attributes tables from processing saga have crippled special chars if input is not a shapefile take a vector with special accented characters in its attribute table if you processing this with processing saga and the input format is not a shapefile then the output will have those chars crippled this do not happen if the input is a shapefile and also do not happen if running from the cli the same saga command that qgis runs
| 1
|
125,571
| 26,688,403,993
|
IssuesEvent
|
2023-01-27 00:53:08
|
WordPress/openverse-catalog
|
https://api.github.com/repos/WordPress/openverse-catalog
|
closed
|
Re-enable Slack load completion notifications for Finnish
|
🟧 priority: high 🛠 goal: fix 💻 aspect: code
|
## Problem
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
Finnish museums has been turned back on after converting it to a dated DAG. There is a backfill of several months worth of daily DagRuns running, which will spam `report_load_completion` messages in Slack at an abnormal rate.
For the duration of the backfill, we're silencing those load completion messages for Finnish. This issue is to track:
1. manually reviewing the state of the backfill (errors will still be reported)
2. re-enabling the messages when the backfill completes
<!-- If you would like to work on this, please comment below separately. -->
|
1.0
|
Re-enable Slack load completion notifications for Finnish - ## Problem
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
Finnish museums has been turned back on after converting it to a dated DAG. There is a backfill of several months worth of daily DagRuns running, which will spam `report_load_completion` messages in Slack at an abnormal rate.
For the duration of the backfill, we're silencing those load completion messages for Finnish. This issue is to track:
1. manually reviewing the state of the backfill (errors will still be reported)
2. re-enabling the messages when the backfill completes
<!-- If you would like to work on this, please comment below separately. -->
|
non_process
|
re enable slack load completion notifications for finnish problem finnish museums has been turned back on after converting it to a dated dag there is a backfill of several months worth of daily dagruns running which will spam report load completion messages in slack at an abnormal rate for the duration of the backfill we re silencing those load completion messages for finnish this issue is to track manually reviewing the state of the backfill errors will still be reported re enabling the messages when the backfill completes
| 0
|
5,078
| 7,873,796,246
|
IssuesEvent
|
2018-06-25 15:11:46
|
AlexsLemonade/refinebio
|
https://api.github.com/repos/AlexsLemonade/refinebio
|
closed
|
Aggregate and Deliver User-Selected Samples for Download
|
api aws database email processor review sci review smasher
|
### Context
Once a user has a created a set of samples that they wish to download, they need to tell us to prepare a download for them, so we need to pull them off of EFS/S3, aggregate them, move them to back to S3, then deliver the result to them either via the web or via email.
This requires three main parts (in addition to the required front-end work):
### Problem or idea
#### API
Two new API endpoints are needed: one which accepts a `POST` with a list of sample IDs, a method of aggregation (and any other desired preferences), and an email, which returns a Download ID, and other which returns the status (and final URL) for a given Download ID which can be polled while the download is being prepared.
_TODO: Investigate potential for abuse and mitigation strategies._
#### Processing
This should be a queue-driven process which fetches the data from EFS/S3, smashes it as desired, packages it with metadata and moves it to S3. I have no idea how much processing power is required for this or how what the time bounds are for these jobs.
#### Delivery
Assuming the processing step takes more than _n_ seconds, we will need a way to deliver the final asset to the user. This will require constructing an email and ensuring its delivery via AWS SES. We can probably look to Cognoma as a model for this.
### Solution or next step
I'd probably start this by figuring out how we want to format the metadata, and figuring how what types of aggregation we want to offer and how to do them. I have some Pandas code [here](https://github.com/Miserlou/science) that may or may not be useful for this.
### New Issue Checklist
- [x] The title is short and descriptive.
- [x] You have explained the context that led you to write this issue.
- [x] You have reported a problem or idea.
- [x] You have proposed a solution or next step.
|
1.0
|
Aggregate and Deliver User-Selected Samples for Download - ### Context
Once a user has a created a set of samples that they wish to download, they need to tell us to prepare a download for them, so we need to pull them off of EFS/S3, aggregate them, move them to back to S3, then deliver the result to them either via the web or via email.
This requires three main parts (in addition to the required front-end work):
### Problem or idea
#### API
Two new API endpoints are needed: one which accepts a `POST` with a list of sample IDs, a method of aggregation (and any other desired preferences), and an email, which returns a Download ID, and other which returns the status (and final URL) for a given Download ID which can be polled while the download is being prepared.
_TODO: Investigate potential for abuse and mitigation strategies._
#### Processing
This should be a queue-driven process which fetches the data from EFS/S3, smashes it as desired, packages it with metadata and moves it to S3. I have no idea how much processing power is required for this or how what the time bounds are for these jobs.
#### Delivery
Assuming the processing step takes more than _n_ seconds, we will need a way to deliver the final asset to the user. This will require constructing an email and ensuring its delivery via AWS SES. We can probably look to Cognoma as a model for this.
### Solution or next step
I'd probably start this by figuring out how we want to format the metadata, and figuring how what types of aggregation we want to offer and how to do them. I have some Pandas code [here](https://github.com/Miserlou/science) that may or may not be useful for this.
### New Issue Checklist
- [x] The title is short and descriptive.
- [x] You have explained the context that led you to write this issue.
- [x] You have reported a problem or idea.
- [x] You have proposed a solution or next step.
|
process
|
aggregate and deliver user selected samples for download context once a user has a created a set of samples that they wish to download they need to tell us to prepare a download for them so we need to pull them off of efs aggregate them move them to back to then deliver the result to them either via the web or via email this requires three main parts in addition to the required front end work problem or idea api two new api endpoints are needed one which accepts a post with a list of sample ids a method of aggregation and any other desired preferences and an email which returns a download id and other which returns the status and final url for a given download id which can be polled while the download is being prepared todo investigate potential for abuse and mitigation strategies processing this should be a queue driven process which fetches the data from efs smashes it as desired packages it with metadata and moves it to i have no idea how much processing power is required for this or how what the time bounds are for these jobs delivery assuming the processing step takes more than n seconds we will need a way to deliver the final asset to the user this will require constructing an email and ensuring its delivery via aws ses we can probably look to cognoma as a model for this solution or next step i d probably start this by figuring out how we want to format the metadata and figuring how what types of aggregation we want to offer and how to do them i have some pandas code that may or may not be useful for this new issue checklist the title is short and descriptive you have explained the context that led you to write this issue you have reported a problem or idea you have proposed a solution or next step
| 1
|
492,857
| 14,221,665,757
|
IssuesEvent
|
2020-11-17 15:58:05
|
clastix/capsule
|
https://api.github.com/repos/clastix/capsule
|
closed
|
Add the node pool label as additionalPrinterColumns
|
enhancement low-priority
|
# Describe the feature
It would be nice to have the node pool label of a Tenant as Additional Printer Column in the output of `kubectl get tenants`
# What would the new user story look like?
Having a tenant oil with a given node selector
```yaml
apiVersion: capsule.clastix.io/v1alpha1
kind: Tenant
metadata:
name: oil
spec:
nodeSelector:
pool: caas
...
```
you get
```
$ kubectl get tenants
NAME NAMESPACE QUOTA NAMESPACE COUNT OWNER NAME OWNER KIND SELECTOR AGE
oil 10 3 alice User pool=caas 39d
```
# Expected behavior
see above
|
1.0
|
Add the node pool label as additionalPrinterColumns - # Describe the feature
It would be nice to have the node pool label of a Tenant as Additional Printer Column in the output of `kubectl get tenants`
# What would the new user story look like?
Having a tenant oil with a given node selector
```yaml
apiVersion: capsule.clastix.io/v1alpha1
kind: Tenant
metadata:
name: oil
spec:
nodeSelector:
pool: caas
...
```
you get
```
$ kubectl get tenants
NAME NAMESPACE QUOTA NAMESPACE COUNT OWNER NAME OWNER KIND SELECTOR AGE
oil 10 3 alice User pool=caas 39d
```
# Expected behavior
see above
|
non_process
|
add the node pool label as additionalprintercolumns describe the feature it would be nice to have the node pool label of a tenant as additional printer column in the output of kubectl get tenants what would the new user story look like having a tenant oil with a given node selector yaml apiversion capsule clastix io kind tenant metadata name oil spec nodeselector pool caas you get kubectl get tenants name namespace quota namespace count owner name owner kind selector age oil alice user pool caas expected behavior see above
| 0
|
8,430
| 11,596,180,898
|
IssuesEvent
|
2020-02-24 18:26:47
|
prisma/prisma-engines
|
https://api.github.com/repos/prisma/prisma-engines
|
opened
|
Panic for empty env var
|
process/candidate
|
When having this schema (models don't really matter)
```prisma
datasource db {
provider = "sqlite"
url = env("DB_URL")
}
generator client {
provider = "prisma-client-js"
binaryTargets = ["native"]
}
model User {
id String @default(cuid()) @id
email String @unique
name String?
posts Post[]
treeHouseMaterial Tree?
}
model Post {
id String @default(cuid()) @id
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
published Boolean
title String
content String?
author User?
}
enum Tree {
OAK DOUGLAS
}
```
And the env var `DB_URL=""` set to an empty string, then starting the query engine in http mode, I get the following error, printed in stdout:
```
{
timestamp: 'Feb 24 10:22:58.628',
level: 'ERROR',
target: 'prisma',
fields: {
message: 'PANIC',
reason: 'called `Option::unwrap()` on a `None` value',
file: '/rustc/5e1a799842ba6ed4a57e91f7ab9435947482f7d8/src/libcore/macros/mod.rs',
line: 15,
column: 40
}
}
```
It would be great if we can turn this into a known error.
|
1.0
|
Panic for empty env var - When having this schema (models don't really matter)
```prisma
datasource db {
provider = "sqlite"
url = env("DB_URL")
}
generator client {
provider = "prisma-client-js"
binaryTargets = ["native"]
}
model User {
id String @default(cuid()) @id
email String @unique
name String?
posts Post[]
treeHouseMaterial Tree?
}
model Post {
id String @default(cuid()) @id
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
published Boolean
title String
content String?
author User?
}
enum Tree {
OAK DOUGLAS
}
```
And the env var `DB_URL=""` set to an empty string, then starting the query engine in http mode, I get the following error, printed in stdout:
```
{
timestamp: 'Feb 24 10:22:58.628',
level: 'ERROR',
target: 'prisma',
fields: {
message: 'PANIC',
reason: 'called `Option::unwrap()` on a `None` value',
file: '/rustc/5e1a799842ba6ed4a57e91f7ab9435947482f7d8/src/libcore/macros/mod.rs',
line: 15,
column: 40
}
}
```
It would be great if we can turn this into a known error.
|
process
|
panic for empty env var when having this schema models don t really matter prisma datasource db provider sqlite url env db url generator client provider prisma client js binarytargets model user id string default cuid id email string unique name string posts post treehousematerial tree model post id string default cuid id createdat datetime default now updatedat datetime updatedat published boolean title string content string author user enum tree oak douglas and the env var db url set to an empty string then starting the query engine in http mode i get the following error printed in stdout timestamp feb level error target prisma fields message panic reason called option unwrap on a none value file rustc src libcore macros mod rs line column it would be great if we can turn this into a known error
| 1
|
509,848
| 14,750,280,664
|
IssuesEvent
|
2021-01-08 01:43:25
|
GentenStudios/Phoenix
|
https://api.github.com/repos/GentenStudios/Phoenix
|
closed
|
Licences for art and other media
|
enhancement priority-high
|
## Story
Currently not all files have a proper licence attached to them. We may wanna add one.
|
1.0
|
Licences for art and other media - ## Story
Currently not all files have a proper licence attached to them. We may wanna add one.
|
non_process
|
licences for art and other media story currently not all files have a proper licence attached to them we may wanna add one
| 0
|
12,669
| 15,037,635,559
|
IssuesEvent
|
2021-02-02 16:33:16
|
retaildevcrews/ngsa
|
https://api.github.com/repos/retaildevcrews/ngsa
|
closed
|
Engineering Fundamentals Process
|
EngPrac Process
|
## Description
What:
- checklists and pre-work documentation
Why:
- to follow CSE recommended engineering fundamentals
When:
- monthly
Where:
- github, ado, and snowball
## Tasks
- [X] Game Plan Document
- [X] Game Plan Review
- [x] CPRs
- [x] Tech Led Engineering Checklist
- [x] Engineering Surveys
## Acceptance Criteria
- Game Plan review completed & plan uploaded
- CPR pre-work completed and uploaded 24hrs in advance
- Artifacts created and uploaded into ADO
## Constraints
- N/A
|
1.0
|
Engineering Fundamentals Process - ## Description
What:
- checklists and pre-work documentation
Why:
- to follow CSE recommended engineering fundamentals
When:
- monthly
Where:
- github, ado, and snowball
## Tasks
- [X] Game Plan Document
- [X] Game Plan Review
- [x] CPRs
- [x] Tech Led Engineering Checklist
- [x] Engineering Surveys
## Acceptance Criteria
- Game Plan review completed & plan uploaded
- CPR pre-work completed and uploaded 24hrs in advance
- Artifacts created and uploaded into ADO
## Constraints
- N/A
|
process
|
engineering fundamentals process description what checklists and pre work documentation why to follow cse recommended engineering fundamentals when monthly where github ado and snowball tasks game plan document game plan review cprs tech led engineering checklist engineering surveys acceptance criteria game plan review completed plan uploaded cpr pre work completed and uploaded in advance artifacts created and uploaded into ado constraints n a
| 1
|
15,937
| 20,160,881,993
|
IssuesEvent
|
2022-02-09 21:24:08
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
Investigate Unit conversions in PH Flue gas
|
Process Heating
|
All the conversions for the inputs are working correctly (including the displayed heating values in the fuel modals)
but the Available heat changes when you do a unit conversion. I wonder if there is a conversion that is getting missed before sending to the suite? Maybe check Ambient Air temp first (recently added field).
If can't find anything, maybe make a note and move to the backlog/icebox.
original

converted to metric

|
1.0
|
Investigate Unit conversions in PH Flue gas - All the conversions for the inputs are working correctly (including the displayed heating values in the fuel modals)
but the Available heat changes when you do a unit conversion. I wonder if there is a conversion that is getting missed before sending to the suite? Maybe check Ambient Air temp first (recently added field).
If can't find anything, maybe make a note and move to the backlog/icebox.
original

converted to metric

|
process
|
investigate unit conversions in ph flue gas all the conversions for the inputs are working correctly including the displayed heating values in the fuel modals but the available heat changes when you do a unit conversion i wonder if there is a conversion that is getting missed before sending to the suite maybe check ambient air temp first recently added field if can t find anything maybe make a note and move to the backlog icebox original converted to metric
| 1
|
170,080
| 13,172,552,647
|
IssuesEvent
|
2020-08-11 18:38:46
|
Azure/azure-sdk-for-js
|
https://api.github.com/repos/Azure/azure-sdk-for-js
|
closed
|
Storage File Datalake Readme Issue
|
Client Docs Storage Up for grabs test-manual-pass
|
1.
Section [Link](https://github.com/Azure/azure-sdk-for-js/tree/master/sdk/storage/storage-file-datalake#examples)

Suggestion:
Add a summary of examples. As follows:

|
1.0
|
Storage File Datalake Readme Issue - 1.
Section [Link](https://github.com/Azure/azure-sdk-for-js/tree/master/sdk/storage/storage-file-datalake#examples)

Suggestion:
Add a summary of examples. As follows:

|
non_process
|
storage file datalake readme issue section suggestion add a summary of examples as follows
| 0
|
25,309
| 18,459,424,305
|
IssuesEvent
|
2021-10-15 21:35:30
|
bolsote/isoduration
|
https://api.github.com/repos/bolsote/isoduration
|
opened
|
Pin development dependencies
|
infrastructure
|
Right now, we have a fairly relaxed development requirements file. We should really pin all those versions, so we have a consistent environment everywhere. A `pip-tools`-based approach seems appropriate.
|
1.0
|
Pin development dependencies - Right now, we have a fairly relaxed development requirements file. We should really pin all those versions, so we have a consistent environment everywhere. A `pip-tools`-based approach seems appropriate.
|
non_process
|
pin development dependencies right now we have a fairly relaxed development requirements file we should really pin all those versions so we have a consistent environment everywhere a pip tools based approach seems appropriate
| 0
|
22,119
| 30,654,934,146
|
IssuesEvent
|
2023-07-25 11:27:54
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
skypilot-nightly 1.0.0.dev20230725 has 2 GuardDog issues
|
guarddog exec-base64 silent-process-execution
|
https://pypi.org/project/skypilot-nightly
https://inspector.pypi.io/project/skypilot-nightly
```{
"dependency": "skypilot-nightly",
"version": "1.0.0.dev20230725",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "skypilot-nightly-1.0.0.dev20230725/sky/skylet/log_lib.py:219",
"code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
],
"exec-base64": [
{
"location": "skypilot-nightly-1.0.0.dev20230725/sky/cloud_stores.py:113",
"code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
]
},
"path": "/tmp/tmp_8rxbnqo/skypilot-nightly"
}
}```
|
1.0
|
skypilot-nightly 1.0.0.dev20230725 has 2 GuardDog issues - https://pypi.org/project/skypilot-nightly
https://inspector.pypi.io/project/skypilot-nightly
```{
"dependency": "skypilot-nightly",
"version": "1.0.0.dev20230725",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "skypilot-nightly-1.0.0.dev20230725/sky/skylet/log_lib.py:219",
"code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
],
"exec-base64": [
{
"location": "skypilot-nightly-1.0.0.dev20230725/sky/cloud_stores.py:113",
"code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
]
},
"path": "/tmp/tmp_8rxbnqo/skypilot-nightly"
}
}```
|
process
|
skypilot nightly has guarddog issues dependency skypilot nightly version result issues errors results silent process execution location skypilot nightly sky skylet log lib py code subprocess popen n daemon cmd n start new session true n suppress output n stdout subprocess devnull n stderr subprocess devnull n disa message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null exec location skypilot nightly sky cloud stores py code p subprocess run command n stdout subprocess pipe n shell true n check true n executable bin bash message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n path tmp tmp skypilot nightly
| 1
|
146,902
| 5,630,374,239
|
IssuesEvent
|
2017-04-05 12:06:05
|
research-resource/research_resource
|
https://api.github.com/repos/research-resource/research_resource
|
closed
|
Add link to log in on sign up form
|
priority-2 T1h
|
Need to add a link underneath / next to (see UI) the sign up module to say "Already signed up? Log in" with the "Log in" linking to the log in page.
|
1.0
|
Add link to log in on sign up form - Need to add a link underneath / next to (see UI) the sign up module to say "Already signed up? Log in" with the "Log in" linking to the log in page.
|
non_process
|
add link to log in on sign up form need to add a link underneath next to see ui the sign up module to say already signed up log in with the log in linking to the log in page
| 0
|
6,110
| 8,969,444,908
|
IssuesEvent
|
2019-01-29 10:49:27
|
pwittchen/prefser
|
https://api.github.com/repos/pwittchen/prefser
|
closed
|
Release 2.2.2-rx2
|
release process
|
**Release notes**:
- updated project dependencies - 74de9db550ff941083448c28c48b6daffaed7b6c
- migrated to `androidx` - 74de9db550ff941083448c28c48b6daffaed7b6c, fixes #130
**Things to do**:
- [x] bump library version
- [x] upload Archives to Maven Central Repository
- [x] close and release artifact on Nexus
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md` after Maven Sync
- [x] create new GitHub release
|
1.0
|
Release 2.2.2-rx2 - **Release notes**:
- updated project dependencies - 74de9db550ff941083448c28c48b6daffaed7b6c
- migrated to `androidx` - 74de9db550ff941083448c28c48b6daffaed7b6c, fixes #130
**Things to do**:
- [x] bump library version
- [x] upload Archives to Maven Central Repository
- [x] close and release artifact on Nexus
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md` after Maven Sync
- [x] create new GitHub release
|
process
|
release release notes updated project dependencies migrated to androidx fixes things to do bump library version upload archives to maven central repository close and release artifact on nexus update changelog md after maven sync bump library version in readme md after maven sync create new github release
| 1
|
63,882
| 18,041,999,766
|
IssuesEvent
|
2021-09-18 07:35:15
|
mozilla-lockwise/lockwise-android
|
https://api.github.com/repos/mozilla-lockwise/lockwise-android
|
opened
|
Doesn't open when two Firefox are installed
|
defect
|
## Steps to reproduce
Install Firefox Beta and Stable in the same device.
Log in each with a different Firefox accounts.
Use fingerprint to unlock Lovkwise (it might not be necessary).
### Expected behavior
Allow choose account to use with Lockwise or show at least the passwords of one account (probably Stable would have sense).
### Actual behavior
After authentication with finger print, the vault stays lock and keep asking for fingerprint.
### Device & build information
* Device: Motorola "moto g⁸ power (XT2041-3)"
* Build version: Android 11, 1st August 2021, kernel 4.14.180-perf+, build RPES31.Q4U-47-35-8
### Notes
Attachments:
|
1.0
|
Doesn't open when two Firefox are installed - ## Steps to reproduce
Install Firefox Beta and Stable in the same device.
Log in each with a different Firefox accounts.
Use fingerprint to unlock Lovkwise (it might not be necessary).
### Expected behavior
Allow choose account to use with Lockwise or show at least the passwords of one account (probably Stable would have sense).
### Actual behavior
After authentication with finger print, the vault stays lock and keep asking for fingerprint.
### Device & build information
* Device: Motorola "moto g⁸ power (XT2041-3)"
* Build version: Android 11, 1st August 2021, kernel 4.14.180-perf+, build RPES31.Q4U-47-35-8
### Notes
Attachments:
|
non_process
|
doesn t open when two firefox are installed steps to reproduce install firefox beta and stable in the same device log in each with a different firefox accounts use fingerprint to unlock lovkwise it might not be necessary expected behavior allow choose account to use with lockwise or show at least the passwords of one account probably stable would have sense actual behavior after authentication with finger print the vault stays lock and keep asking for fingerprint device build information device motorola moto g⁸ power build version android august kernel perf build notes attachments
| 0
|
195,883
| 22,362,799,005
|
IssuesEvent
|
2022-06-15 22:38:23
|
snowflakedb/snowflake-hive-metastore-connector
|
https://api.github.com/repos/snowflakedb/snowflake-hive-metastore-connector
|
closed
|
CVE-2020-24750 (High) detected in jackson-databind-2.6.5.jar
|
security vulnerability
|
## CVE-2020-24750 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.5/jackson-databind-2.6.5.jar</p>
<p>
Dependency Hierarchy:
- hive-metastore-2.3.5.jar (Root Library)
- hive-serde-2.3.5.jar
- hive-common-2.3.5.jar
- :x: **jackson-databind-2.6.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowflakedb/snowflake-hive-metastore-connector/commit/37f5b0ac91898ef82cc1bf4610b729970f6eed58">37f5b0ac91898ef82cc1bf4610b729970f6eed58</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration.
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-09-17</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.6.7.4</p>
<p>Direct dependency fix Resolution (org.apache.hive:hive-metastore): 2.3.6</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.hive","packageName":"hive-metastore","packageVersion":"2.3.5","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.hive:hive-metastore:2.3.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.6","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24750","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-24750 (High) detected in jackson-databind-2.6.5.jar - ## CVE-2020-24750 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.5/jackson-databind-2.6.5.jar</p>
<p>
Dependency Hierarchy:
- hive-metastore-2.3.5.jar (Root Library)
- hive-serde-2.3.5.jar
- hive-common-2.3.5.jar
- :x: **jackson-databind-2.6.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowflakedb/snowflake-hive-metastore-connector/commit/37f5b0ac91898ef82cc1bf4610b729970f6eed58">37f5b0ac91898ef82cc1bf4610b729970f6eed58</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration.
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-09-17</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.6.7.4</p>
<p>Direct dependency fix Resolution (org.apache.hive:hive-metastore): 2.3.6</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.hive","packageName":"hive-metastore","packageVersion":"2.3.5","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.hive:hive-metastore:2.3.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.6","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24750","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy hive metastore jar root library hive serde jar hive common jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com pastdev httpcomponents configuration jndiconfiguration publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution org apache hive hive metastore rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache hive hive metastore isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com pastdev httpcomponents configuration jndiconfiguration vulnerabilityurl
| 0
|
220,034
| 24,549,694,556
|
IssuesEvent
|
2022-10-12 11:37:21
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[Security Solution]: Text not aligned with node icons under analyzer.
|
bug impact:low Team:Threat Hunting Team: SecuritySolution Team:Threat Hunting:Investigations v8.5.0
|
**Describe the bug:**
Text not aligned with node icon under analyzer.
**Build Details:**
```
VERSION: 8.5.0-SNAPSHOT
BUILD: 55925
COMMIT: dc43193d73c5869335a239c7012528bb1fffd509
```
**Preconditions**
1. Kibana should be running.
**Steps to Reproduce**
1. Navigate to Alerts tab.
2. Now, select an alert where analyzer event is available.
3. Open the analyzer and click on node legend icon.
4. Observe that text not aligned with node icons under analyzer.
**Actual Result**
Text not aligned with node icons under analyzer.
**Expected Result**
Text should be aligned with node icons under analyzer.
**Screenshot:**

|
True
|
[Security Solution]: Text not aligned with node icons under analyzer. - **Describe the bug:**
Text not aligned with node icon under analyzer.
**Build Details:**
```
VERSION: 8.5.0-SNAPSHOT
BUILD: 55925
COMMIT: dc43193d73c5869335a239c7012528bb1fffd509
```
**Preconditions**
1. Kibana should be running.
**Steps to Reproduce**
1. Navigate to Alerts tab.
2. Now, select an alert where analyzer event is available.
3. Open the analyzer and click on node legend icon.
4. Observe that text not aligned with node icons under analyzer.
**Actual Result**
Text not aligned with node icons under analyzer.
**Expected Result**
Text should be aligned with node icons under analyzer.
**Screenshot:**

|
non_process
|
text not aligned with node icons under analyzer describe the bug text not aligned with node icon under analyzer build details version snapshot build commit preconditions kibana should be running steps to reproduce navigate to alerts tab now select an alert where analyzer event is available open the analyzer and click on node legend icon observe that text not aligned with node icons under analyzer actual result text not aligned with node icons under analyzer expected result text should be aligned with node icons under analyzer screenshot
| 0
|
17,529
| 23,341,099,823
|
IssuesEvent
|
2022-08-09 14:05:46
|
VolpeUSDOT/FHWA-DANATool
|
https://api.github.com/repos/VolpeUSDOT/FHWA-DANATool
|
closed
|
Geographic intersection for tier 1 matching is inaccurate
|
bug Process Improvements Medium Priority For Next Version
|
There is a deepish dive we could do for the Tier 1 matches to make sure it is working as expected and as desired. It may be matching multiple links at the same time in some cases, and there may be other things that could be improved.
The geographic matching of TMAS count stations to NPMRDS TMCs is inaccurate because it tries to use .2 degrees, which is a different profile in different parts of the globe. It is distorted the further you get from the equator. So we should use a different CRS to convert to some number of miles instead, so we get a more accurate matching function.
|
1.0
|
Geographic intersection for tier 1 matching is inaccurate - There is a deepish dive we could do for the Tier 1 matches to make sure it is working as expected and as desired. It may be matching multiple links at the same time in some cases, and there may be other things that could be improved.
The geographic matching of TMAS count stations to NPMRDS TMCs is inaccurate because it tries to use .2 degrees, which is a different profile in different parts of the globe. It is distorted the further you get from the equator. So we should use a different CRS to convert to some number of miles instead, so we get a more accurate matching function.
|
process
|
geographic intersection for tier matching is inaccurate there is a deepish dive we could do for the tier matches to make sure it is working as expected and as desired it may be matching multiple links at the same time in some cases and there may be other things that could be improved the geographic matching of tmas count stations to npmrds tmcs is inaccurate because it tries to use degrees which is a different profile in different parts of the globe it is distorted the further you get from the equator so we should use a different crs to convert to some number of miles instead so we get a more accurate matching function
| 1
|
623,703
| 19,676,611,151
|
IssuesEvent
|
2022-01-11 13:01:42
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
Same token should not be listed for both Swap from and to
|
bug priority/P3 QA/Yes release-notes/exclude feature/wallet OS/Android
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!-- Provide a brief description of the issue -->
Same token should not be listed for both Swap from and to
## Steps to reproduce <!-- Please add a series of steps to reproduce the issue -->
1. Create wallet
2. Select Swap
3. Able to set same token for both swap from and to
## Actual result <!-- Please add screenshots if needed -->
https://user-images.githubusercontent.com/17010094/145383433-4a0981b3-2a92-43d8-8077-bf7faa33de35.mp4
## Expected result
Similar to Desktop and iOS, token that is set in Swap from should not be listed in Swap to list
## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] -->
Easy
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current Play Store version? Yes when released
- Can you reproduce this issue with the current Play Store Beta version? Yes
- Can you reproduce this issue with the current Play Store Nightly version? Yes
## Device details
- Install type (ARM, x86): ARM
- Device type (Phone, Tablet, Phablet): All
- Android version: All
## Brave version
1.33.103
### Website problems only
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
### Additional information
<!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue -->
cc: @SergeyZhukovsky
|
1.0
|
Same token should not be listed for both Swap from and to - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!-- Provide a brief description of the issue -->
Same token should not be listed for both Swap from and to
## Steps to reproduce <!-- Please add a series of steps to reproduce the issue -->
1. Create wallet
2. Select Swap
3. Able to set same token for both swap from and to
## Actual result <!-- Please add screenshots if needed -->
https://user-images.githubusercontent.com/17010094/145383433-4a0981b3-2a92-43d8-8077-bf7faa33de35.mp4
## Expected result
Similar to Desktop and iOS, token that is set in Swap from should not be listed in Swap to list
## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] -->
Easy
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current Play Store version? Yes when released
- Can you reproduce this issue with the current Play Store Beta version? Yes
- Can you reproduce this issue with the current Play Store Nightly version? Yes
## Device details
- Install type (ARM, x86): ARM
- Device type (Phone, Tablet, Phablet): All
- Android version: All
## Brave version
1.33.103
### Website problems only
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
### Additional information
<!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue -->
cc: @SergeyZhukovsky
|
non_process
|
same token should not be listed for both swap from and to have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description same token should not be listed for both swap from and to steps to reproduce create wallet select swap able to set same token for both swap from and to actual result expected result similar to desktop and ios token that is set in swap from should not be listed in swap to list issue reproduces how often easy version channel information can you reproduce this issue with the current play store version yes when released can you reproduce this issue with the current play store beta version yes can you reproduce this issue with the current play store nightly version yes device details install type arm arm device type phone tablet phablet all android version all brave version website problems only does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na additional information cc sergeyzhukovsky
| 0
|
18,945
| 24,904,757,604
|
IssuesEvent
|
2022-10-29 05:01:51
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
MIPS EABI32 support
|
Feature: Processor/MIPS
|
**Is your feature request related to a problem? Please describe.**
Ghidra does not support the EABI32 ABI for MIPS.
MIPS has supported a fairly large variety of ABIs over the years [[1]](https://en.wikipedia.org/wiki/MIPS_architecture#Calling_conventions). Ghidra should not necessarily be expected to support every ABI out there. Having said that EABI stands out because it is the only MIPS ABI currently supported by GCC that Ghidra does not support. [[2]](https://sourceware.org/gdb/current/onlinedocs/gdb/MIPS.html) Additionally, the commercial disassembler has support for EABI32.
**Describe the solution you'd like**
I think that an update to the processor spec with a new ABI would solve the issue. That looks like a new cspec file (and other surrounding code).
**Describe alternatives you've considered**
I initially wrote up an example that allowed an "__eabicall" calling convention. This isn't a good solution because it's not a calling convention, but an ABI.
**Additional context**
I set up a [repo](https://github.com/lacraig2/Ghidra_MIPS_EABI32_Example) that has source and an example binary that demonstrate the issue. It contains:
- Minimal source code that shows the ABI issue
- A makefile that works with `mips-linux-gnu-gcc` (default from ubuntu apt)
- An example binary
As another example see the below function. It should have five arguments. Instead it shows 4 and one argument "in t0", which should be the 5th argument.
<img width="509" alt="Screen Shot 2021-11-15 at 4 26 41 PM" src="https://user-images.githubusercontent.com/7518336/141856096-839495ea-6f00-4a34-a003-011a05e01493.png">
There is, unfortunately, not a lot of documentation on the EABI32 format. You can get context from either the GCC source or this one [mailing list post](https://sourceware.org/legacy-ml/binutils/2003-06/msg00436.html).
*UPDATE:* I have a PR for review.
|
1.0
|
MIPS EABI32 support - **Is your feature request related to a problem? Please describe.**
Ghidra does not support the EABI32 ABI for MIPS.
MIPS has supported a fairly large variety of ABIs over the years [[1]](https://en.wikipedia.org/wiki/MIPS_architecture#Calling_conventions). Ghidra should not necessarily be expected to support every ABI out there. Having said that EABI stands out because it is the only MIPS ABI currently supported by GCC that Ghidra does not support. [[2]](https://sourceware.org/gdb/current/onlinedocs/gdb/MIPS.html) Additionally, the commercial disassembler has support for EABI32.
**Describe the solution you'd like**
I think that an update to the processor spec with a new ABI would solve the issue. That looks like a new cspec file (and other surrounding code).
**Describe alternatives you've considered**
I initially wrote up an example that allowed an "__eabicall" calling convention. This isn't a good solution because it's not a calling convention, but an ABI.
**Additional context**
I set up a [repo](https://github.com/lacraig2/Ghidra_MIPS_EABI32_Example) that has source and an example binary that demonstrate the issue. It contains:
- Minimal source code that shows the ABI issue
- A makefile that works with `mips-linux-gnu-gcc` (default from ubuntu apt)
- An example binary
As another example see the below function. It should have five arguments. Instead it shows 4 and one argument "in t0", which should be the 5th argument.
<img width="509" alt="Screen Shot 2021-11-15 at 4 26 41 PM" src="https://user-images.githubusercontent.com/7518336/141856096-839495ea-6f00-4a34-a003-011a05e01493.png">
There is, unfortunately, not a lot of documentation on the EABI32 format. You can get context from either the GCC source or this one [mailing list post](https://sourceware.org/legacy-ml/binutils/2003-06/msg00436.html).
*UPDATE:* I have a PR for review.
|
process
|
mips support is your feature request related to a problem please describe ghidra does not support the abi for mips mips has supported a fairly large variety of abis over the years ghidra should not necessarily be expected to support every abi out there having said that eabi stands out because it is the only mips abi currently supported by gcc that ghidra does not support additionally the commercial disassembler has support for describe the solution you d like i think that an update to the processor spec with a new abi would solve the issue that looks like a new cspec file and other surrounding code describe alternatives you ve considered i initially wrote up an example that allowed an eabicall calling convention this isn t a good solution because it s not a calling convention but an abi additional context i set up a that has source and an example binary that demonstrate the issue it contains minimal source code that shows the abi issue a makefile that works with mips linux gnu gcc default from ubuntu apt an example binary as another example see the below function it should have five arguments instead it shows and one argument in which should be the argument img width alt screen shot at pm src there is unfortunately not a lot of documentation on the format you can get context from either the gcc source or this one update i have a pr for review
| 1
|
57,331
| 15,730,654,969
|
IssuesEvent
|
2021-03-29 16:09:27
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
Hung task related to vdev_trim_calculate_progress()
|
Status: Triage Needed Type: Defect
|
### System information
Distribution Name | NixOS
Distribution Version | 20.03
Linux Kernel | 5.4.39
Architecture | x86_64
ZFS Version | 0.8.3-1
SPL Version | 0.8.3-1
### Describe the problem you're observing
This happened to me twice in last 12 months. I'm using SSDs that might handle trim requests poorly, but that's speculation. Model: ADATA SP900.
"zpool import -fFn <pool>" gets me:
```
[178.250374] VERIFY3(0 == metaslab_load(msp)) failed (0 == 52)
[178.250418] PANIC at vdev_trim.c:612:vdev_trim_calculate_progress()
```
The pool *can* be imported with "-o readonly=on". That's how I recovered the last time it happened, about ~8 months ago.
Afterwards, I'm getting:
```
INFO: task txg_sync:1992 blocked for more than <N> seconds.
```
### Describe how to reproduce the problem
No clue. Both times it happened while waking up from hibernation, after a scrub was completed.
I noticed that now autotrim is enabled on this pool again, even though I disabled it the last time this kind of thing happened. I suspect this might've happened during an OS update. It might be that a trim was scheduled on the same day the scrub was.
|
1.0
|
Hung task related to vdev_trim_calculate_progress() - ### System information
Distribution Name | NixOS
Distribution Version | 20.03
Linux Kernel | 5.4.39
Architecture | x86_64
ZFS Version | 0.8.3-1
SPL Version | 0.8.3-1
### Describe the problem you're observing
This happened to me twice in last 12 months. I'm using SSDs that might handle trim requests poorly, but that's speculation. Model: ADATA SP900.
"zpool import -fFn <pool>" gets me:
```
[178.250374] VERIFY3(0 == metaslab_load(msp)) failed (0 == 52)
[178.250418] PANIC at vdev_trim.c:612:vdev_trim_calculate_progress()
```
The pool *can* be imported with "-o readonly=on". That's how I recovered the last time it happened, about ~8 months ago.
Afterwards, I'm getting:
```
INFO: task txg_sync:1992 blocked for more than <N> seconds.
```
### Describe how to reproduce the problem
No clue. Both times it happened while waking up from hibernation, after a scrub was completed.
I noticed that now autotrim is enabled on this pool again, even though I disabled it the last time this kind of thing happened. I suspect this might've happened during an OS update. It might be that a trim was scheduled on the same day the scrub was.
|
non_process
|
hung task related to vdev trim calculate progress system information distribution name nixos distribution version linux kernel architecture zfs version spl version describe the problem you re observing this happened to me twice in last months i m using ssds that might handle trim requests poorly but that s speculation model adata zpool import ffn gets me metaslab load msp failed panic at vdev trim c vdev trim calculate progress the pool can be imported with o readonly on that s how i recovered the last time it happened about months ago afterwards i m getting info task txg sync blocked for more than seconds describe how to reproduce the problem no clue both times it happened while waking up from hibernation after a scrub was completed i noticed that now autotrim is enabled on this pool again even though i disabled it the last time this kind of thing happened i suspect this might ve happened during an os update it might be that a trim was scheduled on the same day the scrub was
| 0
|
21,697
| 30,194,945,692
|
IssuesEvent
|
2023-07-04 19:40:08
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] Add shorter display name like `≠` to display info for filter operators and change existing `:display-name` to `:long-display-name`
|
.Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
In places like joins we actually show just the operator in the UI rather than a nice long display name for filter clause, e.g.


Note the fancy unicode characters like `≠`. We need to return these in the **display info** for `FilterOperator` maps so the FE can use them when we port stuff to MLv2.
MLv1 definitions live here
https://github.com/metabase/metabase/blob/287c5204d30dd36d1c61656da792ae1f6dc7d750/frontend/src/metabase/query_builder/components/notebook/steps/JoinStep/JoinStep.jsx#L78
I believe these are currently only used for joins, so we technically only need to add them for the 6 operators we return for joins, but we might as well return them for everything; I think for everything besides `:!=`, `:>=`, and `:<=`, it's the same as the tag itself.
What should we call this key? I was going to suggest `:short-display-name`, but it's a bit silly to have the display info for filter operators have `:short-display-name` and `:display-name` while everything else has `:display-name` and `:long-display-name`... maybe what we should do here is change the existing `:display-name` to `:long-display-name` and make `≠` the new `:display-name`.
|
1.0
|
[MLv2] Add shorter display name like `≠` to display info for filter operators and change existing `:display-name` to `:long-display-name` - In places like joins we actually show just the operator in the UI rather than a nice long display name for filter clause, e.g.


Note the fancy unicode characters like `≠`. We need to return these in the **display info** for `FilterOperator` maps so the FE can use them when we port stuff to MLv2.
MLv1 definitions live here
https://github.com/metabase/metabase/blob/287c5204d30dd36d1c61656da792ae1f6dc7d750/frontend/src/metabase/query_builder/components/notebook/steps/JoinStep/JoinStep.jsx#L78
I believe these are currently only used for joins, so we technically only need to add them for the 6 operators we return for joins, but we might as well return them for everything; I think for everything besides `:!=`, `:>=`, and `:<=`, it's the same as the tag itself.
What should we call this key? I was going to suggest `:short-display-name`, but it's a bit silly to have the display info for filter operators have `:short-display-name` and `:display-name` while everything else has `:display-name` and `:long-display-name`... maybe what we should do here is change the existing `:display-name` to `:long-display-name` and make `≠` the new `:display-name`.
|
process
|
add shorter display name like ≠ to display info for filter operators and change existing display name to long display name in places like joins we actually show just the operator in the ui rather than a nice long display name for filter clause e g note the fancy unicode characters like ≠ we need to return these in the display info for filteroperator maps so the fe can use them when we port stuff to definitions live here i believe these are currently only used for joins so we technically only need to add them for the operators we return for joins but we might as well return them for everything i think for everything besides and it s the same as the tag itself what should we call this key i was going to suggest short display name but it s a bit silly to have the display info for filter operators have short display name and display name while everything else has display name and long display name maybe what we should do here is change the existing display name to long display name and make ≠ the new display name
| 1
|
19,394
| 25,537,736,353
|
IssuesEvent
|
2022-11-29 13:15:35
|
eosnetworkfoundation/devrel
|
https://api.github.com/repos/eosnetworkfoundation/devrel
|
closed
|
Create a staging URL to review current state of docs before publishing
|
Process
|
need to run every day
need a process doc
Note: It's crucial to make sure that `latest` still links to the latest version published on the site, not to the potential new version, if any, which will be tested using the calvary link. @ericpassmore
|
1.0
|
Create a staging URL to review current state of docs before publishing - need to run every day
need a process doc
Note: It's crucial to make sure that `latest` still links to the latest version published on the site, not to the potential new version, if any, which will be tested using the calvary link. @ericpassmore
|
process
|
create a staging url to review current state of docs before publishing need to run every day need a process doc note it s crucial to make sure that latest still links to the latest version published on the site not to the potential new version if any which will be tested using the calvary link ericpassmore
| 1
|
75,644
| 14,501,526,300
|
IssuesEvent
|
2020-12-11 19:38:32
|
dragonfruitnetwork/dragon6-api
|
https://api.github.com/repos/dragonfruitnetwork/dragon6-api
|
closed
|
reverse the renaming change
|
area:code-quality area:requests type:cleanup
|
Decided to create a new repo for the modern stats as they're so different that they should feel as an optional extension.
This should mean that the naming change (#181) should be partially reverted, #185 cherry-picked into a new repo and the dragonfruit.common.data library updated
|
1.0
|
reverse the renaming change - Decided to create a new repo for the modern stats as they're so different that they should feel as an optional extension.
This should mean that the naming change (#181) should be partially reverted, #185 cherry-picked into a new repo and the dragonfruit.common.data library updated
|
non_process
|
reverse the renaming change decided to create a new repo for the modern stats as they re so different that they should feel as an optional extension this should mean that the naming change should be partially reverted cherry picked into a new repo and the dragonfruit common data library updated
| 0
|
18,519
| 24,551,777,105
|
IssuesEvent
|
2022-10-12 13:08:53
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] [Offline indicator] There should be a consistency in offline error messages in both enrollment flow and review consent flow
|
Bug P2 iOS Process: Fixed Process: Tested dev
|
There should be a consistency in offline error messages in both enrollment flow and review consent flow
**Enrollment flow:**

**Review updated consent:**

|
2.0
|
[iOS] [Offline indicator] There should be a consistency in offline error messages in both enrollment flow and review consent flow - There should be a consistency in offline error messages in both enrollment flow and review consent flow
**Enrollment flow:**

**Review updated consent:**

|
process
|
there should be a consistency in offline error messages in both enrollment flow and review consent flow there should be a consistency in offline error messages in both enrollment flow and review consent flow enrollment flow review updated consent
| 1
|
22,390
| 31,142,286,174
|
IssuesEvent
|
2023-08-16 01:44:17
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Flaky test: Too long with no output (exceeded 10m0s): context deadline exceeded
|
OS: linux process: flaky test topic: flake ❄️ stage: flake stale
|
### Link to dashboard or CircleCI failure
- https://app.circleci.com/pipelines/github/cypress-io/cypress/41757/workflows/edf97bff-eef3-4f3f-8057-4277aef7d046/jobs/1730766
- https://app.circleci.com/pipelines/github/cypress-io/cypress/42240/workflows/b213eedd-524b-4ef3-9170-b15bfc01d1c7/jobs/1753458
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/navigation.cy.js#L495
### Analysis
Tests hang during execution of `navigation.cy.js` file
<img width="700" alt="Screen Shot 2022-08-11 at 8 00 17 PM" src="https://user-images.githubusercontent.com/26726429/184276927-80c83a8d-e2f2-4b69-9794-a18255e7060d.png">
We also see this same error as seen in https://github.com/cypress-io/cypress/issues/23326
```
$ node ../../scripts/cypress run --config experimentalSessionAndOrigin=true,experimentalModifyObstructiveThirdPartyCode=true --record --parallel --group 5x-driver-chrome:beta-experimentalSessionAndOrigin --browser chrome:beta
libva error: vaGetDriverNameByIndex() failed with unknown libva error, driver_name = (null)
[2032:0811/220636.124358:ERROR:gpu_memory_buffer_support_x11.cc(44)] dri3 extension not supported.
```
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
|
1.0
|
Flaky test: Too long with no output (exceeded 10m0s): context deadline exceeded - ### Link to dashboard or CircleCI failure
- https://app.circleci.com/pipelines/github/cypress-io/cypress/41757/workflows/edf97bff-eef3-4f3f-8057-4277aef7d046/jobs/1730766
- https://app.circleci.com/pipelines/github/cypress-io/cypress/42240/workflows/b213eedd-524b-4ef3-9170-b15bfc01d1c7/jobs/1753458
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/navigation.cy.js#L495
### Analysis
Tests hang during execution of `navigation.cy.js` file
<img width="700" alt="Screen Shot 2022-08-11 at 8 00 17 PM" src="https://user-images.githubusercontent.com/26726429/184276927-80c83a8d-e2f2-4b69-9794-a18255e7060d.png">
We also see this same error as seen in https://github.com/cypress-io/cypress/issues/23326
```
$ node ../../scripts/cypress run --config experimentalSessionAndOrigin=true,experimentalModifyObstructiveThirdPartyCode=true --record --parallel --group 5x-driver-chrome:beta-experimentalSessionAndOrigin --browser chrome:beta
libva error: vaGetDriverNameByIndex() failed with unknown libva error, driver_name = (null)
[2032:0811/220636.124358:ERROR:gpu_memory_buffer_support_x11.cc(44)] dri3 extension not supported.
```
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
|
process
|
flaky test too long with no output exceeded context deadline exceeded link to dashboard or circleci failure link to failing test in github analysis tests hang during execution of navigation cy js file img width alt screen shot at pm src we also see this same error as seen in node scripts cypress run config experimentalsessionandorigin true experimentalmodifyobstructivethirdpartycode true record parallel group driver chrome beta experimentalsessionandorigin browser chrome beta libva error vagetdrivernamebyindex failed with unknown libva error driver name null extension not supported cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
| 1
|
543,376
| 15,880,879,724
|
IssuesEvent
|
2021-04-09 14:11:49
|
ansible/awx
|
https://api.github.com/repos/ansible/awx
|
closed
|
When using AWX on plain HTTP, Social Auth is failing
|
component:api priority:low state:needs_devel type:bug
|
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: -->
- API
##### SUMMARY
When user AWX without HTTPS (plain HTTP), logging in with the GitHub OAuth button fails.
##### ENVIRONMENT
* AWX version: 1.0.1.229
* AWX install method: docker on linux
* Ansible version: 2.4.1.0
* Operating System: debian stretch
* Web Browser: firefox
##### STEPS TO REPRODUCE
* Setup AWX, access through port 80 (default)
* Setup GitHub Org auth (but really, it's the same with any Social Auth plugin)
* Log in using the social auth
##### EXPECTED RESULTS
You're logged-in
##### ACTUAL RESULTS
You're not.
awx_web has a log saying :
```
2017-11-28 17:57:48,299 ERROR social Session value state missing.
```
on the `GET /sso/complete/github-org/` call.
##### ADDITIONAL INFORMATION
The session to persist the `state` parameter between the initial and the `complete` view of the oauth dance is defined by:
https://github.com/ansible/awx/blob/9ed2a0da8f4883a7d6311de2958a01e70605b062/awx/settings/defaults.py#L492
Which uses the default Django session store, which uses a secure cookie due to
https://github.com/ansible/awx/blob/9ed2a0da8f4883a7d6311de2958a01e70605b062/awx/settings/defaults.py#L188
So, if we're not using HTTPS, the cookie will not be sent on the second call, so the session will not be retrieved, the `state` value is absent and the oauth dance fails.
##### WORK AROUND
I guess either use HTTPS or override the SESSION_COOKIE_SECURE to False.
|
1.0
|
When using AWX on plain HTTP, Social Auth is failing - ##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: -->
- API
##### SUMMARY
When user AWX without HTTPS (plain HTTP), logging in with the GitHub OAuth button fails.
##### ENVIRONMENT
* AWX version: 1.0.1.229
* AWX install method: docker on linux
* Ansible version: 2.4.1.0
* Operating System: debian stretch
* Web Browser: firefox
##### STEPS TO REPRODUCE
* Setup AWX, access through port 80 (default)
* Setup GitHub Org auth (but really, it's the same with any Social Auth plugin)
* Log in using the social auth
##### EXPECTED RESULTS
You're logged-in
##### ACTUAL RESULTS
You're not.
awx_web has a log saying :
```
2017-11-28 17:57:48,299 ERROR social Session value state missing.
```
on the `GET /sso/complete/github-org/` call.
##### ADDITIONAL INFORMATION
The session to persist the `state` parameter between the initial and the `complete` view of the oauth dance is defined by:
https://github.com/ansible/awx/blob/9ed2a0da8f4883a7d6311de2958a01e70605b062/awx/settings/defaults.py#L492
Which uses the default Django session store, which uses a secure cookie due to
https://github.com/ansible/awx/blob/9ed2a0da8f4883a7d6311de2958a01e70605b062/awx/settings/defaults.py#L188
So, if we're not using HTTPS, the cookie will not be sent on the second call, so the session will not be retrieved, the `state` value is absent and the oauth dance fails.
##### WORK AROUND
I guess either use HTTPS or override the SESSION_COOKIE_SECURE to False.
|
non_process
|
when using awx on plain http social auth is failing issue type bug report component name api summary when user awx without https plain http logging in with the github oauth button fails environment awx version awx install method docker on linux ansible version operating system debian stretch web browser firefox steps to reproduce setup awx access through port default setup github org auth but really it s the same with any social auth plugin log in using the social auth expected results you re logged in actual results you re not awx web has a log saying error social session value state missing on the get sso complete github org call additional information the session to persist the state parameter between the initial and the complete view of the oauth dance is defined by which uses the default django session store which uses a secure cookie due to so if we re not using https the cookie will not be sent on the second call so the session will not be retrieved the state value is absent and the oauth dance fails work around i guess either use https or override the session cookie secure to false
| 0
|
224,547
| 17,192,807,468
|
IssuesEvent
|
2021-07-16 13:25:20
|
MICA-MNI/BrainSpace
|
https://api.github.com/repos/MICA-MNI/BrainSpace
|
closed
|
Subject list for group connectivity matrices
|
documentation
|
I was wondering if the list of HCP subjects used for the group functional connectivity matrices with `load_group_fc`/`load_group_mpc` was available. My understanding/assumption is that these subsets of HCP data are based on the selection procedures described in [Vos de Wael et al (2018)](https://www.pnas.org/content/115/40/10154), but I'm not 100% certain in that.
I'm looking to compute similar connectivity matrices on the samples used in Vos de Wael et al (2018), but I also think it would be great if the documentation could clarify the source of the group connectivity matrices.
|
1.0
|
Subject list for group connectivity matrices - I was wondering if the list of HCP subjects used for the group functional connectivity matrices with `load_group_fc`/`load_group_mpc` was available. My understanding/assumption is that these subsets of HCP data are based on the selection procedures described in [Vos de Wael et al (2018)](https://www.pnas.org/content/115/40/10154), but I'm not 100% certain in that.
I'm looking to compute similar connectivity matrices on the samples used in Vos de Wael et al (2018), but I also think it would be great if the documentation could clarify the source of the group connectivity matrices.
|
non_process
|
subject list for group connectivity matrices i was wondering if the list of hcp subjects used for the group functional connectivity matrices with load group fc load group mpc was available my understanding assumption is that these subsets of hcp data are based on the selection procedures described in but i m not certain in that i m looking to compute similar connectivity matrices on the samples used in vos de wael et al but i also think it would be great if the documentation could clarify the source of the group connectivity matrices
| 0
|
319,674
| 23,785,043,332
|
IssuesEvent
|
2022-09-02 09:17:50
|
Nicmanthecodeman/CMPG-323-Overview-25102737
|
https://api.github.com/repos/Nicmanthecodeman/CMPG-323-Overview-25102737
|
opened
|
Setup repository
|
documentation good first issue project high
|
# Requirements
- [ ] GitHub Repository
- [ ] Create and Configure GitHub Repository
- [ ] Create a repository named \'CMPG 323 Overview - \<add your student number>' <code>Priority: 1</code>
- [ ] Create the following Milestones: <code>Priority: 2</code>
- [ ] Project 1 Submission: Deadline – 18 August
- [ ] Project 2 Submission: Deadline – 8 September
- [ ] Project 3 Submission: Deadline – 29 September
- [ ] Project 4 Submission: Deadline – 20 October
- [ ] Project 5 Submission: Deadline – 3 November
- [ ] Exam (POE) Submission: Deadline – 18 November
- [ ] Create the following Labels: <code>Priority: 2</code>
- [ ] Class
- [ ] Project
- [ ] Training
- [ ] \<any other labels you might find relevant>
- [ ] Add a ReadME.md file to the repository
- [ ] In the ReadME.md, address which repositories will be created and used for each project <code>Priority: 1</code>
- [ ] In the ReadME.md, provide a diagram explaining project and repository context and how they are integrated <code>Priority: 3</code>
- [ ] In the ReadME.md, explain the branching strategy to be used within each project <code>Priority: 3</code>
- [ ] In the ReadME.md, explain the use of a .gitignore file within each project <code>Priority: 3</code>
- [ ] In the ReadME.md, explain the storage of credentials and sensitive information <code>Priority: 3</code>
- [ ] GitHub Project
- [ ] Create and Configure GitHub project
- [ ] Create a GitHub Kanban project <code>Priority: 1</code>
- [ ] Link the GitHub repository to the project <code>Priority: 2</code>
- [ ] Ensure that the project is named appropriately <code>Priority: 3</code>
- [ ] Ensure that the project has a description adequately describing the project <code>Priority: 3</code>
- [ ] Add the following columns: <code>Priority: 2</code>
- [ ] Linked Assessment (field type = text)
- [ ] Due Date (field type = date)
- [ ] Sprint (field type = single select; add all 8 of the sprints as options with sprint start and end date)
- [ ] Effort (in Hours) (field type = number)
- [ ] Create and populate project views
- [ ] Create a table view within the project named ‘Tabular View’ <code>Priority: 2</code>
- [ ] Add Milestone, Label, Linked Pull Request and Repository as fields to the view
- [ ] Hide the Assignees field
- [ ] Create a board view within the project, named ‘Status View’<code>Priority: 2</code>
- [ ] Group by Status
- [ ] Create a board view within the project, named ‘Sprint View’<code>Priority: 2</code>
- [ ] Group by Sprint
- [ ] Create a table view within the project, named ‘Linked Assessment View’<code>Priority: 3</code>
- [ ] Group by Linked Assessment
- [ ] Create an additional customised view that provides you with additional insight into your semester progress<code>Priority: 4</code>
|
1.0
|
Setup repository - # Requirements
- [ ] GitHub Repository
- [ ] Create and Configure GitHub Repository
- [ ] Create a repository named \'CMPG 323 Overview - \<add your student number>' <code>Priority: 1</code>
- [ ] Create the following Milestones: <code>Priority: 2</code>
- [ ] Project 1 Submission: Deadline – 18 August
- [ ] Project 2 Submission: Deadline – 8 September
- [ ] Project 3 Submission: Deadline – 29 September
- [ ] Project 4 Submission: Deadline – 20 October
- [ ] Project 5 Submission: Deadline – 3 November
- [ ] Exam (POE) Submission: Deadline – 18 November
- [ ] Create the following Labels: <code>Priority: 2</code>
- [ ] Class
- [ ] Project
- [ ] Training
- [ ] \<any other labels you might find relevant>
- [ ] Add a ReadME.md file to the repository
- [ ] In the ReadME.md, address which repositories will be created and used for each project <code>Priority: 1</code>
- [ ] In the ReadME.md, provide a diagram explaining project and repository context and how they are integrated <code>Priority: 3</code>
- [ ] In the ReadME.md, explain the branching strategy to be used within each project <code>Priority: 3</code>
- [ ] In the ReadME.md, explain the use of a .gitignore file within each project <code>Priority: 3</code>
- [ ] In the ReadME.md, explain the storage of credentials and sensitive information <code>Priority: 3</code>
- [ ] GitHub Project
- [ ] Create and Configure GitHub project
- [ ] Create a GitHub Kanban project <code>Priority: 1</code>
- [ ] Link the GitHub repository to the project <code>Priority: 2</code>
- [ ] Ensure that the project is named appropriately <code>Priority: 3</code>
- [ ] Ensure that the project has a description adequately describing the project <code>Priority: 3</code>
- [ ] Add the following columns: <code>Priority: 2</code>
- [ ] Linked Assessment (field type = text)
- [ ] Due Date (field type = date)
- [ ] Sprint (field type = single select; add all 8 of the sprints as options with sprint start and end date)
- [ ] Effort (in Hours) (field type = number)
- [ ] Create and populate project views
- [ ] Create a table view within the project named ‘Tabular View’ <code>Priority: 2</code>
- [ ] Add Milestone, Label, Linked Pull Request and Repository as fields to the view
- [ ] Hide the Assignees field
- [ ] Create a board view within the project, named ‘Status View’<code>Priority: 2</code>
- [ ] Group by Status
- [ ] Create a board view within the project, named ‘Sprint View’<code>Priority: 2</code>
- [ ] Group by Sprint
- [ ] Create a table view within the project, named ‘Linked Assessment View’<code>Priority: 3</code>
- [ ] Group by Linked Assessment
- [ ] Create an additional customised view that provides you with additional insight into your semester progress<code>Priority: 4</code>
|
non_process
|
setup repository requirements github repository create and configure github repository create a repository named cmpg overview priority create the following milestones priority project submission deadline – august project submission deadline – september project submission deadline – september project submission deadline – october project submission deadline – november exam poe submission deadline – november create the following labels priority class project training add a readme md file to the repository in the readme md address which repositories will be created and used for each project priority in the readme md provide a diagram explaining project and repository context and how they are integrated priority in the readme md explain the branching strategy to be used within each project priority in the readme md explain the use of a gitignore file within each project priority in the readme md explain the storage of credentials and sensitive information priority github project create and configure github project create a github kanban project priority link the github repository to the project priority ensure that the project is named appropriately priority ensure that the project has a description adequately describing the project priority add the following columns priority linked assessment field type text due date field type date sprint field type single select add all of the sprints as options with sprint start and end date effort in hours field type number create and populate project views create a table view within the project named ‘tabular view’ priority add milestone label linked pull request and repository as fields to the view hide the assignees field create a board view within the project named ‘status view’ priority group by status create a board view within the project named ‘sprint view’ priority group by sprint create a table view within the project named ‘linked assessment view’ priority group by linked assessment create an additional customised view that provides you with additional insight into your semester progress priority
| 0
|
16,883
| 22,162,690,253
|
IssuesEvent
|
2022-06-04 18:46:43
|
RobertCraigie/prisma-client-py
|
https://api.github.com/repos/RobertCraigie/prisma-client-py
|
closed
|
Prisma CLI and Client use different base paths for relative SQLite files
|
bug/2-confirmed kind/bug process/candidate topic: client level/intermediate priority/high
|
<!--
Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output.
See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output.
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
With this schema as `prisma/schema.prisma`:
```prisma
datasource db {
provider = "sqlite"
url = "file:dev.db"
}
model User {
id String @id @default(cuid())
name String
}
```
Running `prisma db push` creates a database file at `prisma/dev.db`.
Now running the client will error as the query engine uses the path relative to the working directory instead of the schema directory, e.g.
`main.py`
```python
import asyncio
from prisma import Prisma
from prisma.models import User
async def bar() -> None:
p = Prisma(auto_register=True)
await p.connect()
await User.prisma().create(data={'name': 'Robert'})
asyncio.run(bar())
```
Will use `dev.db` instead of `prisma/dev.db`.
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
The CLI and the Client should use the same database path.
|
1.0
|
Prisma CLI and Client use different base paths for relative SQLite files - <!--
Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output.
See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output.
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
With this schema as `prisma/schema.prisma`:
```prisma
datasource db {
provider = "sqlite"
url = "file:dev.db"
}
model User {
id String @id @default(cuid())
name String
}
```
Running `prisma db push` creates a database file at `prisma/dev.db`.
Now running the client will error as the query engine uses the path relative to the working directory instead of the schema directory, e.g.
`main.py`
```python
import asyncio
from prisma import Prisma
from prisma.models import User
async def bar() -> None:
p = Prisma(auto_register=True)
await p.connect()
await User.prisma().create(data={'name': 'Robert'})
asyncio.run(bar())
```
Will use `dev.db` instead of `prisma/dev.db`.
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
The CLI and the Client should use the same database path.
|
process
|
prisma cli and client use different base paths for relative sqlite files thanks for helping us improve prisma client python 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by enabling additional logging output see for how to enable additional logging output bug description with this schema as prisma schema prisma prisma datasource db provider sqlite url file dev db model user id string id default cuid name string running prisma db push creates a database file at prisma dev db now running the client will error as the query engine uses the path relative to the working directory instead of the schema directory e g main py python import asyncio from prisma import prisma from prisma models import user async def bar none p prisma auto register true await p connect await user prisma create data name robert asyncio run bar will use dev db instead of prisma dev db expected behavior the cli and the client should use the same database path
| 1
|
20,131
| 26,669,538,478
|
IssuesEvent
|
2023-01-26 09:02:10
|
googleapis/google-cloud-php
|
https://api.github.com/repos/googleapis/google-cloud-php
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'longrunning' invalid in LongRunning/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'longrunning' invalid in LongRunning/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname longrunning invalid in longrunning repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
17,077
| 2,974,593,164
|
IssuesEvent
|
2015-07-15 02:10:24
|
Reimashi/jotai
|
https://api.github.com/repos/Reimashi/jotai
|
closed
|
System is set for F° yet when placed in tray it shows C° it does show F° in gadget box
|
auto-migrated Priority-Medium Type-Defect wontfix
|
```
Using windows XP SP3
```
Original issue reported on code.google.com by `LEGALVID...@aol.com` on 7 May 2012 at 1:26
|
1.0
|
System is set for F° yet when placed in tray it shows C° it does show F° in gadget box - ```
Using windows XP SP3
```
Original issue reported on code.google.com by `LEGALVID...@aol.com` on 7 May 2012 at 1:26
|
non_process
|
system is set for f° yet when placed in tray it shows c° it does show f° in gadget box using windows xp original issue reported on code google com by legalvid aol com on may at
| 0
|
9,307
| 12,321,395,489
|
IssuesEvent
|
2020-05-13 08:39:15
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Step 6 The command is not displayed even if you enter "Connect-AzAccount"
|
Pri1 automation/svc cxp process-automation/subsvc product-question triaged
|
The comments below were machine-translated
@kurihiro0119 commented on [Wed May 06 2020](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/4528)
Follow step 6 and enter the form on the left that says Search Library Items
Nothing is displayed when I enter "Connect-AzAccount".
Therefore, the corresponding command cannot be set and the tutorial ends halfway.
---
#### ドキュメントの詳細
⚠ *このセクションを編集しないでください。 docs.microsoft.com で必須です ➟ GitHub の問題のリンク。*
* ID: 4799ebe4-d654-b9d2-e64f-923d025ad7cf
* Version Independent ID: f6519afb-bf55-c886-e3c3-d11174da3aea
* Content: [Azure Automation でグラフィカル Runbook を作成する](https://docs.microsoft.com/ja-jp/azure/automation/learn/automation-tutorial-runbook-graphical#feedback)
* Content Source: [articles/automation/learn/automation-tutorial-runbook-graphical.md](https://github.com/MicrosoftDocs/azure-docs.ja-jp/blob/master/articles/automation/learn/automation-tutorial-runbook-graphical.md)
* Sub-service: **process-automation**
* GitHub Login: @olprod
---
@srvbpigh commented on [Wed May 06 2020](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/4528#issuecomment-624614712)
Hello, @kurihiro0119
Thank you for your feedback.
We are actively reviewing your comments and will get back to you soon.
Kind regards,
Microsoft DOCS International Team
|
1.0
|
Step 6 The command is not displayed even if you enter "Connect-AzAccount" - The comments below were machine-translated
@kurihiro0119 commented on [Wed May 06 2020](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/4528)
Follow step 6 and enter the form on the left that says Search Library Items
Nothing is displayed when I enter "Connect-AzAccount".
Therefore, the corresponding command cannot be set and the tutorial ends halfway.
---
#### ドキュメントの詳細
⚠ *このセクションを編集しないでください。 docs.microsoft.com で必須です ➟ GitHub の問題のリンク。*
* ID: 4799ebe4-d654-b9d2-e64f-923d025ad7cf
* Version Independent ID: f6519afb-bf55-c886-e3c3-d11174da3aea
* Content: [Azure Automation でグラフィカル Runbook を作成する](https://docs.microsoft.com/ja-jp/azure/automation/learn/automation-tutorial-runbook-graphical#feedback)
* Content Source: [articles/automation/learn/automation-tutorial-runbook-graphical.md](https://github.com/MicrosoftDocs/azure-docs.ja-jp/blob/master/articles/automation/learn/automation-tutorial-runbook-graphical.md)
* Sub-service: **process-automation**
* GitHub Login: @olprod
---
@srvbpigh commented on [Wed May 06 2020](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/4528#issuecomment-624614712)
Hello, @kurihiro0119
Thank you for your feedback.
We are actively reviewing your comments and will get back to you soon.
Kind regards,
Microsoft DOCS International Team
|
process
|
step the command is not displayed even if you enter connect azaccount the comments below were machine translated commented on follow step and enter the form on the left that says search library items nothing is displayed when i enter connect azaccount therefore the corresponding command cannot be set and the tutorial ends halfway ドキュメントの詳細 ⚠ このセクションを編集しないでください。 docs microsoft com で必須です ➟ github の問題のリンク。 id version independent id content content source sub service process automation github login olprod srvbpigh commented on hello thank you for your feedback we are actively reviewing your comments and will get back to you soon kind regards microsoft docs international team
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.