Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
418
| 2,852,501,524
|
IssuesEvent
|
2015-06-01 13:58:04
|
genomizer/genomizer-server
|
https://api.github.com/repos/genomizer/genomizer-server
|
closed
|
File cleanup utility cannot delete all temporary folders after a process execution
|
Medium priority Processing
|
The problem could be due to a process holding a file stream that is not closed at the time file cleanup utility is called.
|
1.0
|
File cleanup utility cannot delete all temporary folders after a process execution - The problem could be due to a process holding a file stream that is not closed at the time file cleanup utility is called.
|
process
|
file cleanup utility cannot delete all temporary folders after a process execution the problem could be due to a process holding a file stream that is not closed at the time file cleanup utility is called
| 1
|
13,142
| 3,117,397,008
|
IssuesEvent
|
2015-09-04 00:48:22
|
meren/anvio
|
https://api.github.com/repos/meren/anvio
|
closed
|
annotation.db needs to know which tables are populated
|
bug design
|
_self_ table in the annotation.db should be updated when any of the papi-populate-*-table scripts are run.
it should be clear to the profiler whether a papi-populate-*-table was run and the results were empty, or it was never run at all.
|
1.0
|
annotation.db needs to know which tables are populated - _self_ table in the annotation.db should be updated when any of the papi-populate-*-table scripts are run.
it should be clear to the profiler whether a papi-populate-*-table was run and the results were empty, or it was never run at all.
|
non_process
|
annotation db needs to know which tables are populated self table in the annotation db should be updated when any of the papi populate table scripts are run it should be clear to the profiler whether a papi populate table was run and the results were empty or it was never run at all
| 0
|
168,658
| 14,169,199,961
|
IssuesEvent
|
2020-11-12 12:53:45
|
selectize/selectize.js
|
https://api.github.com/repos/selectize/selectize.js
|
closed
|
Documentation Request: Nested Properties
|
documentation
|
Looking at current documentation it notes that sifter.js has the ability to filter results based on nested properties. Looking at docs for both selectize and sifter doesn't give a clear indication of how to do so. If the values you are looking for are not in the top level search fields then no results will be displayed. However, doing a blur/focus on a selectize element will display the results.
While I have tried a few different methods I can't find any way to tell selectize which nested fields to look in. If some information can be provided on how to do so, I can help and contribute the basic docs/examples.
|
1.0
|
Documentation Request: Nested Properties - Looking at current documentation it notes that sifter.js has the ability to filter results based on nested properties. Looking at docs for both selectize and sifter doesn't give a clear indication of how to do so. If the values you are looking for are not in the top level search fields then no results will be displayed. However, doing a blur/focus on a selectize element will display the results.
While I have tried a few different methods I can't find any way to tell selectize which nested fields to look in. If some information can be provided on how to do so, I can help and contribute the basic docs/examples.
|
non_process
|
documentation request nested properties looking at current documentation it notes that sifter js has the ability to filter results based on nested properties looking at docs for both selectize and sifter doesn t give a clear indication of how to do so if the values you are looking for are not in the top level search fields then no results will be displayed however doing a blur focus on a selectize element will display the results while i have tried a few different methods i can t find any way to tell selectize which nested fields to look in if some information can be provided on how to do so i can help and contribute the basic docs examples
| 0
|
447,345
| 31,683,961,587
|
IssuesEvent
|
2023-09-08 04:01:27
|
apecloud/kubeblocks
|
https://api.github.com/repos/apecloud/kubeblocks
|
opened
|
[Features] Cost extension
|
kind/feature area/user-interaction documentation
|
Problems:
With the development of customer business, there are more and more database clusters running in KubeBlocks, the resources used by database clusters are a black box for users, opening the black box will help users better manage the database, control costs, and increase the added value of KubeBlocks.
Goals:
With this in mind, we are launching KubeBlocks Cost extension, a cost insight plugin that aims to:
- Provides users with comprehensive insight into the resource information of database clusters in KubeBlocks
- Help users understand the resource utilization of the database cluster and provide cost optimization suggestions
Requirements:
- Run instance analysis
- Resource utilization analysis
- Cost estimation
- Suggestions for cost optimization
|
1.0
|
[Features] Cost extension - Problems:
With the development of customer business, there are more and more database clusters running in KubeBlocks, the resources used by database clusters are a black box for users, opening the black box will help users better manage the database, control costs, and increase the added value of KubeBlocks.
Goals:
With this in mind, we are launching KubeBlocks Cost extension, a cost insight plugin that aims to:
- Provides users with comprehensive insight into the resource information of database clusters in KubeBlocks
- Help users understand the resource utilization of the database cluster and provide cost optimization suggestions
Requirements:
- Run instance analysis
- Resource utilization analysis
- Cost estimation
- Suggestions for cost optimization
|
non_process
|
cost extension problems with the development of customer business there are more and more database clusters running in kubeblocks the resources used by database clusters are a black box for users opening the black box will help users better manage the database control costs and increase the added value of kubeblocks goals with this in mind we are launching kubeblocks cost extension a cost insight plugin that aims to provides users with comprehensive insight into the resource information of database clusters in kubeblocks help users understand the resource utilization of the database cluster and provide cost optimization suggestions requirements run instance analysis resource utilization analysis cost estimation suggestions for cost optimization
| 0
|
2,529
| 5,288,868,862
|
IssuesEvent
|
2017-02-08 16:06:52
|
mesosphere/marathon
|
https://api.github.com/repos/mesosphere/marathon
|
closed
|
scale test better error reporting
|
Epic:Improve CI and Release Process sprint-3
|
hashmap of errors that are in loops,
add errors with counts at the end of loop
|
1.0
|
scale test better error reporting - hashmap of errors that are in loops,
add errors with counts at the end of loop
|
process
|
scale test better error reporting hashmap of errors that are in loops add errors with counts at the end of loop
| 1
|
5,624
| 8,481,798,030
|
IssuesEvent
|
2018-10-25 16:40:32
|
googleapis/google-cloud-node
|
https://api.github.com/repos/googleapis/google-cloud-node
|
closed
|
Add Node 11 to the CI config
|
status: blocked type: process
|
Note: we need to wait for Node.js 11 to be released, which should happen "soon" ™️
|
1.0
|
Add Node 11 to the CI config - Note: we need to wait for Node.js 11 to be released, which should happen "soon" ™️
|
process
|
add node to the ci config note we need to wait for node js to be released which should happen soon ™️
| 1
|
2,492
| 5,267,653,382
|
IssuesEvent
|
2017-02-05 00:54:14
|
coala/teams
|
https://api.github.com/repos/coala/teams
|
closed
|
Core Team Member Application: Rohan VB
|
process/approved
|
# Bio
I'm Rohan. I'm currently a 3rd year undergrad student of Maths & Computing in India.
# coala Contributions so far
My main contributions have been to the coala/coala repo, solving and opening issues. I've also reviewed a few PRs in the same repo. Other contributions have been in the coala/documentation repo.
# Road to the Future
I think reviewing PRs and helping newcomers get adjusted to the coala workflow is extremely important.
Besides that, I'm interested in ways we can make coala more end-user friendly, as mentioned below, I think better plugin support would be a major step forward.
# Specific Responsibilities
I'm currently interested in the project of **better editor support** which involves implementing plugins for other editors and to have CI along with tests functioning. I've started some work on the coala plugin for VScode, and I'm also interested in solving the issues of the plugins which coala already has which don't get much attention as the main repo - including: coala-sublime. Other than that, I'm open to new possibilities :)
|
1.0
|
Core Team Member Application: Rohan VB - # Bio
I'm Rohan. I'm currently a 3rd year undergrad student of Maths & Computing in India.
# coala Contributions so far
My main contributions have been to the coala/coala repo, solving and opening issues. I've also reviewed a few PRs in the same repo. Other contributions have been in the coala/documentation repo.
# Road to the Future
I think reviewing PRs and helping newcomers get adjusted to the coala workflow is extremely important.
Besides that, I'm interested in ways we can make coala more end-user friendly, as mentioned below, I think better plugin support would be a major step forward.
# Specific Responsibilities
I'm currently interested in the project of **better editor support** which involves implementing plugins for other editors and to have CI along with tests functioning. I've started some work on the coala plugin for VScode, and I'm also interested in solving the issues of the plugins which coala already has which don't get much attention as the main repo - including: coala-sublime. Other than that, I'm open to new possibilities :)
|
process
|
core team member application rohan vb bio i m rohan i m currently a year undergrad student of maths computing in india coala contributions so far my main contributions have been to the coala coala repo solving and opening issues i ve also reviewed a few prs in the same repo other contributions have been in the coala documentation repo road to the future i think reviewing prs and helping newcomers get adjusted to the coala workflow is extremely important besides that i m interested in ways we can make coala more end user friendly as mentioned below i think better plugin support would be a major step forward specific responsibilities i m currently interested in the project of better editor support which involves implementing plugins for other editors and to have ci along with tests functioning i ve started some work on the coala plugin for vscode and i m also interested in solving the issues of the plugins which coala already has which don t get much attention as the main repo including coala sublime other than that i m open to new possibilities
| 1
|
17,116
| 22,635,800,957
|
IssuesEvent
|
2022-06-30 18:51:03
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Add optional --no-python argument for qgis_process to disable Python support (Request in QGIS)
|
Processing 3.24
|
### Request for documentation
From pull request QGIS/qgis#46579
Author: @nyalldawson
QGIS version: 3.24
**Add optional --no-python argument for qgis_process to disable Python support**
### PR Description:
Skipping python support when it's not required can result in a significant improvement in qgis_process startup times
### Commits tagged with [need-docs] or [FEATURE]
|
1.0
|
Add optional --no-python argument for qgis_process to disable Python support (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#46579
Author: @nyalldawson
QGIS version: 3.24
**Add optional --no-python argument for qgis_process to disable Python support**
### PR Description:
Skipping python support when it's not required can result in a significant improvement in qgis_process startup times
### Commits tagged with [need-docs] or [FEATURE]
|
process
|
add optional no python argument for qgis process to disable python support request in qgis request for documentation from pull request qgis qgis author nyalldawson qgis version add optional no python argument for qgis process to disable python support pr description skipping python support when it s not required can result in a significant improvement in qgis process startup times commits tagged with or
| 1
|
11,104
| 13,941,864,695
|
IssuesEvent
|
2020-10-22 20:05:21
|
googleapis/google-auth-library-python
|
https://api.github.com/repos/googleapis/google-auth-library-python
|
closed
|
Tests broken on Python < 3.8
|
testing type: process
|
```python
_______ ERROR collecting tests_async/transport/test_aiohttp_requests.py ________
ImportError while importing test module '/home/tseaver/projects/agendaless/Google/src/google-auth/tests_async/transport/test_aiohttp_requests.py'.
Hint: make sure your test modules/packages have valid Python names.
Traceback:
.nox/unit-3-6/lib/python3.6/site-packages/aioresponses/compat.py:14: in <module>
from unittest import IsolatedAsyncioTestCase, skipIf
E ImportError: cannot import name 'IsolatedAsyncioTestCase'
During handling of the above exception, another exception occurred:
/opt/Python-3.6.10/lib/python3.6/importlib/__init__.py:126: in import_module
return _bootstrap._gcd_import(name[level:], package, level)
tests_async/transport/test_aiohttp_requests.py:16: in <module>
from aioresponses import aioresponses, core
.nox/unit-3-6/lib/python3.6/site-packages/aioresponses/__init__.py:2: in <module>
from .core import CallbackResult, aioresponses
.nox/unit-3-6/lib/python3.6/site-packages/aioresponses/core.py:23: in <module>
from .compat import (
.nox/unit-3-6/lib/python3.6/site-packages/aioresponses/compat.py:29: in <module>
from asynctest import fail_on, skipIf
E ModuleNotFoundError: No module named 'asynctest'
```
Due to https://github.com/pnuckowski/aioresponses/issues/172
|
1.0
|
Tests broken on Python < 3.8 - ```python
_______ ERROR collecting tests_async/transport/test_aiohttp_requests.py ________
ImportError while importing test module '/home/tseaver/projects/agendaless/Google/src/google-auth/tests_async/transport/test_aiohttp_requests.py'.
Hint: make sure your test modules/packages have valid Python names.
Traceback:
.nox/unit-3-6/lib/python3.6/site-packages/aioresponses/compat.py:14: in <module>
from unittest import IsolatedAsyncioTestCase, skipIf
E ImportError: cannot import name 'IsolatedAsyncioTestCase'
During handling of the above exception, another exception occurred:
/opt/Python-3.6.10/lib/python3.6/importlib/__init__.py:126: in import_module
return _bootstrap._gcd_import(name[level:], package, level)
tests_async/transport/test_aiohttp_requests.py:16: in <module>
from aioresponses import aioresponses, core
.nox/unit-3-6/lib/python3.6/site-packages/aioresponses/__init__.py:2: in <module>
from .core import CallbackResult, aioresponses
.nox/unit-3-6/lib/python3.6/site-packages/aioresponses/core.py:23: in <module>
from .compat import (
.nox/unit-3-6/lib/python3.6/site-packages/aioresponses/compat.py:29: in <module>
from asynctest import fail_on, skipIf
E ModuleNotFoundError: No module named 'asynctest'
```
Due to https://github.com/pnuckowski/aioresponses/issues/172
|
process
|
tests broken on python python error collecting tests async transport test aiohttp requests py importerror while importing test module home tseaver projects agendaless google src google auth tests async transport test aiohttp requests py hint make sure your test modules packages have valid python names traceback nox unit lib site packages aioresponses compat py in from unittest import isolatedasynciotestcase skipif e importerror cannot import name isolatedasynciotestcase during handling of the above exception another exception occurred opt python lib importlib init py in import module return bootstrap gcd import name package level tests async transport test aiohttp requests py in from aioresponses import aioresponses core nox unit lib site packages aioresponses init py in from core import callbackresult aioresponses nox unit lib site packages aioresponses core py in from compat import nox unit lib site packages aioresponses compat py in from asynctest import fail on skipif e modulenotfounderror no module named asynctest due to
| 1
|
206,873
| 23,405,108,157
|
IssuesEvent
|
2022-08-12 12:03:23
|
nexmo-community/phone-number-pool-manager-node
|
https://api.github.com/repos/nexmo-community/phone-number-pool-manager-node
|
opened
|
axios-0.19.2.tgz: 4 vulnerabilities (highest severity is: 7.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>axios-0.19.2.tgz</b></p></summary>
<p>Promise based HTTP client for the browser and node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.19.2.tgz">https://registry.npmjs.org/axios/-/axios-0.19.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/axios/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-3749](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3749) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | axios-0.19.2.tgz | Direct | 0.20.0 | ✅ |
| [CVE-2022-0155](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | follow-redirects-1.5.10.tgz | Transitive | 0.20.0-0 | ✅ |
| [CVE-2022-0536](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0536) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | follow-redirects-1.5.10.tgz | Transitive | 0.20.0-0 | ✅ |
| [CVE-2020-28168](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28168) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | axios-0.19.2.tgz | Direct | 0.21.1 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3749</summary>
### Vulnerable Library - <b>axios-0.19.2.tgz</b></p>
<p>Promise based HTTP client for the browser and node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.19.2.tgz">https://registry.npmjs.org/axios/-/axios-0.19.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/axios/package.json</p>
<p>
Dependency Hierarchy:
- :x: **axios-0.19.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
axios is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3749>CVE-2021-3749</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/1e8f07fc-c384-4ff9-8498-0690de2e8c31/">https://huntr.dev/bounties/1e8f07fc-c384-4ff9-8498-0690de2e8c31/</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution: 0.20.0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0155</summary>
### Vulnerable Library - <b>follow-redirects-1.5.10.tgz</b></p>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.5.10.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.5.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- axios-0.19.2.tgz (Root Library)
- :x: **follow-redirects-1.5.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
follow-redirects is vulnerable to Exposure of Private Personal Information to an Unauthorized Actor
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155>CVE-2022-0155</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/">https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/</a></p>
<p>Release Date: 2022-01-10</p>
<p>Fix Resolution (follow-redirects): 1.14.7</p>
<p>Direct dependency fix Resolution (axios): 0.20.0-0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0536</summary>
### Vulnerable Library - <b>follow-redirects-1.5.10.tgz</b></p>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.5.10.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.5.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- axios-0.19.2.tgz (Root Library)
- :x: **follow-redirects-1.5.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Exposure of Sensitive Information to an Unauthorized Actor in NPM follow-redirects prior to 1.14.8.
<p>Publish Date: 2022-02-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0536>CVE-2022-0536</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0536">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0536</a></p>
<p>Release Date: 2022-02-09</p>
<p>Fix Resolution (follow-redirects): 1.14.8</p>
<p>Direct dependency fix Resolution (axios): 0.20.0-0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28168</summary>
### Vulnerable Library - <b>axios-0.19.2.tgz</b></p>
<p>Promise based HTTP client for the browser and node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.19.2.tgz">https://registry.npmjs.org/axios/-/axios-0.19.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/axios/package.json</p>
<p>
Dependency Hierarchy:
- :x: **axios-0.19.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Axios NPM package 0.21.0 contains a Server-Side Request Forgery (SSRF) vulnerability where an attacker is able to bypass a proxy by providing a URL that responds with a redirect to a restricted host or IP address.
<p>Publish Date: 2020-11-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28168>CVE-2020-28168</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-11-06</p>
<p>Fix Resolution: 0.21.1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
True
|
axios-0.19.2.tgz: 4 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>axios-0.19.2.tgz</b></p></summary>
<p>Promise based HTTP client for the browser and node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.19.2.tgz">https://registry.npmjs.org/axios/-/axios-0.19.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/axios/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-3749](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3749) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | axios-0.19.2.tgz | Direct | 0.20.0 | ✅ |
| [CVE-2022-0155](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | follow-redirects-1.5.10.tgz | Transitive | 0.20.0-0 | ✅ |
| [CVE-2022-0536](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0536) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | follow-redirects-1.5.10.tgz | Transitive | 0.20.0-0 | ✅ |
| [CVE-2020-28168](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28168) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | axios-0.19.2.tgz | Direct | 0.21.1 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3749</summary>
### Vulnerable Library - <b>axios-0.19.2.tgz</b></p>
<p>Promise based HTTP client for the browser and node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.19.2.tgz">https://registry.npmjs.org/axios/-/axios-0.19.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/axios/package.json</p>
<p>
Dependency Hierarchy:
- :x: **axios-0.19.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
axios is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3749>CVE-2021-3749</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/1e8f07fc-c384-4ff9-8498-0690de2e8c31/">https://huntr.dev/bounties/1e8f07fc-c384-4ff9-8498-0690de2e8c31/</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution: 0.20.0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0155</summary>
### Vulnerable Library - <b>follow-redirects-1.5.10.tgz</b></p>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.5.10.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.5.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- axios-0.19.2.tgz (Root Library)
- :x: **follow-redirects-1.5.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
follow-redirects is vulnerable to Exposure of Private Personal Information to an Unauthorized Actor
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155>CVE-2022-0155</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/">https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/</a></p>
<p>Release Date: 2022-01-10</p>
<p>Fix Resolution (follow-redirects): 1.14.7</p>
<p>Direct dependency fix Resolution (axios): 0.20.0-0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0536</summary>
### Vulnerable Library - <b>follow-redirects-1.5.10.tgz</b></p>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.5.10.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.5.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- axios-0.19.2.tgz (Root Library)
- :x: **follow-redirects-1.5.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Exposure of Sensitive Information to an Unauthorized Actor in NPM follow-redirects prior to 1.14.8.
<p>Publish Date: 2022-02-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0536>CVE-2022-0536</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0536">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0536</a></p>
<p>Release Date: 2022-02-09</p>
<p>Fix Resolution (follow-redirects): 1.14.8</p>
<p>Direct dependency fix Resolution (axios): 0.20.0-0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28168</summary>
### Vulnerable Library - <b>axios-0.19.2.tgz</b></p>
<p>Promise based HTTP client for the browser and node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.19.2.tgz">https://registry.npmjs.org/axios/-/axios-0.19.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/axios/package.json</p>
<p>
Dependency Hierarchy:
- :x: **axios-0.19.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/phone-number-pool-manager-node/commit/7aa742cf79633e383b8b8adfc730c5a72d355f98">7aa742cf79633e383b8b8adfc730c5a72d355f98</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Axios NPM package 0.21.0 contains a Server-Side Request Forgery (SSRF) vulnerability where an attacker is able to bypass a proxy by providing a URL that responds with a redirect to a restricted host or IP address.
<p>Publish Date: 2020-11-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28168>CVE-2020-28168</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-11-06</p>
<p>Fix Resolution: 0.21.1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
non_process
|
axios tgz vulnerabilities highest severity is vulnerable library axios tgz promise based http client for the browser and node js library home page a href path to dependency file package json path to vulnerable library node modules axios package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high axios tgz direct medium follow redirects tgz transitive medium follow redirects tgz transitive medium axios tgz direct details cve vulnerable library axios tgz promise based http client for the browser and node js library home page a href path to dependency file package json path to vulnerable library node modules axios package json dependency hierarchy x axios tgz vulnerable library found in head commit a href found in base branch main vulnerability details axios is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file package json path to vulnerable library node modules follow redirects package json dependency hierarchy axios tgz root library x follow redirects tgz vulnerable library found in head commit a href found in base branch main vulnerability details follow redirects is vulnerable to exposure of private personal information to an unauthorized actor publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution follow redirects direct dependency fix resolution axios rescue worker helmet automatic remediation is available for this issue cve vulnerable library follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file package json path to vulnerable library node modules follow redirects package json dependency hierarchy axios tgz root library x follow redirects tgz vulnerable library found in head commit a href found in base branch main vulnerability details exposure of sensitive information to an unauthorized actor in npm follow redirects prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution follow redirects direct dependency fix resolution axios rescue worker helmet automatic remediation is available for this issue cve vulnerable library axios tgz promise based http client for the browser and node js library home page a href path to dependency file package json path to vulnerable library node modules axios package json dependency hierarchy x axios tgz vulnerable library found in head commit a href found in base branch main vulnerability details axios npm package contains a server side request forgery ssrf vulnerability where an attacker is able to bypass a proxy by providing a url that responds with a redirect to a restricted host or ip address publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
| 0
|
9,787
| 12,804,290,263
|
IssuesEvent
|
2020-07-03 03:58:53
|
OUDcollective/twenty20times
|
https://api.github.com/repos/OUDcollective/twenty20times
|
closed
|
Basic writing and formatting syntax - GitHub Help
|
workflow-process
|

## ON RELATIVE LINKS
---
**Source URL**:
[https://help.github.com/en/github/writing-on-github/basic-writing-and-formatting-syntax](https://help.github.com/en/github/writing-on-github/basic-writing-and-formatting-syntax)
<table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
|
1.0
|
Basic writing and formatting syntax - GitHub Help - 
## ON RELATIVE LINKS
---
**Source URL**:
[https://help.github.com/en/github/writing-on-github/basic-writing-and-formatting-syntax](https://help.github.com/en/github/writing-on-github/basic-writing-and-formatting-syntax)
<table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
|
process
|
basic writing and formatting syntax github help on relative links source url browser chrome os windows bit screen size viewport size pixel ratio zoom level
| 1
|
16,229
| 20,766,312,302
|
IssuesEvent
|
2022-03-15 20:58:41
|
GSA/CIW
|
https://api.github.com/repos/GSA/CIW
|
closed
|
CIW Bug Fixes: 2022-03
|
Type: Bug Topic: Upload/Processing Topic: Validation
|
Fixing issues resulting from the release of #220, listed below:
- [x] Allow valid childcare CIWs to be passed in with no start or end date
- [x] Update "last updated" and "last updated by" fields whenever a contract's end date is updated by the CIW process
|
1.0
|
CIW Bug Fixes: 2022-03 - Fixing issues resulting from the release of #220, listed below:
- [x] Allow valid childcare CIWs to be passed in with no start or end date
- [x] Update "last updated" and "last updated by" fields whenever a contract's end date is updated by the CIW process
|
process
|
ciw bug fixes fixing issues resulting from the release of listed below allow valid childcare ciws to be passed in with no start or end date update last updated and last updated by fields whenever a contract s end date is updated by the ciw process
| 1
|
146,335
| 19,399,935,679
|
IssuesEvent
|
2021-12-19 01:21:08
|
SmartBear/ready-jira-plugin
|
https://api.github.com/repos/SmartBear/ready-jira-plugin
|
opened
|
CVE-2021-45105 (High) detected in log4j-core-2.11.0.jar
|
security vulnerability
|
## CVE-2021-45105 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-core-2.11.0.jar</b></p></summary>
<p>The Apache Log4j Implementation</p>
<p>Library home page: <a href="https://logging.apache.org/log4j/2.x/">https://logging.apache.org/log4j/2.x/</a></p>
<p>Path to dependency file: ready-jira-plugin/pom.xml</p>
<p>Path to vulnerable library: ory/org/apache/logging/log4j/log4j-core/2.11.0/log4j-core-2.11.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-core-2.11.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Log4J 2.0-beta9 before 2.17.0 is vulnerable to Denial of Service (DoS) attacks. Due to lack of protection from uncontrolled recursion from self-referential lookups, attackers with control over Thread Context Map (MDC) input data can craft malicious input data that contains a recursive lookup, resulting in a StackOverflowError that will terminate the process.
<p>Publish Date: 2021-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-45105>CVE-2021-45105</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://logging.apache.org/log4j/2.x/security.html">https://logging.apache.org/log4j/2.x/security.html</a></p>
<p>Release Date: 2021-12-17</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.17.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.11.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.17.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-45105","vulnerabilityDetails":"Apache Log4J 2.0-beta9 before 2.17.0 is vulnerable to Denial of Service (DoS) attacks. Due to lack of protection from uncontrolled recursion from self-referential lookups, attackers with control over Thread Context Map (MDC) input data can craft malicious input data that contains a recursive lookup, resulting in a StackOverflowError that will terminate the process.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-45105","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-45105 (High) detected in log4j-core-2.11.0.jar - ## CVE-2021-45105 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-core-2.11.0.jar</b></p></summary>
<p>The Apache Log4j Implementation</p>
<p>Library home page: <a href="https://logging.apache.org/log4j/2.x/">https://logging.apache.org/log4j/2.x/</a></p>
<p>Path to dependency file: ready-jira-plugin/pom.xml</p>
<p>Path to vulnerable library: ory/org/apache/logging/log4j/log4j-core/2.11.0/log4j-core-2.11.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-core-2.11.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Log4J 2.0-beta9 before 2.17.0 is vulnerable to Denial of Service (DoS) attacks. Due to lack of protection from uncontrolled recursion from self-referential lookups, attackers with control over Thread Context Map (MDC) input data can craft malicious input data that contains a recursive lookup, resulting in a StackOverflowError that will terminate the process.
<p>Publish Date: 2021-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-45105>CVE-2021-45105</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://logging.apache.org/log4j/2.x/security.html">https://logging.apache.org/log4j/2.x/security.html</a></p>
<p>Release Date: 2021-12-17</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.17.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.11.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.17.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-45105","vulnerabilityDetails":"Apache Log4J 2.0-beta9 before 2.17.0 is vulnerable to Denial of Service (DoS) attacks. Due to lack of protection from uncontrolled recursion from self-referential lookups, attackers with control over Thread Context Map (MDC) input data can craft malicious input data that contains a recursive lookup, resulting in a StackOverflowError that will terminate the process.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-45105","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in core jar cve high severity vulnerability vulnerable library core jar the apache implementation library home page a href path to dependency file ready jira plugin pom xml path to vulnerable library ory org apache logging core core jar dependency hierarchy x core jar vulnerable library found in base branch master vulnerability details apache before is vulnerable to denial of service dos attacks due to lack of protection from uncontrolled recursion from self referential lookups attackers with control over thread context map mdc input data can craft malicious input data that contains a recursive lookup resulting in a stackoverflowerror that will terminate the process publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache logging core isminimumfixversionavailable true minimumfixversion org apache logging core isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails apache before is vulnerable to denial of service dos attacks due to lack of protection from uncontrolled recursion from self referential lookups attackers with control over thread context map mdc input data can craft malicious input data that contains a recursive lookup resulting in a stackoverflowerror that will terminate the process vulnerabilityurl
| 0
|
35,196
| 14,629,917,521
|
IssuesEvent
|
2020-12-23 16:42:07
|
trisacrypto/testnet
|
https://api.github.com/repos/trisacrypto/testnet
|
opened
|
Lookup by common name is returning 404
|
bug directory service
|
On the live server lookup by common name is returning 404 even though lookup by id is working and returning the common name that was requested.
|
1.0
|
Lookup by common name is returning 404 - On the live server lookup by common name is returning 404 even though lookup by id is working and returning the common name that was requested.
|
non_process
|
lookup by common name is returning on the live server lookup by common name is returning even though lookup by id is working and returning the common name that was requested
| 0
|
259,279
| 27,621,791,480
|
IssuesEvent
|
2023-03-10 01:11:56
|
artsking/linux-4.19.72
|
https://api.github.com/repos/artsking/linux-4.19.72
|
opened
|
CVE-2023-1074 (Medium) detected in linux-yoctov5.4.51
|
Mend: dependency security vulnerability
|
## CVE-2023-1074 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry.
This page will reflect the classification results once they are available through NVD
<p>Publish Date: 2023-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1074>CVE-2023-1074</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1074">https://www.linuxkernelcves.com/cves/CVE-2023-1074</a></p>
<p>Release Date: 2023-02-28</p>
<p>Fix Resolution: v4.14.305,v4.19.272,v5.4.231,v5.10.166,v5.15.91,v6.1.9,v6.2-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-1074 (Medium) detected in linux-yoctov5.4.51 - ## CVE-2023-1074 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry.
This page will reflect the classification results once they are available through NVD
<p>Publish Date: 2023-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1074>CVE-2023-1074</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1074">https://www.linuxkernelcves.com/cves/CVE-2023-1074</a></p>
<p>Release Date: 2023-02-28</p>
<p>Fix Resolution: v4.14.305,v4.19.272,v5.4.231,v5.10.166,v5.15.91,v6.1.9,v6.2-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in base branch master vulnerable source files net sctp bind addr c net sctp bind addr c vulnerability details this cve was marked as reserved by nvd this means that the cve id is reserved for future use by the cve numbering authority cna or a security researcher still the cve details were not included in the cve entry this is because the original requester of the cve id assignment has not sent an update to mitre with the information needed to populate the cve entry this page will reflect the classification results once they are available through nvd publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
2,403
| 5,193,025,398
|
IssuesEvent
|
2017-01-22 15:18:19
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Wrong header items width
|
inprocess
|
The table header's width is not calculated the same way as the table's body, ending up in situations like this (note the header overflowing on the right):

**[Edit: actually it happens even without this, see below]** It happened because the bar on the left appears after a transition, and the whole layout re-rendered accordingly but the header did not.
**[Edit: just missing CSS files]** Also probably related: the width of a menu item is calculated before the sorting arrows are placed, which can produce this:

I could solve this by setting the `width` property, but I don't want fixed with, and no `minWidth` prop is available.
|
1.0
|
Wrong header items width - The table header's width is not calculated the same way as the table's body, ending up in situations like this (note the header overflowing on the right):

**[Edit: actually it happens even without this, see below]** It happened because the bar on the left appears after a transition, and the whole layout re-rendered accordingly but the header did not.
**[Edit: just missing CSS files]** Also probably related: the width of a menu item is calculated before the sorting arrows are placed, which can produce this:

I could solve this by setting the `width` property, but I don't want fixed with, and no `minWidth` prop is available.
|
process
|
wrong header items width the table header s width is not calculated the same way as the table s body ending up in situations like this note the header overflowing on the right it happened because the bar on the left appears after a transition and the whole layout re rendered accordingly but the header did not also probably related the width of a menu item is calculated before the sorting arrows are placed which can produce this i could solve this by setting the width property but i don t want fixed with and no minwidth prop is available
| 1
|
20,078
| 26,573,081,785
|
IssuesEvent
|
2023-01-21 12:40:38
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
Decompiler interprets SHUFPS Z values as 0.
|
Feature: Processor/x86 Status: Internal
|
**Describe the bug**
A clear and concise description of the bug.
**To Reproduce**
Steps to reproduce the behavior:
1. Compile
```c
#include <cstdio>
#include <xmmintrin.h>
int func(const __m128 v)
{
__m128 x = _mm_shuffle_ps(v, v, 0x00);
__m128 y = _mm_shuffle_ps(v, v, 0x55);
__m128 z = _mm_shuffle_ps(v, v, 0xAA);
__m128 s = _mm_add_ps(x, y);
s = _mm_add_ps(s, z);
return _mm_cvtt_ss2si(s);
}
int main(int argc, char** argv)
{
__m128 v = _mm_setr_ps(30 * argc, 0.3 * argc, 0.4 - argc, 0);
int sum = func(v);
printf("Result: %d", sum);
return 0;
}
```
2. Compile with MSVC (any version, at least from 2013 to latest)
3. The disassembly should say something akin to ``_func__cdecl_int_char_ptr("Result: %d",(int)((float)_Argc * 0.3 + (float)(_Argc * 30) + 0.0));``. Note the incorrect ``0.0`` in place of ``0.4 - (float)_Argc``.
**Expected behavior**
It appears that the result of ``SHUFPS XMM#, 0xAA`` gets turned into 0.0, which is very difficult to work with for many executables. Here's a direct example below:
**Screenshots**
<img width="1026" alt="image" src="https://user-images.githubusercontent.com/13488195/210450610-616ff823-c803-4117-8bfc-772d282ef52a.png">
**File:**
- Language ID: x86:LE:64:default (2.13)
- Compiler ID: windows
- Processor: x86
- Endian: Little
- Address Size: 64
**Environment:**
- OS: Windows 11 22621.963
- Java Version: openjdk 19 2022-09-20
- Ghidra Version: 10.2.2
- Ghidra Origin: https://github.com/NationalSecurityAgency/ghidra/issues/releases
|
1.0
|
Decompiler interprets SHUFPS Z values as 0. - **Describe the bug**
A clear and concise description of the bug.
**To Reproduce**
Steps to reproduce the behavior:
1. Compile
```c
#include <cstdio>
#include <xmmintrin.h>
int func(const __m128 v)
{
__m128 x = _mm_shuffle_ps(v, v, 0x00);
__m128 y = _mm_shuffle_ps(v, v, 0x55);
__m128 z = _mm_shuffle_ps(v, v, 0xAA);
__m128 s = _mm_add_ps(x, y);
s = _mm_add_ps(s, z);
return _mm_cvtt_ss2si(s);
}
int main(int argc, char** argv)
{
__m128 v = _mm_setr_ps(30 * argc, 0.3 * argc, 0.4 - argc, 0);
int sum = func(v);
printf("Result: %d", sum);
return 0;
}
```
2. Compile with MSVC (any version, at least from 2013 to latest)
3. The disassembly should say something akin to ``_func__cdecl_int_char_ptr("Result: %d",(int)((float)_Argc * 0.3 + (float)(_Argc * 30) + 0.0));``. Note the incorrect ``0.0`` in place of ``0.4 - (float)_Argc``.
**Expected behavior**
It appears that the result of ``SHUFPS XMM#, 0xAA`` gets turned into 0.0, which is very difficult to work with for many executables. Here's a direct example below:
**Screenshots**
<img width="1026" alt="image" src="https://user-images.githubusercontent.com/13488195/210450610-616ff823-c803-4117-8bfc-772d282ef52a.png">
**File:**
- Language ID: x86:LE:64:default (2.13)
- Compiler ID: windows
- Processor: x86
- Endian: Little
- Address Size: 64
**Environment:**
- OS: Windows 11 22621.963
- Java Version: openjdk 19 2022-09-20
- Ghidra Version: 10.2.2
- Ghidra Origin: https://github.com/NationalSecurityAgency/ghidra/issues/releases
|
process
|
decompiler interprets shufps z values as describe the bug a clear and concise description of the bug to reproduce steps to reproduce the behavior compile c include include int func const v x mm shuffle ps v v y mm shuffle ps v v z mm shuffle ps v v s mm add ps x y s mm add ps s z return mm cvtt s int main int argc char argv v mm setr ps argc argc argc int sum func v printf result d sum return compile with msvc any version at least from to latest the disassembly should say something akin to func cdecl int char ptr result d int float argc float argc note the incorrect in place of float argc expected behavior it appears that the result of shufps xmm gets turned into which is very difficult to work with for many executables here s a direct example below screenshots img width alt image src file language id le default compiler id windows processor endian little address size environment os windows java version openjdk ghidra version ghidra origin
| 1
|
247,535
| 26,712,343,796
|
IssuesEvent
|
2023-01-28 03:14:32
|
NileshGule/cloud-native-ninja
|
https://api.github.com/repos/NileshGule/cloud-native-ninja
|
closed
|
spring-boot-starter-web-3.0.2.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-boot-starter-web-3.0.2.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /src/java/techtalks-producer/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/NileshGule/cloud-native-ninja/commit/c04fe7000230b4a479ad8e4e4d31201f2e3f6536">c04fe7000230b4a479ad8e4e4d31201f2e3f6536</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (spring-boot-starter-web version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-1471](https://www.mend.io/vulnerability-database/CVE-2022-1471) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | snakeyaml-1.33.jar | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-1471</summary>
### Vulnerable Library - <b>snakeyaml-1.33.jar</b></p>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p>
<p>Path to dependency file: /src/java/techtalks-producer/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-3.0.2.jar (Root Library)
- spring-boot-starter-3.0.2.jar
- :x: **snakeyaml-1.33.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/NileshGule/cloud-native-ninja/commit/c04fe7000230b4a479ad8e4e4d31201f2e3f6536">c04fe7000230b4a479ad8e4e4d31201f2e3f6536</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization.
<p>Publish Date: 2022-12-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
spring-boot-starter-web-3.0.2.jar: 1 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-boot-starter-web-3.0.2.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /src/java/techtalks-producer/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/NileshGule/cloud-native-ninja/commit/c04fe7000230b4a479ad8e4e4d31201f2e3f6536">c04fe7000230b4a479ad8e4e4d31201f2e3f6536</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (spring-boot-starter-web version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-1471](https://www.mend.io/vulnerability-database/CVE-2022-1471) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | snakeyaml-1.33.jar | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-1471</summary>
### Vulnerable Library - <b>snakeyaml-1.33.jar</b></p>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="https://bitbucket.org/snakeyaml/snakeyaml">https://bitbucket.org/snakeyaml/snakeyaml</a></p>
<p>Path to dependency file: /src/java/techtalks-producer/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.33/snakeyaml-1.33.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-3.0.2.jar (Root Library)
- spring-boot-starter-3.0.2.jar
- :x: **snakeyaml-1.33.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/NileshGule/cloud-native-ninja/commit/c04fe7000230b4a479ad8e4e4d31201f2e3f6536">c04fe7000230b4a479ad8e4e4d31201f2e3f6536</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization.
<p>Publish Date: 2022-12-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_process
|
spring boot starter web jar vulnerabilities highest severity is autoclosed vulnerable library spring boot starter web jar path to dependency file src java techtalks producer pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in spring boot starter web version remediation available high snakeyaml jar transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file src java techtalks producer pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter web jar root library spring boot starter jar x snakeyaml jar vulnerable library found in head commit a href found in base branch main vulnerability details snakeyaml s constructor class does not restrict types which can be instantiated during deserialization deserializing yaml content provided by an attacker can lead to remote code execution we recommend using snakeyaml s safeconsturctor when parsing untrusted content to restrict deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with mend
| 0
|
290,145
| 32,035,560,198
|
IssuesEvent
|
2023-09-22 15:06:15
|
flatcar/Flatcar
|
https://api.github.com/repos/flatcar/Flatcar
|
closed
|
update: glibc
|
security advisory cvss/HIGH
|
**Name**: glibc
**CVEs**: [CVE-2023-4527](https://nvd.nist.gov/vuln/detail/CVE-2023-4527), [CVE-2023-4806](https://nvd.nist.gov/vuln/detail/CVE-2023-4806)
**CVSSs**: 8.2, n/a
**Action Needed**: update to >= 2.37-r5 or >= 2.38-r2
**Summary**:
* CVE-2023-4527: A flaw was found in glibc. When the getaddrinfo function is called with the AF_UNSPEC address family and the system is configured with no-aaaa mode via /etc/resolv.conf, a DNS response via TCP larger than 2048 bytes can potentially disclose stack contents through the function returned address data, and may cause a crash.
* CVE-2023-4806: A flaw was found in glibc. In an extremely rare situation, the getaddrinfo function may access memory that has been freed, resulting in an application crash. This issue is only exploitable when a NSS module implements only the _nss_*_gethostbyname2_r and _nss_*_getcanonname_r hooks without implementing the _nss_*_gethostbyname3_r hook. The resolved name should return a large number of IPv6 and IPv4, and the call to the getaddrinfo function should have the AF_INET6 address family with AI_CANONNAME, AI_ALL and AI_V4MAPPED as flags.
**refmap.gentoo**: https://bugs.gentoo.org/914281
|
True
|
update: glibc - **Name**: glibc
**CVEs**: [CVE-2023-4527](https://nvd.nist.gov/vuln/detail/CVE-2023-4527), [CVE-2023-4806](https://nvd.nist.gov/vuln/detail/CVE-2023-4806)
**CVSSs**: 8.2, n/a
**Action Needed**: update to >= 2.37-r5 or >= 2.38-r2
**Summary**:
* CVE-2023-4527: A flaw was found in glibc. When the getaddrinfo function is called with the AF_UNSPEC address family and the system is configured with no-aaaa mode via /etc/resolv.conf, a DNS response via TCP larger than 2048 bytes can potentially disclose stack contents through the function returned address data, and may cause a crash.
* CVE-2023-4806: A flaw was found in glibc. In an extremely rare situation, the getaddrinfo function may access memory that has been freed, resulting in an application crash. This issue is only exploitable when a NSS module implements only the _nss_*_gethostbyname2_r and _nss_*_getcanonname_r hooks without implementing the _nss_*_gethostbyname3_r hook. The resolved name should return a large number of IPv6 and IPv4, and the call to the getaddrinfo function should have the AF_INET6 address family with AI_CANONNAME, AI_ALL and AI_V4MAPPED as flags.
**refmap.gentoo**: https://bugs.gentoo.org/914281
|
non_process
|
update glibc name glibc cves cvsss n a action needed update to or summary cve a flaw was found in glibc when the getaddrinfo function is called with the af unspec address family and the system is configured with no aaaa mode via etc resolv conf a dns response via tcp larger than bytes can potentially disclose stack contents through the function returned address data and may cause a crash cve a flaw was found in glibc in an extremely rare situation the getaddrinfo function may access memory that has been freed resulting in an application crash this issue is only exploitable when a nss module implements only the nss r and nss getcanonname r hooks without implementing the nss r hook the resolved name should return a large number of and and the call to the getaddrinfo function should have the af address family with ai canonname ai all and ai as flags refmap gentoo
| 0
|
1,425
| 3,989,485,766
|
IssuesEvent
|
2016-05-09 14:11:37
|
DynareTeam/dynare
|
https://api.github.com/repos/DynareTeam/dynare
|
closed
|
Crash in preprocessor in unstable branch
|
bug preprocessor
|
I get a crash in the preprocessor in the unstable branch, since I rebased my local unstable branch onto the recent commits [pervious preprocessor versions I used date back October 2015 ...] when dealing with our Global Multicountry model.
I get the error both in Windows and MacOS.
In MacOS I get the following message:
``
dynare_m(554,0x7fff7411b300) malloc:
``
``
*** error for object 0x7fda6a60ccd0: pointer being freed was not allocated
``
``
*** set a breakpoint in malloc_error_break to debug
``
apparently what triggers the error is the use of `trend_var` and `deflator` instructions.
When I filter out detrending, pre-processing completes.
I can share a replication package privately.
with older versions of the unstable preprocessor everything went fine.
|
1.0
|
Crash in preprocessor in unstable branch - I get a crash in the preprocessor in the unstable branch, since I rebased my local unstable branch onto the recent commits [pervious preprocessor versions I used date back October 2015 ...] when dealing with our Global Multicountry model.
I get the error both in Windows and MacOS.
In MacOS I get the following message:
``
dynare_m(554,0x7fff7411b300) malloc:
``
``
*** error for object 0x7fda6a60ccd0: pointer being freed was not allocated
``
``
*** set a breakpoint in malloc_error_break to debug
``
apparently what triggers the error is the use of `trend_var` and `deflator` instructions.
When I filter out detrending, pre-processing completes.
I can share a replication package privately.
with older versions of the unstable preprocessor everything went fine.
|
process
|
crash in preprocessor in unstable branch i get a crash in the preprocessor in the unstable branch since i rebased my local unstable branch onto the recent commits when dealing with our global multicountry model i get the error both in windows and macos in macos i get the following message dynare m malloc error for object pointer being freed was not allocated set a breakpoint in malloc error break to debug apparently what triggers the error is the use of trend var and deflator instructions when i filter out detrending pre processing completes i can share a replication package privately with older versions of the unstable preprocessor everything went fine
| 1
|
10,626
| 13,440,696,407
|
IssuesEvent
|
2020-09-08 01:48:01
|
chavarera/python-mini-projects
|
https://api.github.com/repos/chavarera/python-mini-projects
|
closed
|
Convert an image to cartoon usimg opencv
|
Assigned Image-processing Learn Opencv Project boring-stuffs
|
**problem statement**
Convert an image to cartoon usimg opencv
|
1.0
|
Convert an image to cartoon usimg opencv - **problem statement**
Convert an image to cartoon usimg opencv
|
process
|
convert an image to cartoon usimg opencv problem statement convert an image to cartoon usimg opencv
| 1
|
7,017
| 10,168,222,359
|
IssuesEvent
|
2019-08-07 20:15:20
|
brucemiller/LaTeXML
|
https://api.github.com/repos/brucemiller/LaTeXML
|
closed
|
Postprocessing dies on natbib item with "others" in author field
|
bug postprocessing
|
An MWE to reproduce the issue:
main.tex:
```latex
\documentclass{article}
\usepackage{natbib}
\begin{document}
\cite{key}
\bibliography{mybib}
\bibliographystyle{plainnat}
\end{document}
```
mybib.bib:
```bibtex
@article{key,
title={Title},
author={Lastname, Firstname and others},
year={2019}
}
```
<details>
<summary>Details</summary>
```shell
$ latexmlc --format=html5 main.tex
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/TeX.pool.ltxml...
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/eTeX.pool.ltxml... 0.00 sec)
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/pdfTeX.pool.ltxml... 0.00 sec) 0.10 sec)
latexmlc (LaTeXML version 0.8.4)
processing started Sun Jul 28 12:40:55 2019
(Digesting TeX main...
(Processing content /files/sources/natbib/main.tex...
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/LaTeX.pool.ltxml... 0.12 sec)
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/article.cls.ltxml... 0.01 sec)
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/natbib.sty.ltxml... 0.01 sec) 0.15 sec) 0.15 sec)
(Building...
(Loading compiled schema /usr/local/share/perl/5.28.1/LaTeXML/resources/RelaxNG/LaTeXML.model... 0.01 sec). 0.04 sec)
(Rewriting... 0.00 sec)
(Finalizing... 0.00 sec)
Conversion complete: No obvious problems.
(post-processing...
(Scan processing... 0.00 sec)
(MakeBibliography processing... [Converting bibliography /files/sources/natbib/mybib.bib ...
(Digesting BibTeX mybib...
(Processing content /files/sources/natbib/mybib.bib... 0.00 sec)
(Preparsing Bibliography mybib... 0.00 sec)
(Processing content Literal String... 0.02 sec) 0.02 sec)
(Building...
(Loading compiled schema /usr/local/share/perl/5.28.1/LaTeXML/resources/RelaxNG/LaTeXML.model... 0.01 sec). 0.03 sec)
(Rewriting... 0.00 sec)
(Finalizing... 0.00 sec)
... converted!] [using bibliographies mybib.bib] [1 bibentries, 1 cited]
Fatal:perl:die Perl died
Can't call method "cloneNode" on unblessed reference at /usr/local/share/perl/5.28.1/LaTeXML/Post.pm line 1184.
In Post::Document[] ->cloneNode
<= Post::MakeBibliography[@0x56022ad59bc0]
1 fatal error
Post-processing complete: 1 fatal error
```
</details>
Works OK with bibtex.
|
1.0
|
Postprocessing dies on natbib item with "others" in author field - An MWE to reproduce the issue:
main.tex:
```latex
\documentclass{article}
\usepackage{natbib}
\begin{document}
\cite{key}
\bibliography{mybib}
\bibliographystyle{plainnat}
\end{document}
```
mybib.bib:
```bibtex
@article{key,
title={Title},
author={Lastname, Firstname and others},
year={2019}
}
```
<details>
<summary>Details</summary>
```shell
$ latexmlc --format=html5 main.tex
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/TeX.pool.ltxml...
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/eTeX.pool.ltxml... 0.00 sec)
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/pdfTeX.pool.ltxml... 0.00 sec) 0.10 sec)
latexmlc (LaTeXML version 0.8.4)
processing started Sun Jul 28 12:40:55 2019
(Digesting TeX main...
(Processing content /files/sources/natbib/main.tex...
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/LaTeX.pool.ltxml... 0.12 sec)
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/article.cls.ltxml... 0.01 sec)
(Loading /usr/local/share/perl/5.28.1/LaTeXML/Package/natbib.sty.ltxml... 0.01 sec) 0.15 sec) 0.15 sec)
(Building...
(Loading compiled schema /usr/local/share/perl/5.28.1/LaTeXML/resources/RelaxNG/LaTeXML.model... 0.01 sec). 0.04 sec)
(Rewriting... 0.00 sec)
(Finalizing... 0.00 sec)
Conversion complete: No obvious problems.
(post-processing...
(Scan processing... 0.00 sec)
(MakeBibliography processing... [Converting bibliography /files/sources/natbib/mybib.bib ...
(Digesting BibTeX mybib...
(Processing content /files/sources/natbib/mybib.bib... 0.00 sec)
(Preparsing Bibliography mybib... 0.00 sec)
(Processing content Literal String... 0.02 sec) 0.02 sec)
(Building...
(Loading compiled schema /usr/local/share/perl/5.28.1/LaTeXML/resources/RelaxNG/LaTeXML.model... 0.01 sec). 0.03 sec)
(Rewriting... 0.00 sec)
(Finalizing... 0.00 sec)
... converted!] [using bibliographies mybib.bib] [1 bibentries, 1 cited]
Fatal:perl:die Perl died
Can't call method "cloneNode" on unblessed reference at /usr/local/share/perl/5.28.1/LaTeXML/Post.pm line 1184.
In Post::Document[] ->cloneNode
<= Post::MakeBibliography[@0x56022ad59bc0]
1 fatal error
Post-processing complete: 1 fatal error
```
</details>
Works OK with bibtex.
|
process
|
postprocessing dies on natbib item with others in author field an mwe to reproduce the issue main tex latex documentclass article usepackage natbib begin document cite key bibliography mybib bibliographystyle plainnat end document mybib bib bibtex article key title title author lastname firstname and others year details shell latexmlc format main tex loading usr local share perl latexml package tex pool ltxml loading usr local share perl latexml package etex pool ltxml sec loading usr local share perl latexml package pdftex pool ltxml sec sec latexmlc latexml version processing started sun jul digesting tex main processing content files sources natbib main tex loading usr local share perl latexml package latex pool ltxml sec loading usr local share perl latexml package article cls ltxml sec loading usr local share perl latexml package natbib sty ltxml sec sec sec building loading compiled schema usr local share perl latexml resources relaxng latexml model sec sec rewriting sec finalizing sec conversion complete no obvious problems post processing scan processing sec makebibliography processing converting bibliography files sources natbib mybib bib digesting bibtex mybib processing content files sources natbib mybib bib sec preparsing bibliography mybib sec processing content literal string sec sec building loading compiled schema usr local share perl latexml resources relaxng latexml model sec sec rewriting sec finalizing sec converted fatal perl die perl died can t call method clonenode on unblessed reference at usr local share perl latexml post pm line in post document clonenode post makebibliography fatal error post processing complete fatal error works ok with bibtex
| 1
|
17,535
| 23,345,563,584
|
IssuesEvent
|
2022-08-09 17:36:58
|
AdyanRios-NOAA/SEFSC-MH-Processing
|
https://api.github.com/repos/AdyanRios-NOAA/SEFSC-MH-Processing
|
closed
|
Redundancy check improvements
|
enhancement icebox Processing
|
https://github.com/SEFSC-Management-History-Project/SEFSC-MH-Processing/blob/dcd6f518b1ee081472ac2d1a20577386de035c64/MH_process.R#L112-L125
# Notes have been added to explain the redundancy check. However, we can still improve it.
I think that we do not need to create the vector to track change events that later gets merged into the cluster. I think we could just create a new REG_CHANGE variable for the entire mh_prep_use at the start of this process and write directly to that column when we check how many rows are unique after the comparison.
- [ ] Revisit what flags are exceptions to redundancy
- [ ] Try to implement the suggestion above
- [ ] Try changing to a function (we may not need to create blank list at the start!)
- [ ] Check if one-record clusters actually need to be specially treated
|
1.0
|
Redundancy check improvements - https://github.com/SEFSC-Management-History-Project/SEFSC-MH-Processing/blob/dcd6f518b1ee081472ac2d1a20577386de035c64/MH_process.R#L112-L125
# Notes have been added to explain the redundancy check. However, we can still improve it.
I think that we do not need to create the vector to track change events that later gets merged into the cluster. I think we could just create a new REG_CHANGE variable for the entire mh_prep_use at the start of this process and write directly to that column when we check how many rows are unique after the comparison.
- [ ] Revisit what flags are exceptions to redundancy
- [ ] Try to implement the suggestion above
- [ ] Try changing to a function (we may not need to create blank list at the start!)
- [ ] Check if one-record clusters actually need to be specially treated
|
process
|
redundancy check improvements notes have been added to explain the redundancy check however we can still improve it i think that we do not need to create the vector to track change events that later gets merged into the cluster i think we could just create a new reg change variable for the entire mh prep use at the start of this process and write directly to that column when we check how many rows are unique after the comparison revisit what flags are exceptions to redundancy try to implement the suggestion above try changing to a function we may not need to create blank list at the start check if one record clusters actually need to be specially treated
| 1
|
19,779
| 26,162,801,628
|
IssuesEvent
|
2022-12-31 21:09:01
|
RobertCraigie/prisma-client-py
|
https://api.github.com/repos/RobertCraigie/prisma-client-py
|
closed
|
Add support for CockroachDB
|
topic: types kind/feature process/candidate topic: client level/intermediate priority/high
|
## Problem
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
We already have pretty good support for CockroachDB but there is one blind spot, the `push` operation for arrays is not supported.
## Suggested solution
<!-- A clear and concise description of what you want to happen. -->
We need to disable generating this type.
## Additional context
<!-- Add any other context or screenshots about the feature request here. -->
https://github.com/prisma/prisma/issues/13892
|
1.0
|
Add support for CockroachDB - ## Problem
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
We already have pretty good support for CockroachDB but there is one blind spot, the `push` operation for arrays is not supported.
## Suggested solution
<!-- A clear and concise description of what you want to happen. -->
We need to disable generating this type.
## Additional context
<!-- Add any other context or screenshots about the feature request here. -->
https://github.com/prisma/prisma/issues/13892
|
process
|
add support for cockroachdb problem we already have pretty good support for cockroachdb but there is one blind spot the push operation for arrays is not supported suggested solution we need to disable generating this type additional context
| 1
|
158,860
| 20,035,460,204
|
IssuesEvent
|
2022-02-02 11:23:05
|
kapseliboi/crowdfunding-frontend
|
https://api.github.com/repos/kapseliboi/crowdfunding-frontend
|
opened
|
CVE-2020-7793 (High) detected in ua-parser-js-0.7.14.tgz
|
security vulnerability
|
## CVE-2020-7793 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.14.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.14.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.14.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- glamor-2.20.40.tgz (Root Library)
- fbjs-0.8.14.tgz
- :x: **ua-parser-js-0.7.14.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/crowdfunding-frontend/commit/eba5cd0cb6ec1f3de56f69546c8bb139ae4b245d">eba5cd0cb6ec1f3de56f69546c8bb139ae4b245d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info).
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793>CVE-2020-7793</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18">https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution (ua-parser-js): 0.7.23</p>
<p>Direct dependency fix Resolution (glamor): 3.0.0-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7793 (High) detected in ua-parser-js-0.7.14.tgz - ## CVE-2020-7793 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.14.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.14.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.14.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- glamor-2.20.40.tgz (Root Library)
- fbjs-0.8.14.tgz
- :x: **ua-parser-js-0.7.14.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/crowdfunding-frontend/commit/eba5cd0cb6ec1f3de56f69546c8bb139ae4b245d">eba5cd0cb6ec1f3de56f69546c8bb139ae4b245d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info).
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793>CVE-2020-7793</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18">https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution (ua-parser-js): 0.7.23</p>
<p>Direct dependency fix Resolution (glamor): 3.0.0-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in ua parser js tgz cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file package json path to vulnerable library node modules ua parser js package json dependency hierarchy glamor tgz root library fbjs tgz x ua parser js tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos in multiple regexes see linked commit for more info publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ua parser js direct dependency fix resolution glamor step up your open source security game with whitesource
| 0
|
19,902
| 26,357,040,115
|
IssuesEvent
|
2023-01-11 10:28:49
|
tradingstrategy-ai/trading-strategy
|
https://api.github.com/repos/tradingstrategy-ai/trading-strategy
|
opened
|
The version specifier in the code does not match the actual release version
|
process priority: P2 size: S
|
Looking at the [\_\_version\_\_](https://github.com/tradingstrategy-ai/trading-strategy/blob/c2b3cea698c9adb1b0f2d884549b1fdcbe6f8da9/tradingstrategy/__init__.py#L1) marker in the source, it just hardcodes the initial version `0.1.0`, while the actual [latest release](https://pypi.org/project/trading-strategy/) version is `0.9.1` at the time of writing.
The release process needs to be updated, so that the version specifier in the code is properly bumped, and that release metadata is generated from that (in other words, there needs to be a single source of truth).
|
1.0
|
The version specifier in the code does not match the actual release version - Looking at the [\_\_version\_\_](https://github.com/tradingstrategy-ai/trading-strategy/blob/c2b3cea698c9adb1b0f2d884549b1fdcbe6f8da9/tradingstrategy/__init__.py#L1) marker in the source, it just hardcodes the initial version `0.1.0`, while the actual [latest release](https://pypi.org/project/trading-strategy/) version is `0.9.1` at the time of writing.
The release process needs to be updated, so that the version specifier in the code is properly bumped, and that release metadata is generated from that (in other words, there needs to be a single source of truth).
|
process
|
the version specifier in the code does not match the actual release version looking at the marker in the source it just hardcodes the initial version while the actual version is at the time of writing the release process needs to be updated so that the version specifier in the code is properly bumped and that release metadata is generated from that in other words there needs to be a single source of truth
| 1
|
81,768
| 10,256,675,571
|
IssuesEvent
|
2019-08-21 18:14:52
|
ddnexus/pagy
|
https://api.github.com/repos/ddnexus/pagy
|
closed
|
Incorrect documentation on rendering templates?
|
documentation
|
The documentation for rendering a customizable template for the navigation links shows:
```
<%== render 'pagy/nav', locals: {pagy: @pagy} %>
```
This did not work for me and I had to change it to:
```
<%== render partial: 'pagy/nav', locals: {pagy: @pagy} %>
```
The [Rails documentation ](https://guides.rubyonrails.org/layouts_and_rendering.html#partial-layouts) says that 'partial' must be explicitly added whenever there are options passed to the render call, unless I'm missing something.
|
1.0
|
Incorrect documentation on rendering templates? - The documentation for rendering a customizable template for the navigation links shows:
```
<%== render 'pagy/nav', locals: {pagy: @pagy} %>
```
This did not work for me and I had to change it to:
```
<%== render partial: 'pagy/nav', locals: {pagy: @pagy} %>
```
The [Rails documentation ](https://guides.rubyonrails.org/layouts_and_rendering.html#partial-layouts) says that 'partial' must be explicitly added whenever there are options passed to the render call, unless I'm missing something.
|
non_process
|
incorrect documentation on rendering templates the documentation for rendering a customizable template for the navigation links shows this did not work for me and i had to change it to the says that partial must be explicitly added whenever there are options passed to the render call unless i m missing something
| 0
|
594,738
| 18,052,311,070
|
IssuesEvent
|
2021-09-19 23:49:13
|
CraftTweaker/CraftTweaker
|
https://api.github.com/repos/CraftTweaker/CraftTweaker
|
closed
|
[1.15.2] Using .reuse() in CT recipes causes items to duplicate
|
bug Low Priority
|
#### Issue Description:
If you make a crafting table recipe with some items having .reuse() on them then it will cause these items to duplicate themselves when crafting with said item. The .reuse() items must have at least 2 items in the stack for this to happen. I have verified that these are not just ghost items.
#### What happens:
https://imgur.com/a/bhlxk6z
#### What you expected to happen:
I expect the stacked .reuse() items to stay at the same value, not duplicate until they poof.
#### Script used (Please pastebin.com or gist.github.com your script, posting an unpasted or ungist'd script will automatically close this issue. We are **NOT** going to download files, do **NOT** upload your file to this issue, please give us a link to a paste / gist of the file):
https://pastebin.com/ai2r2x2D No other scripts are present.
#### crafttweaker.log file (Please pastebin.com or gist.github.com your file, posting an unpasted or ungist'd file will automatically close this issue. We are **NOT** going to download files, do **NOT** upload your file to this issue, please give us a link to a paste / gist of the file.):
https://pastebin.com/sJEuJ6s9
____
#### Affected Versions (Do *not* use "latest"):
- Minecraft: 1.15.2
- Forge: 31.2.47
- Crafttweaker: CraftTweaker-1.15.2-6.0.0.45
- Using a server: Nope.
- If yes, does the client have the exact same scripts?
#### Your most recent log file where the issue was present (Please pastebin.com or gist.github.com your file, posting an unpasted or ungist'd file will automatically close this issue. We are **NOT** going to download files, do **NOT** upload your file to this issue, please give us a link to a paste / gist of the file.):
https://pastebin.com/kUND66GJ
This bug also happens in 1.16.4, using the exact same script.
- Forge: 35.1.13
- Crafttweaker: CraftTweaker-1.16.4-7.0.0.63
|
1.0
|
[1.15.2] Using .reuse() in CT recipes causes items to duplicate - #### Issue Description:
If you make a crafting table recipe with some items having .reuse() on them then it will cause these items to duplicate themselves when crafting with said item. The .reuse() items must have at least 2 items in the stack for this to happen. I have verified that these are not just ghost items.
#### What happens:
https://imgur.com/a/bhlxk6z
#### What you expected to happen:
I expect the stacked .reuse() items to stay at the same value, not duplicate until they poof.
#### Script used (Please pastebin.com or gist.github.com your script, posting an unpasted or ungist'd script will automatically close this issue. We are **NOT** going to download files, do **NOT** upload your file to this issue, please give us a link to a paste / gist of the file):
https://pastebin.com/ai2r2x2D No other scripts are present.
#### crafttweaker.log file (Please pastebin.com or gist.github.com your file, posting an unpasted or ungist'd file will automatically close this issue. We are **NOT** going to download files, do **NOT** upload your file to this issue, please give us a link to a paste / gist of the file.):
https://pastebin.com/sJEuJ6s9
____
#### Affected Versions (Do *not* use "latest"):
- Minecraft: 1.15.2
- Forge: 31.2.47
- Crafttweaker: CraftTweaker-1.15.2-6.0.0.45
- Using a server: Nope.
- If yes, does the client have the exact same scripts?
#### Your most recent log file where the issue was present (Please pastebin.com or gist.github.com your file, posting an unpasted or ungist'd file will automatically close this issue. We are **NOT** going to download files, do **NOT** upload your file to this issue, please give us a link to a paste / gist of the file.):
https://pastebin.com/kUND66GJ
This bug also happens in 1.16.4, using the exact same script.
- Forge: 35.1.13
- Crafttweaker: CraftTweaker-1.16.4-7.0.0.63
|
non_process
|
using reuse in ct recipes causes items to duplicate issue description if you make a crafting table recipe with some items having reuse on them then it will cause these items to duplicate themselves when crafting with said item the reuse items must have at least items in the stack for this to happen i have verified that these are not just ghost items what happens what you expected to happen i expect the stacked reuse items to stay at the same value not duplicate until they poof script used please pastebin com or gist github com your script posting an unpasted or ungist d script will automatically close this issue we are not going to download files do not upload your file to this issue please give us a link to a paste gist of the file no other scripts are present crafttweaker log file please pastebin com or gist github com your file posting an unpasted or ungist d file will automatically close this issue we are not going to download files do not upload your file to this issue please give us a link to a paste gist of the file affected versions do not use latest minecraft forge crafttweaker crafttweaker using a server nope if yes does the client have the exact same scripts your most recent log file where the issue was present please pastebin com or gist github com your file posting an unpasted or ungist d file will automatically close this issue we are not going to download files do not upload your file to this issue please give us a link to a paste gist of the file this bug also happens in using the exact same script forge crafttweaker crafttweaker
| 0
|
15,858
| 20,034,181,257
|
IssuesEvent
|
2022-02-02 10:06:03
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Allow inline many-to-many relation syntax for mongoDB
|
bug/2-confirmed kind/bug process/candidate topic: schema validation engines/data model parser team/migrations topic: mongodb team/psl-wg
|
There is an untested feature on mongodb that the query engine relies on: embedded inline many-to-many relations:
```prisma
model A {
id String @id @map("_id") @db.ObjectId
gql String?
b_ids String[] @db.Array(ObjectId)
bs B[] @relation(fields: [b_ids])
}
model B {
id String @id @map("_id") @db.ObjectId
gql String?
a_ids String[] @db.Array(ObjectId)
as A[] @relation(fields: [a_ids])
}
```
We improved relation validations recently, and that case that slipped through now triggers a working validation error. Unfortunately, this version of many to many relations on mongo doesn't follow the same pattern and it became impossible to write it the way the Query Engine expects it. We should stop validating that the types of the referencing fields match the referenced fields on MongoDB as a first fix, then validate that the type of the embedded IDs in the array match the type of the id field in the referenced records, maybe as a second pull request. We could assume it is always ObjectId, but I don't have enough mongo knowledge to say if that assumption will hold, so let's do go with the laxer solution first.
|
1.0
|
Allow inline many-to-many relation syntax for mongoDB - There is an untested feature on mongodb that the query engine relies on: embedded inline many-to-many relations:
```prisma
model A {
id String @id @map("_id") @db.ObjectId
gql String?
b_ids String[] @db.Array(ObjectId)
bs B[] @relation(fields: [b_ids])
}
model B {
id String @id @map("_id") @db.ObjectId
gql String?
a_ids String[] @db.Array(ObjectId)
as A[] @relation(fields: [a_ids])
}
```
We improved relation validations recently, and that case that slipped through now triggers a working validation error. Unfortunately, this version of many to many relations on mongo doesn't follow the same pattern and it became impossible to write it the way the Query Engine expects it. We should stop validating that the types of the referencing fields match the referenced fields on MongoDB as a first fix, then validate that the type of the embedded IDs in the array match the type of the id field in the referenced records, maybe as a second pull request. We could assume it is always ObjectId, but I don't have enough mongo knowledge to say if that assumption will hold, so let's do go with the laxer solution first.
|
process
|
allow inline many to many relation syntax for mongodb there is an untested feature on mongodb that the query engine relies on embedded inline many to many relations prisma model a id string id map id db objectid gql string b ids string db array objectid bs b relation fields model b id string id map id db objectid gql string a ids string db array objectid as a relation fields we improved relation validations recently and that case that slipped through now triggers a working validation error unfortunately this version of many to many relations on mongo doesn t follow the same pattern and it became impossible to write it the way the query engine expects it we should stop validating that the types of the referencing fields match the referenced fields on mongodb as a first fix then validate that the type of the embedded ids in the array match the type of the id field in the referenced records maybe as a second pull request we could assume it is always objectid but i don t have enough mongo knowledge to say if that assumption will hold so let s do go with the laxer solution first
| 1
|
6,726
| 9,829,285,375
|
IssuesEvent
|
2019-06-15 19:17:59
|
ION28/BLUESPAWN
|
https://api.github.com/repos/ION28/BLUESPAWN
|
opened
|
Investigate subscribing to PsSetCreateProcessNotifyRoutine
|
enhancement hard other os processes
|
Pssetcreateprocessnotifyroutine for managing process creation and blocking new procs
|
1.0
|
Investigate subscribing to PsSetCreateProcessNotifyRoutine - Pssetcreateprocessnotifyroutine for managing process creation and blocking new procs
|
process
|
investigate subscribing to pssetcreateprocessnotifyroutine pssetcreateprocessnotifyroutine for managing process creation and blocking new procs
| 1
|
10,350
| 13,175,401,165
|
IssuesEvent
|
2020-08-12 01:30:25
|
googleapis/python-spanner
|
https://api.github.com/repos/googleapis/python-spanner
|
closed
|
docs-presubmit failing
|
api: spanner priority: p2 type: process
|
```
[18:13:16][ERROR] Failed to get build config
com.google.devtools.kokoro.config.ConfigException: Couldn't find build configuration file docs-presubmit.cfg or docs-presubmit.gcl under /tmp/workspace/workspace/cloud-devrel/client-libraries/python/googleapis/python-spanner/docs/docs-presubmit/src/github/python-spanner/.kokoro/docs.
at com.google.devtools.kokoro.config.BuildConfigReader.lambda$read$2(BuildConfigReader.java:54)
at java.util.Optional.orElseThrow(Optional.java:290)
at com.google.devtools.kokoro.config.BuildConfigReader.read(BuildConfigReader.java:51)
at com.google.devtools.kokoro.jenkins.plugin.kokorojob.store.NodeBuildConfigReader.invoke(NodeBuildConfigReader.java:39)
at com.google.devtools.kokoro.jenkins.plugin.kokorojob.store.NodeBuildConfigReader.invoke(NodeBuildConfigReader.java:13)
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2731)
at hudson.remoting.UserRequest.perform(UserRequest.java:153)
at hudson.remoting.UserRequest.perform(UserRequest.java:50)
at hudson.remoting.Request$2.run(Request.java:336)
at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:68)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at org.jenkinsci.remoting.kokoro.RpcSlaveEngine$1$1.run(RpcSlaveEngine.java:107)
at java.lang.Thread.run(Thread.java:748)
at ......remote call to gcp_ubuntu-prod-yoshi-ubuntu-ir-819542672(Native Method)
at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1537)
at hudson.remoting.UserResponse.retrieve(UserRequest.java:253)
at hudson.remoting.Channel.call(Channel.java:822)
at hudson.FilePath.act(FilePath.java:985)
at hudson.FilePath.act(FilePath.java:974)
at com.google.devtools.kokoro.jenkins.plugin.kokorojob.store.ConfigStore.getKokoroBuildConfig(ConfigStore.java:102)
at com.google.devtools.kokoro.jenkins.plugin.pipeline.KokoroFlowExecution.getBuildConfig(KokoroFlowExecution.java:661)
at com.google.devtools.kokoro.jenkins.plugin.pipeline.KokoroFlowExecution.addPostScmSteps(KokoroFlowExecution.java:608)
at com.google.devtools.kokoro.jenkins.plugin.pipeline.KokoroScmStepContext.onSuccess(KokoroScmStepContext.java:25)
at org.jenkinsci.plugins.workflow.steps.AbstractSynchronousNonBlockingStepExecution$1.run(AbstractSynchronousNonBlockingStepExecution.java:44)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
```
|
1.0
|
docs-presubmit failing - ```
[18:13:16][ERROR] Failed to get build config
com.google.devtools.kokoro.config.ConfigException: Couldn't find build configuration file docs-presubmit.cfg or docs-presubmit.gcl under /tmp/workspace/workspace/cloud-devrel/client-libraries/python/googleapis/python-spanner/docs/docs-presubmit/src/github/python-spanner/.kokoro/docs.
at com.google.devtools.kokoro.config.BuildConfigReader.lambda$read$2(BuildConfigReader.java:54)
at java.util.Optional.orElseThrow(Optional.java:290)
at com.google.devtools.kokoro.config.BuildConfigReader.read(BuildConfigReader.java:51)
at com.google.devtools.kokoro.jenkins.plugin.kokorojob.store.NodeBuildConfigReader.invoke(NodeBuildConfigReader.java:39)
at com.google.devtools.kokoro.jenkins.plugin.kokorojob.store.NodeBuildConfigReader.invoke(NodeBuildConfigReader.java:13)
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2731)
at hudson.remoting.UserRequest.perform(UserRequest.java:153)
at hudson.remoting.UserRequest.perform(UserRequest.java:50)
at hudson.remoting.Request$2.run(Request.java:336)
at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:68)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at org.jenkinsci.remoting.kokoro.RpcSlaveEngine$1$1.run(RpcSlaveEngine.java:107)
at java.lang.Thread.run(Thread.java:748)
at ......remote call to gcp_ubuntu-prod-yoshi-ubuntu-ir-819542672(Native Method)
at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1537)
at hudson.remoting.UserResponse.retrieve(UserRequest.java:253)
at hudson.remoting.Channel.call(Channel.java:822)
at hudson.FilePath.act(FilePath.java:985)
at hudson.FilePath.act(FilePath.java:974)
at com.google.devtools.kokoro.jenkins.plugin.kokorojob.store.ConfigStore.getKokoroBuildConfig(ConfigStore.java:102)
at com.google.devtools.kokoro.jenkins.plugin.pipeline.KokoroFlowExecution.getBuildConfig(KokoroFlowExecution.java:661)
at com.google.devtools.kokoro.jenkins.plugin.pipeline.KokoroFlowExecution.addPostScmSteps(KokoroFlowExecution.java:608)
at com.google.devtools.kokoro.jenkins.plugin.pipeline.KokoroScmStepContext.onSuccess(KokoroScmStepContext.java:25)
at org.jenkinsci.plugins.workflow.steps.AbstractSynchronousNonBlockingStepExecution$1.run(AbstractSynchronousNonBlockingStepExecution.java:44)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
```
|
process
|
docs presubmit failing failed to get build config com google devtools kokoro config configexception couldn t find build configuration file docs presubmit cfg or docs presubmit gcl under tmp workspace workspace cloud devrel client libraries python googleapis python spanner docs docs presubmit src github python spanner kokoro docs at com google devtools kokoro config buildconfigreader lambda read buildconfigreader java at java util optional orelsethrow optional java at com google devtools kokoro config buildconfigreader read buildconfigreader java at com google devtools kokoro jenkins plugin kokorojob store nodebuildconfigreader invoke nodebuildconfigreader java at com google devtools kokoro jenkins plugin kokorojob store nodebuildconfigreader invoke nodebuildconfigreader java at hudson filepath filecallablewrapper call filepath java at hudson remoting userrequest perform userrequest java at hudson remoting userrequest perform userrequest java at hudson remoting request run request java at hudson remoting interceptingexecutorservice call interceptingexecutorservice java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org jenkinsci remoting kokoro rpcslaveengine run rpcslaveengine java at java lang thread run thread java at remote call to gcp ubuntu prod yoshi ubuntu ir native method at hudson remoting channel attachcallsitestacktrace channel java at hudson remoting userresponse retrieve userrequest java at hudson remoting channel call channel java at hudson filepath act filepath java at hudson filepath act filepath java at com google devtools kokoro jenkins plugin kokorojob store configstore getkokorobuildconfig configstore java at com google devtools kokoro jenkins plugin pipeline kokoroflowexecution getbuildconfig kokoroflowexecution java at com google devtools kokoro jenkins plugin pipeline kokoroflowexecution addpostscmsteps kokoroflowexecution java at com google devtools kokoro jenkins plugin pipeline kokoroscmstepcontext onsuccess kokoroscmstepcontext java at org jenkinsci plugins workflow steps abstractsynchronousnonblockingstepexecution run abstractsynchronousnonblockingstepexecution java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
| 1
|
123,794
| 12,218,562,484
|
IssuesEvent
|
2020-05-01 19:37:42
|
koala-il/origame-backend
|
https://api.github.com/repos/koala-il/origame-backend
|
opened
|
Plan the UI/UX of the application
|
documentation
|
How does the User see customers?
How to save customers?
How to register?
How to set events on the schedule?
|
1.0
|
Plan the UI/UX of the application - How does the User see customers?
How to save customers?
How to register?
How to set events on the schedule?
|
non_process
|
plan the ui ux of the application how does the user see customers how to save customers how to register how to set events on the schedule
| 0
|
279,286
| 24,211,726,314
|
IssuesEvent
|
2022-09-26 00:04:20
|
envoyproxy/envoy
|
https://api.github.com/repos/envoyproxy/envoy
|
closed
|
ProtocolIntegrationTest.HandleUpstreamSocketFail is flaky
|
bug area/test flakes stale
|
The test is flaky in various ways:
Fails at:
```
EXPECT_THAT(waitForAccessLog(access_log_name_),
HasSubstr("upstream_reset_before_response_started{connection_termination}"));
```
with corresponding log observed:
[2022-08-12 18:49:00.193][14][critical][assert] [test/integration/base_integration_test.cc:502] assert failure: 0. Details: Timed out waiting for access log. Found:
[2022-08-12 18:49:00.193][14][critical][backtrace] [./source/server/backtrace.h:104] Caught Aborted, suspect faulting address 0x50e8d0000000e
[2022-08-12 18:49:00.193][14][critical][backtrace] [./source/server/backtrace.h:91] Backtrace (use tools/stack_decode.py to get line numbers):
[2022-08-12 18:49:00.193][14][critical][backtrace] [./source/server/backtrace.h:92] Envoy version: 0/1.24.0-dev/test/DEBUG/BoringSSL
[2022-08-12 18:49:00.217][14][critical][backtrace] [./source/server/backtrace.h:96] #0: Envoy::SignalAction::sigHandler() [0x46fab78]
[2022-08-12 18:49:00.217][14][critical][backtrace] [./source/server/backtrace.h:96] #1: __restore_rt [0x7f06aad14200]
[2022-08-12 18:49:00.237][14][critical][backtrace] [./source/server/backtrace.h:96] #2: Envoy::ProtocolIntegrationTest_HandleUpstreamSocketFail_Test::TestBody() [0x1ea0757]
[2022-08-12 18:49:00.257][14][critical][backtrace] [./source/server/backtrace.h:96] #3: testing::internal::HandleSehExceptionsInMethodIfSupported<>() [0x53cfc4b]
[2022-08-12 18:49:00.277][14][critical][backtrace] [./source/server/backtrace.h:96] #4: testing::internal::HandleExceptionsInMethodIfSupported<>() [0x53c05bd]
[2022-08-12 18:49:00.296][14][critical][backtrace] [./source/server/backtrace.h:96] #5: testing::Test::Run() [0x53a8e93]
[2022-08-12 18:49:00.315][14][critical][backtrace] [./source/server/backtrace.h:96] #6: testing::TestInfo::Run() [0x53a9a5a]
[2022-08-12 18:49:00.334][14][critical][backtrace] [./source/server/backtrace.h:96] #7: testing::TestSuite::Run() [0x53aa2ab]
And it also fails with errors related to singleton::get() before initialize().
[ RUN ] UpstreamProtocols/ProtocolIntegrationTest.HandleUpstreamSocketFail/IPv6_Http2Downstream_Http3UpstreamWrappedNghttp2WithDeferredProcessing
[2022-08-19 16:05:12.336][2862][error][quic] [external/com_github_google_quiche/quiche/quic/core/quic_connection.cc:3899] Client: Write failed with error: 9 (Bad file descriptor)
[2022-08-19 16:05:12.370][2854][critical][assert] [./source/common/singleton/threadsafe_singleton.h:56] assert failure: loader_ != nullptr. Details: InjectableSingleton used prior to initialization
[2022-08-19 16:05:12.370][2854][critical][backtrace] [./source/server/backtrace.h:104] Caught Aborted, suspect faulting address 0x50e8d0000000e
[2022-08-19 16:05:12.370][2854][critical][backtrace] [./source/server/backtrace.h:91] Backtrace (use tools/stack_decode.py to get line numbers):
[2022-08-19 16:05:12.370][2854][critical][backtrace] [./source/server/backtrace.h:92] Envoy version: 0/1.24.0-dev/test/DEBUG/BoringSSL
[2022-08-19 16:05:12.394][2854][critical][backtrace] [./source/server/backtrace.h:96] #0: Envoy::SignalAction::sigHandler() [0x471f468]
[2022-08-19 16:05:12.394][2854][critical][backtrace] [./source/server/backtrace.h:96] #1: __restore_rt [0x7fdb42e67200]
[2022-08-19 16:05:12.426][2854][critical][backtrace] [./source/server/backtrace.h:96] #2: Envoy::Network::Address::(anonymous namespace)::sockInterfaceOrDefault() [0x46b291c]
[2022-08-19 16:05:12.454][2854][critical][backtrace] [./source/server/backtrace.h:96] #3: Envoy::Network::Address::Ipv6Instance::Ipv6Instance() [0x46b5a40]
[2022-08-19 16:05:12.488][2854][critical][backtrace] [./source/server/backtrace.h:96] #4: Envoy::Network::Address::InstanceFactory::createInstancePtr<>() [0x46b7765]
[2022-08-19 16:05:12.516][2854][critical][backtrace] [./source/server/backtrace.h:96] #5: Envoy::Network::Address::addressFromSockAddr() [0x46b1056]
[2022-08-19 16:05:12.550][2854][critical][backtrace] [./source/server/backtrace.h:96] #6: Envoy::Network::Address::addressFromSockAddrOrDie() [0x46b1ef8]
[2022-08-19 16:05:12.578][2854][critical][backtrace] [./source/server/backtrace.h:96] #7: Envoy::Quic::quicAddressToEnvoyAddressInstance() [0x3fa1dc2]
[2022-08-19 16:05:12.611][2854][critical][backtrace] [./source/server/backtrace.h:96] #8: Envoy::Quic::EnvoyQuicPacketWriter::GetNextWriteLocation() [0x3ddf037]
[2022-08-19 16:05:12.637][2854][critical][backtrace] [./source/server/backtrace.h:96] #9: quic::QuicConnection::GetPacketBuffer() [0x416afa9]
[2022-08-19 16:05:12.662][2854][critical][backtrace] [./source/server/backtrace.h:96] #10: q
repro: $bazel test test/integration:quic_protocol_integration_test --test_filter="*HandleUpstreamSocketFail/IPv6_Http2Downstream_Http3UpstreamWrapp*" --test_arg="-l trace" --runs_per_test=5000
|
1.0
|
ProtocolIntegrationTest.HandleUpstreamSocketFail is flaky - The test is flaky in various ways:
Fails at:
```
EXPECT_THAT(waitForAccessLog(access_log_name_),
HasSubstr("upstream_reset_before_response_started{connection_termination}"));
```
with corresponding log observed:
[2022-08-12 18:49:00.193][14][critical][assert] [test/integration/base_integration_test.cc:502] assert failure: 0. Details: Timed out waiting for access log. Found:
[2022-08-12 18:49:00.193][14][critical][backtrace] [./source/server/backtrace.h:104] Caught Aborted, suspect faulting address 0x50e8d0000000e
[2022-08-12 18:49:00.193][14][critical][backtrace] [./source/server/backtrace.h:91] Backtrace (use tools/stack_decode.py to get line numbers):
[2022-08-12 18:49:00.193][14][critical][backtrace] [./source/server/backtrace.h:92] Envoy version: 0/1.24.0-dev/test/DEBUG/BoringSSL
[2022-08-12 18:49:00.217][14][critical][backtrace] [./source/server/backtrace.h:96] #0: Envoy::SignalAction::sigHandler() [0x46fab78]
[2022-08-12 18:49:00.217][14][critical][backtrace] [./source/server/backtrace.h:96] #1: __restore_rt [0x7f06aad14200]
[2022-08-12 18:49:00.237][14][critical][backtrace] [./source/server/backtrace.h:96] #2: Envoy::ProtocolIntegrationTest_HandleUpstreamSocketFail_Test::TestBody() [0x1ea0757]
[2022-08-12 18:49:00.257][14][critical][backtrace] [./source/server/backtrace.h:96] #3: testing::internal::HandleSehExceptionsInMethodIfSupported<>() [0x53cfc4b]
[2022-08-12 18:49:00.277][14][critical][backtrace] [./source/server/backtrace.h:96] #4: testing::internal::HandleExceptionsInMethodIfSupported<>() [0x53c05bd]
[2022-08-12 18:49:00.296][14][critical][backtrace] [./source/server/backtrace.h:96] #5: testing::Test::Run() [0x53a8e93]
[2022-08-12 18:49:00.315][14][critical][backtrace] [./source/server/backtrace.h:96] #6: testing::TestInfo::Run() [0x53a9a5a]
[2022-08-12 18:49:00.334][14][critical][backtrace] [./source/server/backtrace.h:96] #7: testing::TestSuite::Run() [0x53aa2ab]
And it also fails with errors related to singleton::get() before initialize().
[ RUN ] UpstreamProtocols/ProtocolIntegrationTest.HandleUpstreamSocketFail/IPv6_Http2Downstream_Http3UpstreamWrappedNghttp2WithDeferredProcessing
[2022-08-19 16:05:12.336][2862][error][quic] [external/com_github_google_quiche/quiche/quic/core/quic_connection.cc:3899] Client: Write failed with error: 9 (Bad file descriptor)
[2022-08-19 16:05:12.370][2854][critical][assert] [./source/common/singleton/threadsafe_singleton.h:56] assert failure: loader_ != nullptr. Details: InjectableSingleton used prior to initialization
[2022-08-19 16:05:12.370][2854][critical][backtrace] [./source/server/backtrace.h:104] Caught Aborted, suspect faulting address 0x50e8d0000000e
[2022-08-19 16:05:12.370][2854][critical][backtrace] [./source/server/backtrace.h:91] Backtrace (use tools/stack_decode.py to get line numbers):
[2022-08-19 16:05:12.370][2854][critical][backtrace] [./source/server/backtrace.h:92] Envoy version: 0/1.24.0-dev/test/DEBUG/BoringSSL
[2022-08-19 16:05:12.394][2854][critical][backtrace] [./source/server/backtrace.h:96] #0: Envoy::SignalAction::sigHandler() [0x471f468]
[2022-08-19 16:05:12.394][2854][critical][backtrace] [./source/server/backtrace.h:96] #1: __restore_rt [0x7fdb42e67200]
[2022-08-19 16:05:12.426][2854][critical][backtrace] [./source/server/backtrace.h:96] #2: Envoy::Network::Address::(anonymous namespace)::sockInterfaceOrDefault() [0x46b291c]
[2022-08-19 16:05:12.454][2854][critical][backtrace] [./source/server/backtrace.h:96] #3: Envoy::Network::Address::Ipv6Instance::Ipv6Instance() [0x46b5a40]
[2022-08-19 16:05:12.488][2854][critical][backtrace] [./source/server/backtrace.h:96] #4: Envoy::Network::Address::InstanceFactory::createInstancePtr<>() [0x46b7765]
[2022-08-19 16:05:12.516][2854][critical][backtrace] [./source/server/backtrace.h:96] #5: Envoy::Network::Address::addressFromSockAddr() [0x46b1056]
[2022-08-19 16:05:12.550][2854][critical][backtrace] [./source/server/backtrace.h:96] #6: Envoy::Network::Address::addressFromSockAddrOrDie() [0x46b1ef8]
[2022-08-19 16:05:12.578][2854][critical][backtrace] [./source/server/backtrace.h:96] #7: Envoy::Quic::quicAddressToEnvoyAddressInstance() [0x3fa1dc2]
[2022-08-19 16:05:12.611][2854][critical][backtrace] [./source/server/backtrace.h:96] #8: Envoy::Quic::EnvoyQuicPacketWriter::GetNextWriteLocation() [0x3ddf037]
[2022-08-19 16:05:12.637][2854][critical][backtrace] [./source/server/backtrace.h:96] #9: quic::QuicConnection::GetPacketBuffer() [0x416afa9]
[2022-08-19 16:05:12.662][2854][critical][backtrace] [./source/server/backtrace.h:96] #10: q
repro: $bazel test test/integration:quic_protocol_integration_test --test_filter="*HandleUpstreamSocketFail/IPv6_Http2Downstream_Http3UpstreamWrapp*" --test_arg="-l trace" --runs_per_test=5000
|
non_process
|
protocolintegrationtest handleupstreamsocketfail is flaky the test is flaky in various ways fails at expect that waitforaccesslog access log name hassubstr upstream reset before response started connection termination with corresponding log observed assert failure details timed out waiting for access log found caught aborted suspect faulting address backtrace use tools stack decode py to get line numbers envoy version dev test debug boringssl envoy signalaction sighandler restore rt envoy protocolintegrationtest handleupstreamsocketfail test testbody testing internal handlesehexceptionsinmethodifsupported testing internal handleexceptionsinmethodifsupported testing test run testing testinfo run testing testsuite run and it also fails with errors related to singleton get before initialize upstreamprotocols protocolintegrationtest handleupstreamsocketfail client write failed with error bad file descriptor assert failure loader nullptr details injectablesingleton used prior to initialization caught aborted suspect faulting address backtrace use tools stack decode py to get line numbers envoy version dev test debug boringssl envoy signalaction sighandler restore rt envoy network address anonymous namespace sockinterfaceordefault envoy network address envoy network address instancefactory createinstanceptr envoy network address addressfromsockaddr envoy network address addressfromsockaddrordie envoy quic quicaddresstoenvoyaddressinstance envoy quic envoyquicpacketwriter getnextwritelocation quic quicconnection getpacketbuffer q repro bazel test test integration quic protocol integration test test filter handleupstreamsocketfail test arg l trace runs per test
| 0
|
9,946
| 12,976,232,683
|
IssuesEvent
|
2020-07-21 18:24:20
|
obinnaokechukwu/internship-2020
|
https://api.github.com/repos/obinnaokechukwu/internship-2020
|
opened
|
Update OBS API documentation with info about newly added APIs
|
process
|
Update OBS API documentation with info about newly added APIs
|
1.0
|
Update OBS API documentation with info about newly added APIs - Update OBS API documentation with info about newly added APIs
|
process
|
update obs api documentation with info about newly added apis update obs api documentation with info about newly added apis
| 1
|
15,742
| 19,910,538,565
|
IssuesEvent
|
2022-01-25 16:44:12
|
input-output-hk/high-assurance-legacy
|
https://api.github.com/repos/input-output-hk/high-assurance-legacy
|
closed
|
Formally prove `bidirectional_bridge` core lemmas
|
type: enhancement language: isabelle topic: process calculus
|
Our goal is to formally prove the `bidirectional_bridge` core lemmas described in #35.
Informal proofs of these lemmas are given in https://github.com/input-output-hk/fm-ouroboros/issues/15#issuecomment-486381795.
|
1.0
|
Formally prove `bidirectional_bridge` core lemmas - Our goal is to formally prove the `bidirectional_bridge` core lemmas described in #35.
Informal proofs of these lemmas are given in https://github.com/input-output-hk/fm-ouroboros/issues/15#issuecomment-486381795.
|
process
|
formally prove bidirectional bridge core lemmas our goal is to formally prove the bidirectional bridge core lemmas described in informal proofs of these lemmas are given in
| 1
|
285,930
| 21,558,595,107
|
IssuesEvent
|
2022-04-30 21:06:14
|
magidoc-org/magidoc
|
https://api.github.com/repos/magidoc-org/magidoc
|
closed
|
Update documentation with SDL introspection and Rollup-parse-graphql-schema
|
documentation
|
Update documentation with SDL introspection and Rollup-parse-graphql-schema. See https://github.com/magidoc-org/magidoc/pull/40
|
1.0
|
Update documentation with SDL introspection and Rollup-parse-graphql-schema - Update documentation with SDL introspection and Rollup-parse-graphql-schema. See https://github.com/magidoc-org/magidoc/pull/40
|
non_process
|
update documentation with sdl introspection and rollup parse graphql schema update documentation with sdl introspection and rollup parse graphql schema see
| 0
|
552,547
| 16,243,319,885
|
IssuesEvent
|
2021-05-07 12:11:20
|
aiidateam/aiida-quantumespresso
|
https://api.github.com/repos/aiidateam/aiida-quantumespresso
|
opened
|
`PwBandStructureWorkChain`: Deprecate and fix issues
|
priority/important topic/workflows type/bug type/refactoring
|
The `PwBandStructureWorkChain` is a legacy version of a band structure work chain with a pre-defined protocol, mostly used for previous tutorials. However, now that the `PwBandsWorkChain` has been equipped with a protocol through the `get_builder_from_protocol()` method, it no longer serves a purpose and should be deprecated.
However, we should still introduce a fix for the issues people have been experiencing when trying to run this work chain, so it's still functional until v4.0.0 is released and the work chain is removed. Two issues reported so far:
* No support for `aiida-pseudo` pseudos.
* An issue @atztogo ran into recently related to the `calculation` tag being missing:
```Python
File "/home/max/.virtualenvs/aiida/lib/python3.7/site-packages/plumpy/process_states.py",
line 230, in execute
result = self.run_fn(*self.args, **self.kwargs)
File "/home/max/codes/aiida-core/aiida/engine/processes/workchains/workchain.py",
line 214, in _do_step
finished, stepper_result = self._stepper.step()
File "/home/max/.virtualenvs/aiida/lib/python3.7/site-packages/plumpy/workchains.py",
line 299, in step
finished, result = self._child_stepper.step()
File "/home/max/.virtualenvs/aiida/lib/python3.7/site-packages/plumpy/workchains.py",
line 250, in step
return True, self._fn(self._workchain)
File "/home/max/.virtualenvs/aiida/lib/python3.7/site-packages/aiida_quantumespresso/workflows/pw/relax.py",
line 362, in results
if self.inputs.base.pw.parameters['CONTROL']['calculation'] != 'scf':
KeyError: 'calculation'
```
|
1.0
|
`PwBandStructureWorkChain`: Deprecate and fix issues - The `PwBandStructureWorkChain` is a legacy version of a band structure work chain with a pre-defined protocol, mostly used for previous tutorials. However, now that the `PwBandsWorkChain` has been equipped with a protocol through the `get_builder_from_protocol()` method, it no longer serves a purpose and should be deprecated.
However, we should still introduce a fix for the issues people have been experiencing when trying to run this work chain, so it's still functional until v4.0.0 is released and the work chain is removed. Two issues reported so far:
* No support for `aiida-pseudo` pseudos.
* An issue @atztogo ran into recently related to the `calculation` tag being missing:
```Python
File "/home/max/.virtualenvs/aiida/lib/python3.7/site-packages/plumpy/process_states.py",
line 230, in execute
result = self.run_fn(*self.args, **self.kwargs)
File "/home/max/codes/aiida-core/aiida/engine/processes/workchains/workchain.py",
line 214, in _do_step
finished, stepper_result = self._stepper.step()
File "/home/max/.virtualenvs/aiida/lib/python3.7/site-packages/plumpy/workchains.py",
line 299, in step
finished, result = self._child_stepper.step()
File "/home/max/.virtualenvs/aiida/lib/python3.7/site-packages/plumpy/workchains.py",
line 250, in step
return True, self._fn(self._workchain)
File "/home/max/.virtualenvs/aiida/lib/python3.7/site-packages/aiida_quantumespresso/workflows/pw/relax.py",
line 362, in results
if self.inputs.base.pw.parameters['CONTROL']['calculation'] != 'scf':
KeyError: 'calculation'
```
|
non_process
|
pwbandstructureworkchain deprecate and fix issues the pwbandstructureworkchain is a legacy version of a band structure work chain with a pre defined protocol mostly used for previous tutorials however now that the pwbandsworkchain has been equipped with a protocol through the get builder from protocol method it no longer serves a purpose and should be deprecated however we should still introduce a fix for the issues people have been experiencing when trying to run this work chain so it s still functional until is released and the work chain is removed two issues reported so far no support for aiida pseudo pseudos an issue atztogo ran into recently related to the calculation tag being missing python file home max virtualenvs aiida lib site packages plumpy process states py line in execute result self run fn self args self kwargs file home max codes aiida core aiida engine processes workchains workchain py line in do step finished stepper result self stepper step file home max virtualenvs aiida lib site packages plumpy workchains py line in step finished result self child stepper step file home max virtualenvs aiida lib site packages plumpy workchains py line in step return true self fn self workchain file home max virtualenvs aiida lib site packages aiida quantumespresso workflows pw relax py line in results if self inputs base pw parameters scf keyerror calculation
| 0
|
6,865
| 9,998,709,255
|
IssuesEvent
|
2019-07-12 08:52:18
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Changing ZenHub pipeline should change label on the GitHub issue
|
process: contributing stage: needs investigating type: chore
|
If we start working on the issue and move it in the pipeline
<img width="440" alt="screen shot 2019-01-23 at 4 03 05 pm" src="https://user-images.githubusercontent.com/2212006/51636950-7abefa80-1f28-11e9-8cef-c70d1654f1e2.png">
it should change the "state: ..." label on that GitHub issue.
|
1.0
|
Changing ZenHub pipeline should change label on the GitHub issue - If we start working on the issue and move it in the pipeline
<img width="440" alt="screen shot 2019-01-23 at 4 03 05 pm" src="https://user-images.githubusercontent.com/2212006/51636950-7abefa80-1f28-11e9-8cef-c70d1654f1e2.png">
it should change the "state: ..." label on that GitHub issue.
|
process
|
changing zenhub pipeline should change label on the github issue if we start working on the issue and move it in the pipeline img width alt screen shot at pm src it should change the state label on that github issue
| 1
|
22,018
| 30,523,569,569
|
IssuesEvent
|
2023-07-19 09:41:14
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
@mitech/onit-cli 3.0.1 has 2 guarddog issues
|
npm-silent-process-execution
|
```{"npm-silent-process-execution":[{"code":" const subprocess = (0, child_process_1.spawn)(pm2exec, ['stop', 'all'], {\n detached: true,\n stdio: 'ignore'\n });","location":"package/dist/bin/serve/_versions/2.0.0/lib/pm2.js:67","message":"This package is silently executing another executable"},{"code":" const subprocess = _spawn(pm2exec, ['stop', 'all'], {\n detached: true,\n stdio: 'ignore'\n });","location":"package/src/bin/serve/_versions/2.0.0/lib/pm2.ts:66","message":"This package is silently executing another executable"}]}```
|
1.0
|
@mitech/onit-cli 3.0.1 has 2 guarddog issues - ```{"npm-silent-process-execution":[{"code":" const subprocess = (0, child_process_1.spawn)(pm2exec, ['stop', 'all'], {\n detached: true,\n stdio: 'ignore'\n });","location":"package/dist/bin/serve/_versions/2.0.0/lib/pm2.js:67","message":"This package is silently executing another executable"},{"code":" const subprocess = _spawn(pm2exec, ['stop', 'all'], {\n detached: true,\n stdio: 'ignore'\n });","location":"package/src/bin/serve/_versions/2.0.0/lib/pm2.ts:66","message":"This package is silently executing another executable"}]}```
|
process
|
mitech onit cli has guarddog issues npm silent process execution n detached true n stdio ignore n location package dist bin serve versions lib js message this package is silently executing another executable code const subprocess spawn n detached true n stdio ignore n location package src bin serve versions lib ts message this package is silently executing another executable
| 1
|
51,181
| 21,582,106,579
|
IssuesEvent
|
2022-05-02 19:55:32
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
Create Prototype "My Work Order Details" page
|
Workgroup: AMD Service: Apps Type: Enhancement Product: AMD Data Tracker
|
Created a new "My Work Orders - Prototype" table on the `my-work-orders` page
- Hidden to others that do not have the "System Administrator" role

|
1.0
|
Create Prototype "My Work Order Details" page - Created a new "My Work Orders - Prototype" table on the `my-work-orders` page
- Hidden to others that do not have the "System Administrator" role

|
non_process
|
create prototype my work order details page created a new my work orders prototype table on the my work orders page hidden to others that do not have the system administrator role
| 0
|
420,701
| 28,293,184,910
|
IssuesEvent
|
2023-04-09 13:19:24
|
AY2223S2-CS2113-W15-2/tp
|
https://api.github.com/repos/AY2223S2-CS2113-W15-2/tp
|
closed
|
testing: improve code coverage
|
type.Documentation
|
@AY2223S2-CS2113-W15-2/developers: We should aim to minimally achieve 100% LOC coverage (currently 79%) in testing to avoid unexpected bugs surfacing, with the long-term goal of achieving 100% branch coverage.
__Todo:__
- [ ] Achieve 100% line coverage in sub-packages
- [x] Integration testing for `Frontend` and `Backend`
- [ ] System testing for `PocketPal`
Code coverage as of ee5e416e8c81fa48c12df28dc49a9f9760e71571:

|
1.0
|
testing: improve code coverage - @AY2223S2-CS2113-W15-2/developers: We should aim to minimally achieve 100% LOC coverage (currently 79%) in testing to avoid unexpected bugs surfacing, with the long-term goal of achieving 100% branch coverage.
__Todo:__
- [ ] Achieve 100% line coverage in sub-packages
- [x] Integration testing for `Frontend` and `Backend`
- [ ] System testing for `PocketPal`
Code coverage as of ee5e416e8c81fa48c12df28dc49a9f9760e71571:

|
non_process
|
testing improve code coverage developers we should aim to minimally achieve loc coverage currently in testing to avoid unexpected bugs surfacing with the long term goal of achieving branch coverage todo achieve line coverage in sub packages integration testing for frontend and backend system testing for pocketpal code coverage as of
| 0
|
54,938
| 11,352,940,052
|
IssuesEvent
|
2020-01-24 14:39:20
|
Altinn/altinn-studio
|
https://api.github.com/repos/Altinn/altinn-studio
|
opened
|
docker-compose build fails for repos - "max depth exceeded"
|
kind/bug quality/code solution/studio/repos
|
## Describe the bug
After using docker-compose for Altinn Studio with `--build` parameter multiple times, you will eventually get this error, and need to remove the gitea-image.
```
ERROR: Service 'altinn_repositories' failed to build: max depth exceeded
```
This is the same problem as described [here](https://benfi.ca/docker-failed-to-build-max-depth-exceeded/), and requires a simple fix in `docker-compose.yml`.
## To Reproduce
Build altinn studio using docker-compose multiple times.
## Expected behavior
No errors.
|
1.0
|
docker-compose build fails for repos - "max depth exceeded" - ## Describe the bug
After using docker-compose for Altinn Studio with `--build` parameter multiple times, you will eventually get this error, and need to remove the gitea-image.
```
ERROR: Service 'altinn_repositories' failed to build: max depth exceeded
```
This is the same problem as described [here](https://benfi.ca/docker-failed-to-build-max-depth-exceeded/), and requires a simple fix in `docker-compose.yml`.
## To Reproduce
Build altinn studio using docker-compose multiple times.
## Expected behavior
No errors.
|
non_process
|
docker compose build fails for repos max depth exceeded describe the bug after using docker compose for altinn studio with build parameter multiple times you will eventually get this error and need to remove the gitea image error service altinn repositories failed to build max depth exceeded this is the same problem as described and requires a simple fix in docker compose yml to reproduce build altinn studio using docker compose multiple times expected behavior no errors
| 0
|
18,708
| 24,601,712,149
|
IssuesEvent
|
2022-10-14 13:03:00
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process.spawn not checking null byte in args
|
child_process
|
### Version
v18.9.1
### Platform
Linux tsc-ubuntu2204 5.15.0-48-generic #54-Ubuntu SMP Fri Aug 26 13:26:29 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
### Subsystem
child_process
### What steps will reproduce the bug?
```js
//// spawner.js
const {spawn} = require('child_process');
const myArgs = ['dump.js','AAA','BBB\0XXX', 'CCC'];
console.log('spawner.js:', myArgs);
const childProcess = spawn('node', myArgs);
childProcess.stdout.on('data', (chunk) => {
console.log(chunk.toString());
});
```
```js
//// dump.js
console.log('dump.js:', process.argv);
```
```
$ node spawner.js
spawner.js: [ 'dump.js', 'AAA', 'BBB\x00XXX', 'CCC' ]
dump.js: [ '/usr/local/bin/node', '/tmp/dump.js', 'AAA', 'BBB', 'CCC' ]
```
### How often does it reproduce? Is there a required condition?
No particular condition required.
### What is the expected behavior?
Node.js should raise error when invalid args (containing null byte) are given.
### What do you see instead?
No error raised. Null byte and subsequent bytes are silently truncated.
### Additional information
Other languages below have null byte checking and raise error in the same situation.
```
Java: java.io.IOException: invalid null character in command
PHP: Uncaught ValueError: Command array element 4 contains a null byte
Python: ValueError: embedded null byte
Ruby: ArgumentError (string contains null byte)
```
IMO raising error is safer to avoid null byte injection.
|
1.0
|
child_process.spawn not checking null byte in args - ### Version
v18.9.1
### Platform
Linux tsc-ubuntu2204 5.15.0-48-generic #54-Ubuntu SMP Fri Aug 26 13:26:29 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
### Subsystem
child_process
### What steps will reproduce the bug?
```js
//// spawner.js
const {spawn} = require('child_process');
const myArgs = ['dump.js','AAA','BBB\0XXX', 'CCC'];
console.log('spawner.js:', myArgs);
const childProcess = spawn('node', myArgs);
childProcess.stdout.on('data', (chunk) => {
console.log(chunk.toString());
});
```
```js
//// dump.js
console.log('dump.js:', process.argv);
```
```
$ node spawner.js
spawner.js: [ 'dump.js', 'AAA', 'BBB\x00XXX', 'CCC' ]
dump.js: [ '/usr/local/bin/node', '/tmp/dump.js', 'AAA', 'BBB', 'CCC' ]
```
### How often does it reproduce? Is there a required condition?
No particular condition required.
### What is the expected behavior?
Node.js should raise error when invalid args (containing null byte) are given.
### What do you see instead?
No error raised. Null byte and subsequent bytes are silently truncated.
### Additional information
Other languages below have null byte checking and raise error in the same situation.
```
Java: java.io.IOException: invalid null character in command
PHP: Uncaught ValueError: Command array element 4 contains a null byte
Python: ValueError: embedded null byte
Ruby: ArgumentError (string contains null byte)
```
IMO raising error is safer to avoid null byte injection.
|
process
|
child process spawn not checking null byte in args version platform linux tsc generic ubuntu smp fri aug utc gnu linux subsystem child process what steps will reproduce the bug js spawner js const spawn require child process const myargs console log spawner js myargs const childprocess spawn node myargs childprocess stdout on data chunk console log chunk tostring js dump js console log dump js process argv node spawner js spawner js dump js how often does it reproduce is there a required condition no particular condition required what is the expected behavior node js should raise error when invalid args containing null byte are given what do you see instead no error raised null byte and subsequent bytes are silently truncated additional information other languages below have null byte checking and raise error in the same situation java java io ioexception invalid null character in command php uncaught valueerror command array element contains a null byte python valueerror embedded null byte ruby argumenterror string contains null byte imo raising error is safer to avoid null byte injection
| 1
|
4,524
| 7,187,307,395
|
IssuesEvent
|
2018-02-02 04:18:26
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
The HttpClient.SendAsync throws WebException instead of wrapping it with HttpRequestException
|
area-System.Net.Http bug needs more info tenet-compatibility
|
The 4.6.1 framework version did not do this.
When moving to .NET Standard 2.0 + .NET Framework 4.7.1, the same exception is no longer wrapped into an HttpRequestException (while the docs still state that it should).
This breaks robustness wrappers and retry mechanisms built around the HttpClient.SendAsync.
Consider the following stacktrace (processed with the [AsyncFriendlyStackTrace](https://github.com/aelij/AsyncFriendlyStackTrace))
```
System.Net.WebException: The request was aborted: The request was canceled.
System.Net.ConnectStream.BeginRead(Byte[] buffer, Int32 offset, Int32 size, AsyncCallback callback, Object state)
System.IO.Stream.<>c.<BeginEndReadAsync>b__43_0(Stream stream, ReadWriteParameters args, AsyncCallback callback, Object state)
System.Threading.Tasks.TaskFactory`1.FromAsyncTrim[TInstance,TArgs](TInstance thisRef, TArgs args, Func`5 beginMethod, Func`3 endMethod)
System.IO.Stream.BeginEndReadAsync(Byte[] buffer, Int32 offset, Int32 count)
System.IO.Stream.ReadAsync(Byte[] buffer, Int32 offset, Int32 count, CancellationToken cancellationToken)
async System.IO.Stream.CopyToAsyncInternal(?)
async System.Net.Http.HttpContent.LoadIntoBufferAsyncCore(?)
async System.Net.Http.HttpClient.FinishSendAsyncBuffered(?)
async System.Net.Http.HttpClient.SendAsync(?)
```
|
True
|
The HttpClient.SendAsync throws WebException instead of wrapping it with HttpRequestException - The 4.6.1 framework version did not do this.
When moving to .NET Standard 2.0 + .NET Framework 4.7.1, the same exception is no longer wrapped into an HttpRequestException (while the docs still state that it should).
This breaks robustness wrappers and retry mechanisms built around the HttpClient.SendAsync.
Consider the following stacktrace (processed with the [AsyncFriendlyStackTrace](https://github.com/aelij/AsyncFriendlyStackTrace))
```
System.Net.WebException: The request was aborted: The request was canceled.
System.Net.ConnectStream.BeginRead(Byte[] buffer, Int32 offset, Int32 size, AsyncCallback callback, Object state)
System.IO.Stream.<>c.<BeginEndReadAsync>b__43_0(Stream stream, ReadWriteParameters args, AsyncCallback callback, Object state)
System.Threading.Tasks.TaskFactory`1.FromAsyncTrim[TInstance,TArgs](TInstance thisRef, TArgs args, Func`5 beginMethod, Func`3 endMethod)
System.IO.Stream.BeginEndReadAsync(Byte[] buffer, Int32 offset, Int32 count)
System.IO.Stream.ReadAsync(Byte[] buffer, Int32 offset, Int32 count, CancellationToken cancellationToken)
async System.IO.Stream.CopyToAsyncInternal(?)
async System.Net.Http.HttpContent.LoadIntoBufferAsyncCore(?)
async System.Net.Http.HttpClient.FinishSendAsyncBuffered(?)
async System.Net.Http.HttpClient.SendAsync(?)
```
|
non_process
|
the httpclient sendasync throws webexception instead of wrapping it with httprequestexception the framework version did not do this when moving to net standard net framework the same exception is no longer wrapped into an httprequestexception while the docs still state that it should this breaks robustness wrappers and retry mechanisms built around the httpclient sendasync consider the following stacktrace processed with the system net webexception the request was aborted the request was canceled system net connectstream beginread byte buffer offset size asynccallback callback object state system io stream c b stream stream readwriteparameters args asynccallback callback object state system threading tasks taskfactory fromasynctrim tinstance thisref targs args func beginmethod func endmethod system io stream beginendreadasync byte buffer offset count system io stream readasync byte buffer offset count cancellationtoken cancellationtoken async system io stream copytoasyncinternal async system net http httpcontent loadintobufferasynccore async system net http httpclient finishsendasyncbuffered async system net http httpclient sendasync
| 0
|
21,179
| 28,149,381,438
|
IssuesEvent
|
2023-04-02 21:19:32
|
AbdElAziz333/Pluto
|
https://api.github.com/repos/AbdElAziz333/Pluto
|
closed
|
Mod causes server connection issues
|
bug vanilla parity in processing
|
Game version: Forge 40.1.84 (Minecraft 1.18.2)
Mod version: 0.0.2 (https://www.curseforge.com/minecraft/mc-mods/pluto/files/4049858)
Steps to reproduce:
-Install Pluto on both server and client
-Join server
-Disconnect from server
You will now be unable to reconnect to the server until the client is restarted.
|
1.0
|
Mod causes server connection issues - Game version: Forge 40.1.84 (Minecraft 1.18.2)
Mod version: 0.0.2 (https://www.curseforge.com/minecraft/mc-mods/pluto/files/4049858)
Steps to reproduce:
-Install Pluto on both server and client
-Join server
-Disconnect from server
You will now be unable to reconnect to the server until the client is restarted.
|
process
|
mod causes server connection issues game version forge minecraft mod version steps to reproduce install pluto on both server and client join server disconnect from server you will now be unable to reconnect to the server until the client is restarted
| 1
|
6,157
| 9,037,760,463
|
IssuesEvent
|
2019-02-09 14:00:00
|
kmycode/sangokukmy
|
https://api.github.com/repos/kmycode/sangokukmy
|
closed
|
城壁耐久力の、守兵への置き換え
|
enhancement process-processing
|
新規要素として、都市のパラメータ「城壁耐久力」を削除し、守兵へ置き換えることを検討しています。
## 城壁耐久力とは
原作からあった仕様で、耐久力が高いほど、戦争で城壁を攻め込まれたときの城壁の減りを遅らせることができました。
反面、三国志NET KMY Versionでは能力のインフレが起きやすく、戦争における耐久力はあまり意味をなしていませんでした。
## 修正概要
* 「城壁耐久力」パラメータを削除し、「守兵」に置き換える
* 守兵の最大は、標準で2000とする
* 「守兵募集(仮)」コマンドで守兵を増やすことができる。実行結果は武将の人望に依存する
* 守兵は、都市の技術に応じた兵種(弓系)をもつ(技術0で弓兵、技術300で弩兵、技術500で強弩、技術999で連弩)
* 守兵のいる都市を攻めると、守兵より反撃を受ける。守兵の能力は、武力:都市技術の5分の1程度とする(要バランス調整)
* 守兵のいない都市を攻めても、反撃を受けない
## その他
* 守兵の訓練度という要素も考えましたが、三国志NETの利用者が少ない今では、武将1人あたりの仕事を過剰に増やしてしまうので保留とします
|
2.0
|
城壁耐久力の、守兵への置き換え - 新規要素として、都市のパラメータ「城壁耐久力」を削除し、守兵へ置き換えることを検討しています。
## 城壁耐久力とは
原作からあった仕様で、耐久力が高いほど、戦争で城壁を攻め込まれたときの城壁の減りを遅らせることができました。
反面、三国志NET KMY Versionでは能力のインフレが起きやすく、戦争における耐久力はあまり意味をなしていませんでした。
## 修正概要
* 「城壁耐久力」パラメータを削除し、「守兵」に置き換える
* 守兵の最大は、標準で2000とする
* 「守兵募集(仮)」コマンドで守兵を増やすことができる。実行結果は武将の人望に依存する
* 守兵は、都市の技術に応じた兵種(弓系)をもつ(技術0で弓兵、技術300で弩兵、技術500で強弩、技術999で連弩)
* 守兵のいる都市を攻めると、守兵より反撃を受ける。守兵の能力は、武力:都市技術の5分の1程度とする(要バランス調整)
* 守兵のいない都市を攻めても、反撃を受けない
## その他
* 守兵の訓練度という要素も考えましたが、三国志NETの利用者が少ない今では、武将1人あたりの仕事を過剰に増やしてしまうので保留とします
|
process
|
城壁耐久力の、守兵への置き換え 新規要素として、都市のパラメータ「城壁耐久力」を削除し、守兵へ置き換えることを検討しています。 城壁耐久力とは 原作からあった仕様で、耐久力が高いほど、戦争で城壁を攻め込まれたときの城壁の減りを遅らせることができました。 反面、三国志net kmy versionでは能力のインフレが起きやすく、戦争における耐久力はあまり意味をなしていませんでした。 修正概要 「城壁耐久力」パラメータを削除し、「守兵」に置き換える 守兵の最大は、 「守兵募集(仮)」コマンドで守兵を増やすことができる。実行結果は武将の人望に依存する 守兵は、都市の技術に応じた兵種(弓系)をもつ( 、 、 、 ) 守兵のいる都市を攻めると、守兵より反撃を受ける。守兵の能力は、武力: (要バランス調整) 守兵のいない都市を攻めても、反撃を受けない その他 守兵の訓練度という要素も考えましたが、三国志netの利用者が少ない今では、
| 1
|
29,553
| 11,759,835,868
|
IssuesEvent
|
2020-03-13 18:06:22
|
01binary/elevator
|
https://api.github.com/repos/01binary/elevator
|
opened
|
CVE-2020-8116 (High) detected in dot-prop-4.2.0.tgz
|
security vulnerability
|
## CVE-2020-8116 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dot-prop-4.2.0.tgz</b></p></summary>
<p>Get, set, or delete a property from a nested object using a dot path</p>
<p>Library home page: <a href="https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz">https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/elevator/ClientApp/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/elevator/ClientApp/node_modules/dot-prop/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.1.5.tgz (Root Library)
- sw-precache-webpack-plugin-0.11.4.tgz
- sw-precache-5.2.1.tgz
- update-notifier-2.5.0.tgz
- configstore-3.1.2.tgz
- :x: **dot-prop-4.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/01binary/elevator/commit/c03855450ce69cbe684e2d0017a95692e42f929f">c03855450ce69cbe684e2d0017a95692e42f929f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in dot-prop npm package version 5.1.0 and earlier allows an attacker to add arbitrary properties to JavaScript language constructs such as objects.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: dot-prop - 5.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8116 (High) detected in dot-prop-4.2.0.tgz - ## CVE-2020-8116 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dot-prop-4.2.0.tgz</b></p></summary>
<p>Get, set, or delete a property from a nested object using a dot path</p>
<p>Library home page: <a href="https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz">https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/elevator/ClientApp/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/elevator/ClientApp/node_modules/dot-prop/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.1.5.tgz (Root Library)
- sw-precache-webpack-plugin-0.11.4.tgz
- sw-precache-5.2.1.tgz
- update-notifier-2.5.0.tgz
- configstore-3.1.2.tgz
- :x: **dot-prop-4.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/01binary/elevator/commit/c03855450ce69cbe684e2d0017a95692e42f929f">c03855450ce69cbe684e2d0017a95692e42f929f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in dot-prop npm package version 5.1.0 and earlier allows an attacker to add arbitrary properties to JavaScript language constructs such as objects.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: dot-prop - 5.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in dot prop tgz cve high severity vulnerability vulnerable library dot prop tgz get set or delete a property from a nested object using a dot path library home page a href path to dependency file tmp ws scm elevator clientapp package json path to vulnerable library tmp ws scm elevator clientapp node modules dot prop package json dependency hierarchy react scripts tgz root library sw precache webpack plugin tgz sw precache tgz update notifier tgz configstore tgz x dot prop tgz vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability in dot prop npm package version and earlier allows an attacker to add arbitrary properties to javascript language constructs such as objects publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution dot prop step up your open source security game with whitesource
| 0
|
19,627
| 25,982,546,182
|
IssuesEvent
|
2022-12-19 20:15:26
|
aws/sagemaker-python-sdk
|
https://api.github.com/repos/aws/sagemaker-python-sdk
|
closed
|
FrameworkProcessor is broken with SageMaker Pipelines
|
bug component: processing component: pipelines
|
**Describe the bug**
Trying to use any Processor derived from FrameworkProcessor is bugged with SageMaker Pipelines. There is a problem with the `command` and `entrypoint` parameter, where `command` does not pass `python3`, causing the following error:
> `line 2: import: command not found`
**To reproduce**
1. Create a FrameworkProcessor (i.e. PyTorchProcessor, TensorFlowProcessor)
2. Create a ProcessingStep and a Pipeline
3. Execute it
4. See it fail
**Expected behavior**
The pipeline should go through.
**Screenshots or logs**
Screenshot from Pipelines:

Logs from CloudWatch:
```
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 2: import: command not found
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 3: import: command not found
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 4: import: command not found
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 5: import: command not found
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 6: from: command not found
```
**System information**
A description of your system. Please provide:
- **SageMaker Python SDK version**: 2.57.0
- **Framework name (eg. PyTorch) or algorithm (eg. KMeans)**: Every Framework
- **Framework version**: Every version supported by SM
- **Python version**: 3.8
- **CPU or GPU**: CPU and GPU
- **Custom Docker image (Y/N)**: N
**Additional context**
N/A
|
1.0
|
FrameworkProcessor is broken with SageMaker Pipelines - **Describe the bug**
Trying to use any Processor derived from FrameworkProcessor is bugged with SageMaker Pipelines. There is a problem with the `command` and `entrypoint` parameter, where `command` does not pass `python3`, causing the following error:
> `line 2: import: command not found`
**To reproduce**
1. Create a FrameworkProcessor (i.e. PyTorchProcessor, TensorFlowProcessor)
2. Create a ProcessingStep and a Pipeline
3. Execute it
4. See it fail
**Expected behavior**
The pipeline should go through.
**Screenshots or logs**
Screenshot from Pipelines:

Logs from CloudWatch:
```
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 2: import: command not found
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 3: import: command not found
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 4: import: command not found
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 5: import: command not found
/opt/ml/processing/input/entrypoint/inference_with_processing.py: line 6: from: command not found
```
**System information**
A description of your system. Please provide:
- **SageMaker Python SDK version**: 2.57.0
- **Framework name (eg. PyTorch) or algorithm (eg. KMeans)**: Every Framework
- **Framework version**: Every version supported by SM
- **Python version**: 3.8
- **CPU or GPU**: CPU and GPU
- **Custom Docker image (Y/N)**: N
**Additional context**
N/A
|
process
|
frameworkprocessor is broken with sagemaker pipelines describe the bug trying to use any processor derived from frameworkprocessor is bugged with sagemaker pipelines there is a problem with the command and entrypoint parameter where command does not pass causing the following error line import command not found to reproduce create a frameworkprocessor i e pytorchprocessor tensorflowprocessor create a processingstep and a pipeline execute it see it fail expected behavior the pipeline should go through screenshots or logs screenshot from pipelines logs from cloudwatch opt ml processing input entrypoint inference with processing py line import command not found opt ml processing input entrypoint inference with processing py line import command not found opt ml processing input entrypoint inference with processing py line import command not found opt ml processing input entrypoint inference with processing py line import command not found opt ml processing input entrypoint inference with processing py line from command not found system information a description of your system please provide sagemaker python sdk version framework name eg pytorch or algorithm eg kmeans every framework framework version every version supported by sm python version cpu or gpu cpu and gpu custom docker image y n n additional context n a
| 1
|
126,144
| 26,786,983,792
|
IssuesEvent
|
2023-02-01 04:11:20
|
microsoft/vsmarketplace
|
https://api.github.com/repos/microsoft/vsmarketplace
|
closed
|
Extensions malfunction
|
vscode
|
Type: <b>Performance Issue</b>
Extensions keep disappearing every time I startup the app.
When I download an extension then close the app, the extensions don't show again when I reopen the app. I have been having this issue for the past one month.
VS Code version: Code 1.74.3 (97dec172d3256f8ca4bfb2143f3f76b503ca0534, 2023-01-09T16:59:02.252Z)
OS version: Windows_NT x64 10.0.18362
Modes:
Sandboxed: No
<!-- generated by issue reporter -->
|
1.0
|
Extensions malfunction - Type: <b>Performance Issue</b>
Extensions keep disappearing every time I startup the app.
When I download an extension then close the app, the extensions don't show again when I reopen the app. I have been having this issue for the past one month.
VS Code version: Code 1.74.3 (97dec172d3256f8ca4bfb2143f3f76b503ca0534, 2023-01-09T16:59:02.252Z)
OS version: Windows_NT x64 10.0.18362
Modes:
Sandboxed: No
<!-- generated by issue reporter -->
|
non_process
|
extensions malfunction type performance issue extensions keep disappearing every time i startup the app when i download an extension then close the app the extensions don t show again when i reopen the app i have been having this issue for the past one month vs code version code os version windows nt modes sandboxed no
| 0
|
283
| 2,722,682,804
|
IssuesEvent
|
2015-04-14 06:58:14
|
mkdocs/mkdocs
|
https://api.github.com/repos/mkdocs/mkdocs
|
closed
|
Create google group for discussions.
|
Process
|
Also link to this and the IRC channel in both README and the front page of the docs.
|
1.0
|
Create google group for discussions. - Also link to this and the IRC channel in both README and the front page of the docs.
|
process
|
create google group for discussions also link to this and the irc channel in both readme and the front page of the docs
| 1
|
6,197
| 3,770,583,899
|
IssuesEvent
|
2016-03-16 15:02:26
|
rust-lang/rust
|
https://api.github.com/repos/rust-lang/rust
|
closed
|
rustbuild: can't cross compile compiler-rt for ARM
|
A-rustbuild
|
rustbuild is using cmake to build compiler-rt instead of Makefiles like the old build system does,
and there are some issues that make the build fail when the target is ARM.
### Background info
For ARM, there 3 floats ABI (see `-mfloat-abi` in [1]):
- `soft`: processor has no Floating Point Unit (FPU), floating point operations are software
routines/library calls.
- `hard`: processor has FPU, floating point operations are hardware instructions.
- `softfp`: processor has FPU, can use the hardware instructions but still uses `soft` calling
convention.
As far as Rust is concerned:
- `arm(v7)-unknown-linux-gnueabihf` uses the `hard` ABI.
- `arm-unknown-linux-gnueabi` uses the `soft` ABI.
### Observations when using cmake
In general:
- compiler-rt makes a [distinction][2] between an `arm` target and an `armhf` target in its sanity
checks, but then goes off and generates the [same set of symbols][3] for both targets. Not a
problem, but why then bother making a distinction in the first place?
- (a) compiler-rt demands a C++ compiler in its checks even though it's not used at all to compile
the builtins that form part of libcompiler-rt.a. This forces the user to install a g++ compiler
were gcc was sufficient before (in the Makefile case). This particularly annoying for `musl`
targets where only a `musl-gcc` wrapper is available in the buildbots. Note: It seems the C++
compiler is required to compile the sanitizers, but we don't build those right now.
- When libcompiler-rt.a is built using cmake for ARM it has [more symbols][4] than when is built
using the Makefile system. Note the `*vfp` symbols which are in the cmake version but not in the
Makefile version; they are relevant in the next section.
When targeting `arm-unknown-linux-gnueabi`:
- (b) The extra `*vfp` symbols can't be compiled for this target; you get a compiler error: "Error:
selected processor does not support ARM mode". At least that's the case with Ubuntu's
`arm-linux-gnueabi-gcc` toolchain and a crosstool-ng generated toolchain.
When targeting `arm(v7)-unknown-linux-gnueabihf`:
- (c) compiler-rt [expects][2] the arm compiler to be able to generate programs for both the `soft`
ABI and the `hard` ABI. The arm toolchains shipped in Ubuntu have no problems with this check
(both tests pass), but when a crosstool-ng compiler is used this check fails in the `soft`
case and the cmake build is aborted.
Other than that, I have had no problems (cross) compiling compiler-rt to i686, x86_64, mips or
mipsel apart from the annoying extra g++ requirement.
### What to do?
The simplest way to get rid of all these issues is to build compiler-rt using Makefiles at least
when the target is ARM. The other alternative is to patch compiler-rt:
- To fix (a), we can disable the C++ compiler check because we are not compiling the sanitizers
(yet).
- For (b), we can filter out the problematic symbols (about 34 assembly files) when compiling for
soft float ARM. It's OK, I think, to exclude these symbols because our current releases don't include
them for any ARM target (AFAICT).
- For (c), we could disable the `soft` float check when the target is hard float ARM.
[1]: https://gcc.gnu.org/onlinedocs/gcc/ARM-Options.html
[2]: https://github.com/rust-lang/compiler-rt/blob/b6087e82ba1384c4af3adf2dc68e92316f0d4caf/cmake/config-ix.cmake#L210-L211
[3]: https://github.com/rust-lang/compiler-rt/blob/b6087e82ba1384c4af3adf2dc68e92316f0d4caf/lib/builtins/CMakeLists.txt#L331
[4]: https://gist.github.com/japaric/97bce36543e7c55149a4
Thoughts @alexcrichton?
|
1.0
|
rustbuild: can't cross compile compiler-rt for ARM - rustbuild is using cmake to build compiler-rt instead of Makefiles like the old build system does,
and there are some issues that make the build fail when the target is ARM.
### Background info
For ARM, there 3 floats ABI (see `-mfloat-abi` in [1]):
- `soft`: processor has no Floating Point Unit (FPU), floating point operations are software
routines/library calls.
- `hard`: processor has FPU, floating point operations are hardware instructions.
- `softfp`: processor has FPU, can use the hardware instructions but still uses `soft` calling
convention.
As far as Rust is concerned:
- `arm(v7)-unknown-linux-gnueabihf` uses the `hard` ABI.
- `arm-unknown-linux-gnueabi` uses the `soft` ABI.
### Observations when using cmake
In general:
- compiler-rt makes a [distinction][2] between an `arm` target and an `armhf` target in its sanity
checks, but then goes off and generates the [same set of symbols][3] for both targets. Not a
problem, but why then bother making a distinction in the first place?
- (a) compiler-rt demands a C++ compiler in its checks even though it's not used at all to compile
the builtins that form part of libcompiler-rt.a. This forces the user to install a g++ compiler
were gcc was sufficient before (in the Makefile case). This particularly annoying for `musl`
targets where only a `musl-gcc` wrapper is available in the buildbots. Note: It seems the C++
compiler is required to compile the sanitizers, but we don't build those right now.
- When libcompiler-rt.a is built using cmake for ARM it has [more symbols][4] than when is built
using the Makefile system. Note the `*vfp` symbols which are in the cmake version but not in the
Makefile version; they are relevant in the next section.
When targeting `arm-unknown-linux-gnueabi`:
- (b) The extra `*vfp` symbols can't be compiled for this target; you get a compiler error: "Error:
selected processor does not support ARM mode". At least that's the case with Ubuntu's
`arm-linux-gnueabi-gcc` toolchain and a crosstool-ng generated toolchain.
When targeting `arm(v7)-unknown-linux-gnueabihf`:
- (c) compiler-rt [expects][2] the arm compiler to be able to generate programs for both the `soft`
ABI and the `hard` ABI. The arm toolchains shipped in Ubuntu have no problems with this check
(both tests pass), but when a crosstool-ng compiler is used this check fails in the `soft`
case and the cmake build is aborted.
Other than that, I have had no problems (cross) compiling compiler-rt to i686, x86_64, mips or
mipsel apart from the annoying extra g++ requirement.
### What to do?
The simplest way to get rid of all these issues is to build compiler-rt using Makefiles at least
when the target is ARM. The other alternative is to patch compiler-rt:
- To fix (a), we can disable the C++ compiler check because we are not compiling the sanitizers
(yet).
- For (b), we can filter out the problematic symbols (about 34 assembly files) when compiling for
soft float ARM. It's OK, I think, to exclude these symbols because our current releases don't include
them for any ARM target (AFAICT).
- For (c), we could disable the `soft` float check when the target is hard float ARM.
[1]: https://gcc.gnu.org/onlinedocs/gcc/ARM-Options.html
[2]: https://github.com/rust-lang/compiler-rt/blob/b6087e82ba1384c4af3adf2dc68e92316f0d4caf/cmake/config-ix.cmake#L210-L211
[3]: https://github.com/rust-lang/compiler-rt/blob/b6087e82ba1384c4af3adf2dc68e92316f0d4caf/lib/builtins/CMakeLists.txt#L331
[4]: https://gist.github.com/japaric/97bce36543e7c55149a4
Thoughts @alexcrichton?
|
non_process
|
rustbuild can t cross compile compiler rt for arm rustbuild is using cmake to build compiler rt instead of makefiles like the old build system does and there are some issues that make the build fail when the target is arm background info for arm there floats abi see mfloat abi in soft processor has no floating point unit fpu floating point operations are software routines library calls hard processor has fpu floating point operations are hardware instructions softfp processor has fpu can use the hardware instructions but still uses soft calling convention as far as rust is concerned arm unknown linux gnueabihf uses the hard abi arm unknown linux gnueabi uses the soft abi observations when using cmake in general compiler rt makes a between an arm target and an armhf target in its sanity checks but then goes off and generates the for both targets not a problem but why then bother making a distinction in the first place a compiler rt demands a c compiler in its checks even though it s not used at all to compile the builtins that form part of libcompiler rt a this forces the user to install a g compiler were gcc was sufficient before in the makefile case this particularly annoying for musl targets where only a musl gcc wrapper is available in the buildbots note it seems the c compiler is required to compile the sanitizers but we don t build those right now when libcompiler rt a is built using cmake for arm it has than when is built using the makefile system note the vfp symbols which are in the cmake version but not in the makefile version they are relevant in the next section when targeting arm unknown linux gnueabi b the extra vfp symbols can t be compiled for this target you get a compiler error error selected processor does not support arm mode at least that s the case with ubuntu s arm linux gnueabi gcc toolchain and a crosstool ng generated toolchain when targeting arm unknown linux gnueabihf c compiler rt the arm compiler to be able to generate programs for both the soft abi and the hard abi the arm toolchains shipped in ubuntu have no problems with this check both tests pass but when a crosstool ng compiler is used this check fails in the soft case and the cmake build is aborted other than that i have had no problems cross compiling compiler rt to mips or mipsel apart from the annoying extra g requirement what to do the simplest way to get rid of all these issues is to build compiler rt using makefiles at least when the target is arm the other alternative is to patch compiler rt to fix a we can disable the c compiler check because we are not compiling the sanitizers yet for b we can filter out the problematic symbols about assembly files when compiling for soft float arm it s ok i think to exclude these symbols because our current releases don t include them for any arm target afaict for c we could disable the soft float check when the target is hard float arm thoughts alexcrichton
| 0
|
10,777
| 13,607,753,721
|
IssuesEvent
|
2020-09-23 00:21:22
|
GoogleCloudPlatform/cloud-ops-sandbox
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
|
reopened
|
Secure e2e tests & the test runner since it has many IAM roles
|
priority: p2 type: process
|
Our e2e test runner has a lot of permissions in IAM, with ever increasing scope as we add on features. This makes us vulnerable to strangers making malicious PRs to start the test runner and then abusing this access in code.
This is something that really should be resolved at the Github level. At the moment [it isn't](https://github.com/actions/runner/issues/494).
Discussed with @Daniel-Sanche, atm, our best option may be to:
1. Require external PRs come from forks, which can only run the base workflow tests
2. Constrain base workflow tests to only run on the following conditions:
- External push from fork > code is reviewed by owner > base test can run (this is so external PRs can't run automatically)
- Internal push from branch -> latest test can run
This issue blocks [Workload ID PR](https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/pull/448) which requires IAM permissions to set up workload.
|
1.0
|
Secure e2e tests & the test runner since it has many IAM roles - Our e2e test runner has a lot of permissions in IAM, with ever increasing scope as we add on features. This makes us vulnerable to strangers making malicious PRs to start the test runner and then abusing this access in code.
This is something that really should be resolved at the Github level. At the moment [it isn't](https://github.com/actions/runner/issues/494).
Discussed with @Daniel-Sanche, atm, our best option may be to:
1. Require external PRs come from forks, which can only run the base workflow tests
2. Constrain base workflow tests to only run on the following conditions:
- External push from fork > code is reviewed by owner > base test can run (this is so external PRs can't run automatically)
- Internal push from branch -> latest test can run
This issue blocks [Workload ID PR](https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/pull/448) which requires IAM permissions to set up workload.
|
process
|
secure tests the test runner since it has many iam roles our test runner has a lot of permissions in iam with ever increasing scope as we add on features this makes us vulnerable to strangers making malicious prs to start the test runner and then abusing this access in code this is something that really should be resolved at the github level at the moment discussed with daniel sanche atm our best option may be to require external prs come from forks which can only run the base workflow tests constrain base workflow tests to only run on the following conditions external push from fork code is reviewed by owner base test can run this is so external prs can t run automatically internal push from branch latest test can run this issue blocks which requires iam permissions to set up workload
| 1
|
4,385
| 7,274,977,874
|
IssuesEvent
|
2018-02-21 11:56:11
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Function wrappers break is-defined checks
|
AREA: client AREA: server SYSTEM: resource processing TYPE: bug
|
Our site runs some code in a webworker, using an internal package that in some circumstances accesses `localStorage`:
```javascript
if (typeof (localStorage) !== 'undefined') {
// ...
}
```
However, this completely breaks after Hammerhead (for some reason?) surrounds this with a wrapper.
```javascript
// near the script top:
__get$Storage=function(s){return s}
// later on:
if (typeof __get$Storage(localStorage) !== 'undefined') {
// ...
}
```
Now, it attempts to evaluate `localStorage` before checking if it exists, so the webworker crashes.
|
1.0
|
Function wrappers break is-defined checks - Our site runs some code in a webworker, using an internal package that in some circumstances accesses `localStorage`:
```javascript
if (typeof (localStorage) !== 'undefined') {
// ...
}
```
However, this completely breaks after Hammerhead (for some reason?) surrounds this with a wrapper.
```javascript
// near the script top:
__get$Storage=function(s){return s}
// later on:
if (typeof __get$Storage(localStorage) !== 'undefined') {
// ...
}
```
Now, it attempts to evaluate `localStorage` before checking if it exists, so the webworker crashes.
|
process
|
function wrappers break is defined checks our site runs some code in a webworker using an internal package that in some circumstances accesses localstorage javascript if typeof localstorage undefined however this completely breaks after hammerhead for some reason surrounds this with a wrapper javascript near the script top get storage function s return s later on if typeof get storage localstorage undefined now it attempts to evaluate localstorage before checking if it exists so the webworker crashes
| 1
|
96,841
| 28,028,403,080
|
IssuesEvent
|
2023-03-28 10:40:47
|
yeahitsjan/pawxel
|
https://api.github.com/repos/yeahitsjan/pawxel
|
closed
|
leftover files after `make clean`
|
linux build
|
After running `make clean` there are still some `.o` files left, is this expected?
```
$ make clean
rm -f qrc_framelesshelpercore.cpp qrc_resources.cpp
rm -f moc_predefs.h
rm -f moc_application.cpp moc_sniparea.cpp moc_preview_window.cpp moc_editor_view.cpp moc_items.cpp moc_editor_window.cpp moc_preferences.cpp moc_preferences_window.cpp moc_systray.cpp moc_about_window.cpp moc_emoji_widget.cpp moc_chromepalette.cpp moc_framelesshelper_qt.cpp moc_framelesshelpercore_global.cpp moc_framelessmanager.cpp moc_micamaterial.cpp moc_windowborderpainter.cpp moc_chromepalette_p.cpp moc_framelessconfig_p.cpp moc_framelessmanager_p.cpp moc_micamaterial_p.cpp moc_sysapiloader_p.cpp moc_windowborderpainter_p.cpp moc_framelesswidget.cpp moc_framelessmainwindow.cpp moc_standardsystembutton.cpp moc_framelesswidgetshelper.cpp moc_standardtitlebar.cpp moc_framelessdialog.cpp moc_framelesswidgetshelper_p.cpp moc_standardsystembutton_p.cpp moc_standardtitlebar_p.cpp moc_framelesswidget_p.cpp moc_framelessmainwindow_p.cpp moc_widgetssharedhelper_p.cpp moc_framelessdialog_p.cpp moc_phantomstyle.cpp moc_singleapplication.cpp moc_singleapplication_p.cpp moc_qhotkey.cpp moc_qhotkey_p.cpp moc_graphicsitemresizer.cpp moc_FHNotifyWidget.cpp moc_FHMessageBox.cpp
rm -f application.o main.o sniparea.o preview_window.o editor_view.o items.o editor_window.o preferences.o preferences_window.o systray.o about_window.o qt_flowlayout.o emoji_widget.o easylogging++.o chromepalette.o framelessconfig.o framelesshelper_qt.o framelessmanager.o framelesshelpercore_global.o micamaterial.o sysapiloader.o utils.o windowborderpainter.o utils_linux.o framelessmainwindow.o framelesswidgetshelper.o framelesswidget.o standardsystembutton.o standardtitlebar.o widgetssharedhelper.o framelesshelperwidgets_global.o framelessdialog.o phantomcolor.o phantomstyle.o singleapplication.o singleapplication_p.o qhotkey.o qhotkey_x11.o graphicsitemresizer.o handlerstrategies.o resizehandleitem.o FHNotifyWidget.o FHMessageBox.o qrc_framelesshelpercore.o qrc_resources.o moc_application.o moc_sniparea.o moc_preview_window.o moc_editor_view.o moc_items.o moc_editor_window.o moc_preferences.o moc_preferences_window.o moc_systray.o moc_about_window.o moc_emoji_widget.o moc_chromepalette.o moc_framelesshelper_qt.o moc_framelesshelpercore_global.o moc_framelessmanager.o moc_micamaterial.o moc_windowborderpainter.o moc_chromepalette_p.o moc_framelessconfig_p.o moc_framelessmanager_p.o moc_micamaterial_p.o moc_sysapiloader_p.o moc_windowborderpainter_p.o moc_framelesswidget.o moc_framelessmainwindow.o moc_standardsystembutton.o moc_framelesswidgetshelper.o moc_standardtitlebar.o moc_framelessdialog.o moc_framelesswidgetshelper_p.o moc_standardsystembutton_p.o moc_standardtitlebar_p.o moc_framelesswidget_p.o moc_framelessmainwindow_p.o moc_widgetssharedhelper_p.o moc_framelessdialog_p.o moc_singleapplication.o moc_singleapplication_p.o moc_qhotkey.o moc_qhotkey_p.o moc_graphicsitemresizer.o moc_FHNotifyWidget.o moc_FHMessageBox.o
rm -f *~ core *.core
$ ls *.o
moc_select_window.o select_window.o
```
|
1.0
|
leftover files after `make clean` - After running `make clean` there are still some `.o` files left, is this expected?
```
$ make clean
rm -f qrc_framelesshelpercore.cpp qrc_resources.cpp
rm -f moc_predefs.h
rm -f moc_application.cpp moc_sniparea.cpp moc_preview_window.cpp moc_editor_view.cpp moc_items.cpp moc_editor_window.cpp moc_preferences.cpp moc_preferences_window.cpp moc_systray.cpp moc_about_window.cpp moc_emoji_widget.cpp moc_chromepalette.cpp moc_framelesshelper_qt.cpp moc_framelesshelpercore_global.cpp moc_framelessmanager.cpp moc_micamaterial.cpp moc_windowborderpainter.cpp moc_chromepalette_p.cpp moc_framelessconfig_p.cpp moc_framelessmanager_p.cpp moc_micamaterial_p.cpp moc_sysapiloader_p.cpp moc_windowborderpainter_p.cpp moc_framelesswidget.cpp moc_framelessmainwindow.cpp moc_standardsystembutton.cpp moc_framelesswidgetshelper.cpp moc_standardtitlebar.cpp moc_framelessdialog.cpp moc_framelesswidgetshelper_p.cpp moc_standardsystembutton_p.cpp moc_standardtitlebar_p.cpp moc_framelesswidget_p.cpp moc_framelessmainwindow_p.cpp moc_widgetssharedhelper_p.cpp moc_framelessdialog_p.cpp moc_phantomstyle.cpp moc_singleapplication.cpp moc_singleapplication_p.cpp moc_qhotkey.cpp moc_qhotkey_p.cpp moc_graphicsitemresizer.cpp moc_FHNotifyWidget.cpp moc_FHMessageBox.cpp
rm -f application.o main.o sniparea.o preview_window.o editor_view.o items.o editor_window.o preferences.o preferences_window.o systray.o about_window.o qt_flowlayout.o emoji_widget.o easylogging++.o chromepalette.o framelessconfig.o framelesshelper_qt.o framelessmanager.o framelesshelpercore_global.o micamaterial.o sysapiloader.o utils.o windowborderpainter.o utils_linux.o framelessmainwindow.o framelesswidgetshelper.o framelesswidget.o standardsystembutton.o standardtitlebar.o widgetssharedhelper.o framelesshelperwidgets_global.o framelessdialog.o phantomcolor.o phantomstyle.o singleapplication.o singleapplication_p.o qhotkey.o qhotkey_x11.o graphicsitemresizer.o handlerstrategies.o resizehandleitem.o FHNotifyWidget.o FHMessageBox.o qrc_framelesshelpercore.o qrc_resources.o moc_application.o moc_sniparea.o moc_preview_window.o moc_editor_view.o moc_items.o moc_editor_window.o moc_preferences.o moc_preferences_window.o moc_systray.o moc_about_window.o moc_emoji_widget.o moc_chromepalette.o moc_framelesshelper_qt.o moc_framelesshelpercore_global.o moc_framelessmanager.o moc_micamaterial.o moc_windowborderpainter.o moc_chromepalette_p.o moc_framelessconfig_p.o moc_framelessmanager_p.o moc_micamaterial_p.o moc_sysapiloader_p.o moc_windowborderpainter_p.o moc_framelesswidget.o moc_framelessmainwindow.o moc_standardsystembutton.o moc_framelesswidgetshelper.o moc_standardtitlebar.o moc_framelessdialog.o moc_framelesswidgetshelper_p.o moc_standardsystembutton_p.o moc_standardtitlebar_p.o moc_framelesswidget_p.o moc_framelessmainwindow_p.o moc_widgetssharedhelper_p.o moc_framelessdialog_p.o moc_singleapplication.o moc_singleapplication_p.o moc_qhotkey.o moc_qhotkey_p.o moc_graphicsitemresizer.o moc_FHNotifyWidget.o moc_FHMessageBox.o
rm -f *~ core *.core
$ ls *.o
moc_select_window.o select_window.o
```
|
non_process
|
leftover files after make clean after running make clean there are still some o files left is this expected make clean rm f qrc framelesshelpercore cpp qrc resources cpp rm f moc predefs h rm f moc application cpp moc sniparea cpp moc preview window cpp moc editor view cpp moc items cpp moc editor window cpp moc preferences cpp moc preferences window cpp moc systray cpp moc about window cpp moc emoji widget cpp moc chromepalette cpp moc framelesshelper qt cpp moc framelesshelpercore global cpp moc framelessmanager cpp moc micamaterial cpp moc windowborderpainter cpp moc chromepalette p cpp moc framelessconfig p cpp moc framelessmanager p cpp moc micamaterial p cpp moc sysapiloader p cpp moc windowborderpainter p cpp moc framelesswidget cpp moc framelessmainwindow cpp moc standardsystembutton cpp moc framelesswidgetshelper cpp moc standardtitlebar cpp moc framelessdialog cpp moc framelesswidgetshelper p cpp moc standardsystembutton p cpp moc standardtitlebar p cpp moc framelesswidget p cpp moc framelessmainwindow p cpp moc widgetssharedhelper p cpp moc framelessdialog p cpp moc phantomstyle cpp moc singleapplication cpp moc singleapplication p cpp moc qhotkey cpp moc qhotkey p cpp moc graphicsitemresizer cpp moc fhnotifywidget cpp moc fhmessagebox cpp rm f application o main o sniparea o preview window o editor view o items o editor window o preferences o preferences window o systray o about window o qt flowlayout o emoji widget o easylogging o chromepalette o framelessconfig o framelesshelper qt o framelessmanager o framelesshelpercore global o micamaterial o sysapiloader o utils o windowborderpainter o utils linux o framelessmainwindow o framelesswidgetshelper o framelesswidget o standardsystembutton o standardtitlebar o widgetssharedhelper o framelesshelperwidgets global o framelessdialog o phantomcolor o phantomstyle o singleapplication o singleapplication p o qhotkey o qhotkey o graphicsitemresizer o handlerstrategies o resizehandleitem o fhnotifywidget o fhmessagebox o qrc framelesshelpercore o qrc resources o moc application o moc sniparea o moc preview window o moc editor view o moc items o moc editor window o moc preferences o moc preferences window o moc systray o moc about window o moc emoji widget o moc chromepalette o moc framelesshelper qt o moc framelesshelpercore global o moc framelessmanager o moc micamaterial o moc windowborderpainter o moc chromepalette p o moc framelessconfig p o moc framelessmanager p o moc micamaterial p o moc sysapiloader p o moc windowborderpainter p o moc framelesswidget o moc framelessmainwindow o moc standardsystembutton o moc framelesswidgetshelper o moc standardtitlebar o moc framelessdialog o moc framelesswidgetshelper p o moc standardsystembutton p o moc standardtitlebar p o moc framelesswidget p o moc framelessmainwindow p o moc widgetssharedhelper p o moc framelessdialog p o moc singleapplication o moc singleapplication p o moc qhotkey o moc qhotkey p o moc graphicsitemresizer o moc fhnotifywidget o moc fhmessagebox o rm f core core ls o moc select window o select window o
| 0
|
1,175
| 2,599,630,398
|
IssuesEvent
|
2015-02-23 10:23:54
|
v-l-m/vlm
|
https://api.github.com/repos/v-l-m/vlm
|
closed
|
Utiliser le redimensionnement auto coté client pour les infos bulles (images)
|
C: site P: major R: fixed T: defect
|
**Reported by paparazzia on 4 Oct 2009 10:04 UTC**
* ca vite de faire des accs locaux juste pour lire la dimension de l'image
* c'est de toute faon ncessaire avec la faon de grer les images dynamiques.
|
1.0
|
Utiliser le redimensionnement auto coté client pour les infos bulles (images) - **Reported by paparazzia on 4 Oct 2009 10:04 UTC**
* ca vite de faire des accs locaux juste pour lire la dimension de l'image
* c'est de toute faon ncessaire avec la faon de grer les images dynamiques.
|
non_process
|
utiliser le redimensionnement auto coté client pour les infos bulles images reported by paparazzia on oct utc ca vite de faire des accs locaux juste pour lire la dimension de l image c est de toute faon ncessaire avec la faon de grer les images dynamiques
| 0
|
4,904
| 7,783,536,285
|
IssuesEvent
|
2018-06-06 10:11:19
|
gvwilson/teachtogether.tech
|
https://api.github.com/repos/gvwilson/teachtogether.tech
|
opened
|
Ch06 Joel Ostblom
|
Ch06 Process
|
- on writing things up in a logical order: this can reinforce the reader's impostor syndrome, since they *aren't* doing things in that order
- "Indistinguishable from what good teachers do using..." are there specific references for this?
|
1.0
|
Ch06 Joel Ostblom - - on writing things up in a logical order: this can reinforce the reader's impostor syndrome, since they *aren't* doing things in that order
- "Indistinguishable from what good teachers do using..." are there specific references for this?
|
process
|
joel ostblom on writing things up in a logical order this can reinforce the reader s impostor syndrome since they aren t doing things in that order indistinguishable from what good teachers do using are there specific references for this
| 1
|
9,447
| 12,427,312,279
|
IssuesEvent
|
2020-05-25 01:53:01
|
GetTerminus/terminus-ui
|
https://api.github.com/repos/GetTerminus/terminus-ui
|
opened
|
Look into test component harness
|
Focus: community Goal: Process Improvement Needs: exploration
|
Should we use this concept? Is it easier to use or maintain vs standalone functions like we currently support?
- https://material.angular.io/cdk/test-harnesses/overview
- https://material.angular.io/guide/using-component-harnesses
- https://medium.com/@mhamel06/component-test-harnesses-angular-9s-second-best-feature-4261cabedc77
- https://indepth.dev/create-a-component-harness-for-your-tests-with-angular-cdk/
---
- [ ] Should we leverage this concept?
- [ ] What is the path forward?
|
1.0
|
Look into test component harness - Should we use this concept? Is it easier to use or maintain vs standalone functions like we currently support?
- https://material.angular.io/cdk/test-harnesses/overview
- https://material.angular.io/guide/using-component-harnesses
- https://medium.com/@mhamel06/component-test-harnesses-angular-9s-second-best-feature-4261cabedc77
- https://indepth.dev/create-a-component-harness-for-your-tests-with-angular-cdk/
---
- [ ] Should we leverage this concept?
- [ ] What is the path forward?
|
process
|
look into test component harness should we use this concept is it easier to use or maintain vs standalone functions like we currently support should we leverage this concept what is the path forward
| 1
|
294,814
| 9,048,723,164
|
IssuesEvent
|
2019-02-12 01:17:35
|
less/less.js
|
https://api.github.com/repos/less/less.js
|
closed
|
Comments inside ruleset maps are parsed by the each-function
|
feature request medium priority
|
Let's say I have the following map and iterate over it with the each-function:

This will result in the following compiled output (with less.js):

The lines that are commented out are also parsed and the index is used as the key-variable somehow.
I don't think this is the intended behaviour.
|
1.0
|
Comments inside ruleset maps are parsed by the each-function - Let's say I have the following map and iterate over it with the each-function:

This will result in the following compiled output (with less.js):

The lines that are commented out are also parsed and the index is used as the key-variable somehow.
I don't think this is the intended behaviour.
|
non_process
|
comments inside ruleset maps are parsed by the each function let s say i have the following map and iterate over it with the each function this will result in the following compiled output with less js the lines that are commented out are also parsed and the index is used as the key variable somehow i don t think this is the intended behaviour
| 0
|
384,566
| 26,593,402,164
|
IssuesEvent
|
2023-01-23 10:31:49
|
pyodide/pyodide
|
https://api.github.com/repos/pyodide/pyodide
|
opened
|
Detectiong Pyodide, at build time
|
documentation
|
## 📚 Documentation
The current documentation reads:
To detect Pyodide, at build time use,
```
import os
if "PYODIDE" in os.environ:
# building for Pyodide
```
This used to work on my computer using pyodide-build v0.21.
When I tried it using pyodide-build v0.22 however, the test `if "PYODIDE" in os.environ` returned false.
Inspecting my `os.environ`, I changed the test to `if ("PYODIDE_ROOT" in os.environ):`, and my setup.py detects when the code is being built with pyodide.
This lead me to believe the docs needed updating. However, searching the pyodide repo, I saw `if "PYODIDE" in os.environ` in the following patch file: https://github.com/pyodide/pyodide/blob/main/packages/Pillow/patches/0001-Enable-image-formats.patch
So I'm not really sure anymore ...
|
1.0
|
Detectiong Pyodide, at build time - ## 📚 Documentation
The current documentation reads:
To detect Pyodide, at build time use,
```
import os
if "PYODIDE" in os.environ:
# building for Pyodide
```
This used to work on my computer using pyodide-build v0.21.
When I tried it using pyodide-build v0.22 however, the test `if "PYODIDE" in os.environ` returned false.
Inspecting my `os.environ`, I changed the test to `if ("PYODIDE_ROOT" in os.environ):`, and my setup.py detects when the code is being built with pyodide.
This lead me to believe the docs needed updating. However, searching the pyodide repo, I saw `if "PYODIDE" in os.environ` in the following patch file: https://github.com/pyodide/pyodide/blob/main/packages/Pillow/patches/0001-Enable-image-formats.patch
So I'm not really sure anymore ...
|
non_process
|
detectiong pyodide at build time 📚 documentation the current documentation reads to detect pyodide at build time use import os if pyodide in os environ building for pyodide this used to work on my computer using pyodide build when i tried it using pyodide build however the test if pyodide in os environ returned false inspecting my os environ i changed the test to if pyodide root in os environ and my setup py detects when the code is being built with pyodide this lead me to believe the docs needed updating however searching the pyodide repo i saw if pyodide in os environ in the following patch file so i m not really sure anymore
| 0
|
46,067
| 7,235,638,980
|
IssuesEvent
|
2018-02-13 01:50:10
|
Marak/colors.js
|
https://api.github.com/repos/Marak/colors.js
|
opened
|
Add CONTRIBUTING.md
|
documentation
|
It would be good to have a guide on how to propose useful new features/fixes to the repo. In particular, we should be more explicit about what kinds of things we will/won't accept, how to format/test code, etc. We should also make things clear enough that first-time contributors don't feel discouraged from pushing fixes.
|
1.0
|
Add CONTRIBUTING.md - It would be good to have a guide on how to propose useful new features/fixes to the repo. In particular, we should be more explicit about what kinds of things we will/won't accept, how to format/test code, etc. We should also make things clear enough that first-time contributors don't feel discouraged from pushing fixes.
|
non_process
|
add contributing md it would be good to have a guide on how to propose useful new features fixes to the repo in particular we should be more explicit about what kinds of things we will won t accept how to format test code etc we should also make things clear enough that first time contributors don t feel discouraged from pushing fixes
| 0
|
273,577
| 20,798,422,606
|
IssuesEvent
|
2022-03-17 11:35:54
|
bkaczkowski/xcore
|
https://api.github.com/repos/bkaczkowski/xcore
|
closed
|
modelGeneExpression improve output description
|
documentation
|
Improve modelGeneExpression output description.
|
1.0
|
modelGeneExpression improve output description - Improve modelGeneExpression output description.
|
non_process
|
modelgeneexpression improve output description improve modelgeneexpression output description
| 0
|
231,442
| 25,499,199,649
|
IssuesEvent
|
2022-11-28 01:17:38
|
MValle21/Intelehealth-WebApp
|
https://api.github.com/repos/MValle21/Intelehealth-WebApp
|
opened
|
CVE-2022-24999 (Medium) detected in qs-6.5.2.tgz, qs-6.7.0.tgz
|
security vulnerability
|
## CVE-2022-24999 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>qs-6.5.2.tgz</b>, <b>qs-6.7.0.tgz</b></p></summary>
<p>
<details><summary><b>qs-6.5.2.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.5.2.tgz">https://registry.npmjs.org/qs/-/qs-6.5.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/request/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- cli-10.0.6.tgz (Root Library)
- universal-analytics-0.4.20.tgz
- request-2.88.2.tgz
- :x: **qs-6.5.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>qs-6.7.0.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.7.0.tgz">https://registry.npmjs.org/qs/-/qs-6.7.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- karma-5.1.1.tgz (Root Library)
- body-parser-1.19.0.tgz
- :x: **qs-6.7.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/MValle21/Intelehealth-WebApp/commit/bcbacdc3090e8dddcad46176ac579f0fb3db5a5f">bcbacdc3090e8dddcad46176ac579f0fb3db5a5f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable).
<p>Publish Date: 2022-11-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p>
<p>Release Date: 2022-11-26</p>
<p>Fix Resolution: qs - 6.2.4,6.3.3,6.4.1,6.5.3,6.6.1,6.7.3,6.8.3,6.9.7,6.10.3</p>
</p>
</details>
<p></p>
|
True
|
CVE-2022-24999 (Medium) detected in qs-6.5.2.tgz, qs-6.7.0.tgz - ## CVE-2022-24999 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>qs-6.5.2.tgz</b>, <b>qs-6.7.0.tgz</b></p></summary>
<p>
<details><summary><b>qs-6.5.2.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.5.2.tgz">https://registry.npmjs.org/qs/-/qs-6.5.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/request/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- cli-10.0.6.tgz (Root Library)
- universal-analytics-0.4.20.tgz
- request-2.88.2.tgz
- :x: **qs-6.5.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>qs-6.7.0.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.7.0.tgz">https://registry.npmjs.org/qs/-/qs-6.7.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- karma-5.1.1.tgz (Root Library)
- body-parser-1.19.0.tgz
- :x: **qs-6.7.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/MValle21/Intelehealth-WebApp/commit/bcbacdc3090e8dddcad46176ac579f0fb3db5a5f">bcbacdc3090e8dddcad46176ac579f0fb3db5a5f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable).
<p>Publish Date: 2022-11-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p>
<p>Release Date: 2022-11-26</p>
<p>Fix Resolution: qs - 6.2.4,6.3.3,6.4.1,6.5.3,6.6.1,6.7.3,6.8.3,6.9.7,6.10.3</p>
</p>
</details>
<p></p>
|
non_process
|
cve medium detected in qs tgz qs tgz cve medium severity vulnerability vulnerable libraries qs tgz qs tgz qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules request node modules qs package json dependency hierarchy cli tgz root library universal analytics tgz request tgz x qs tgz vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules qs package json dependency hierarchy karma tgz root library body parser tgz x qs tgz vulnerable library found in head commit a href found in base branch master vulnerability details qs before as used in express before and other products allows attackers to cause a node process hang for an express application because an proto key can be used in many typical express use cases an unauthenticated remote attacker can place the attack payload in the query string of the url that is used to visit the application such as a b a a the fix was backported to qs and and therefore express which has deps qs in its release description is not vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs
| 0
|
1,541
| 6,572,233,159
|
IssuesEvent
|
2017-09-11 00:23:20
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
Haproxy module doesn't check if the service is present on the given host
|
affects_2.0 bug_report networking waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
haproxy
##### ANSIBLE VERSION
2.0.2.0
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
The haproxy module assumes that a given host presents in _all_ proxies, and if it is not true, an error occures.
##### STEPS TO REPRODUCE
haproxy: host=a_backend_host socket=/run/haproxy/admin.sock state=enabled
If a service exists in haproxy, but there's no backend at the backend host, the module fails.
##### EXPECTED RESULTS
It should skip the service instead.
##### ACTUAL RESULTS
It fails, with the message: 'unable to find server backend/a_backend_host'
|
True
|
Haproxy module doesn't check if the service is present on the given host - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
haproxy
##### ANSIBLE VERSION
2.0.2.0
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
The haproxy module assumes that a given host presents in _all_ proxies, and if it is not true, an error occures.
##### STEPS TO REPRODUCE
haproxy: host=a_backend_host socket=/run/haproxy/admin.sock state=enabled
If a service exists in haproxy, but there's no backend at the backend host, the module fails.
##### EXPECTED RESULTS
It should skip the service instead.
##### ACTUAL RESULTS
It fails, with the message: 'unable to find server backend/a_backend_host'
|
non_process
|
haproxy module doesn t check if the service is present on the given host issue type bug report component name haproxy ansible version configuration os environment summary the haproxy module assumes that a given host presents in all proxies and if it is not true an error occures steps to reproduce haproxy host a backend host socket run haproxy admin sock state enabled if a service exists in haproxy but there s no backend at the backend host the module fails expected results it should skip the service instead actual results it fails with the message unable to find server backend a backend host
| 0
|
830,715
| 32,022,074,870
|
IssuesEvent
|
2023-09-22 05:52:02
|
oceanbase/odc
|
https://api.github.com/repos/oceanbase/odc
|
opened
|
[Feature]: White-screen configuration for integrating bastion user login
|
type-feature priority-medium module-Bastion integration
|
### Is your feature request related to a problem?
ODC support bastion user login integration since 3.x release.
But, the configuration is complex, because you should modifying meta DB.
### Describe the solution you'd like
ODC support a white-screen configuration for integrating bastion user login
### Additional context
_No response_
|
1.0
|
[Feature]: White-screen configuration for integrating bastion user login - ### Is your feature request related to a problem?
ODC support bastion user login integration since 3.x release.
But, the configuration is complex, because you should modifying meta DB.
### Describe the solution you'd like
ODC support a white-screen configuration for integrating bastion user login
### Additional context
_No response_
|
non_process
|
white screen configuration for integrating bastion user login is your feature request related to a problem odc support bastion user login integration since x release but the configuration is complex because you should modifying meta db describe the solution you d like odc support a white screen configuration for integrating bastion user login additional context no response
| 0
|
33,603
| 7,735,752,047
|
IssuesEvent
|
2018-05-27 18:34:40
|
fga-gpp-mds/2018.1_Nexte
|
https://api.github.com/repos/fga-gpp-mds/2018.1_Nexte
|
closed
|
Eu, como usuário, desejo realizar efetivamente meu login.
|
code eps infrastructure medium refactoring
|
## Descrição
Essa issue destina-se a integrar o login ao servidor.
## Critérios de Aceitação
- [ ] Integrar login ao servidor;
- [ ] Fazer login funcionar tanto com informações normais, quanto com o account kit;
- [ ] Atualizar o user do singleton de usuário.
|
1.0
|
Eu, como usuário, desejo realizar efetivamente meu login. - ## Descrição
Essa issue destina-se a integrar o login ao servidor.
## Critérios de Aceitação
- [ ] Integrar login ao servidor;
- [ ] Fazer login funcionar tanto com informações normais, quanto com o account kit;
- [ ] Atualizar o user do singleton de usuário.
|
non_process
|
eu como usuário desejo realizar efetivamente meu login descrição essa issue destina se a integrar o login ao servidor critérios de aceitação integrar login ao servidor fazer login funcionar tanto com informações normais quanto com o account kit atualizar o user do singleton de usuário
| 0
|
9,414
| 12,407,418,306
|
IssuesEvent
|
2020-05-21 21:03:06
|
burnpiro/wod-bike-dataset-generator
|
https://api.github.com/repos/burnpiro/wod-bike-dataset-generator
|
closed
|
Preprocessing for csv files from './data'
|
data processing
|
We need to clean and save preprocessed files from `./data` folder
|
1.0
|
Preprocessing for csv files from './data' - We need to clean and save preprocessed files from `./data` folder
|
process
|
preprocessing for csv files from data we need to clean and save preprocessed files from data folder
| 1
|
9,381
| 6,277,930,102
|
IssuesEvent
|
2017-07-18 13:24:04
|
giantswarm/cli
|
https://api.github.com/repos/giantswarm/cli
|
closed
|
how to delete orgs and envs is inconsistent
|
0 - status/inbox area/consistency area/usability
|
deleting orgs `swarm org delete <org>` ... and deleting envs `swarm env -d <env>`
Similar issue: https://github.com/giantswarm/cli/issues/137
|
True
|
how to delete orgs and envs is inconsistent - deleting orgs `swarm org delete <org>` ... and deleting envs `swarm env -d <env>`
Similar issue: https://github.com/giantswarm/cli/issues/137
|
non_process
|
how to delete orgs and envs is inconsistent deleting orgs swarm org delete and deleting envs swarm env d similar issue
| 0
|
13,937
| 16,704,593,542
|
IssuesEvent
|
2021-06-09 08:29:12
|
hochschule-darmstadt/openartbrowser
|
https://api.github.com/repos/hochschule-darmstadt/openartbrowser
|
opened
|
Blacklist functionality in etl
|
etl process medium priority
|
**Reason (Why?)**
If some categories seem inappropriate, there should be a functionality to filter them from the dataset.
**Solution (What?)**
Implementation of a blacklist, which is applied to the crawling process in a performant way. The blacklist should be a json array, containing all QIDs of blacklisteted entities (all types possible). As a result, the entity and all references to it should not appear in the dataset.
**Acceptance criteria**
Blacklisted entities no longer appear in the dataset.
|
1.0
|
Blacklist functionality in etl - **Reason (Why?)**
If some categories seem inappropriate, there should be a functionality to filter them from the dataset.
**Solution (What?)**
Implementation of a blacklist, which is applied to the crawling process in a performant way. The blacklist should be a json array, containing all QIDs of blacklisteted entities (all types possible). As a result, the entity and all references to it should not appear in the dataset.
**Acceptance criteria**
Blacklisted entities no longer appear in the dataset.
|
process
|
blacklist functionality in etl reason why if some categories seem inappropriate there should be a functionality to filter them from the dataset solution what implementation of a blacklist which is applied to the crawling process in a performant way the blacklist should be a json array containing all qids of blacklisteted entities all types possible as a result the entity and all references to it should not appear in the dataset acceptance criteria blacklisted entities no longer appear in the dataset
| 1
|
43,706
| 17,633,749,553
|
IssuesEvent
|
2021-08-19 11:17:09
|
hashicorp/terraform-provider-azurerm
|
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
|
closed
|
Support for function_app vnet integration
|
question service/functions
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
The field `virtual_network_name` was removed from the `site_config` block since azurerm provider version 2.0.
I'm currently working in a previous version which allow me to use this field in an `ignore_changes` block like following (for both app_service and app_function):
```
lifecycle {
ignore_changes = [
site_config[0].vnetName,
app_settings
]
```
I need this to avoid breaking the vnet integration did by ARM templates.
How can I do the same in 2.0 version?
**New or Affected Resource(s)**
- azurerm_app_service
- azurerm_function_app
|
1.0
|
Support for function_app vnet integration - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
The field `virtual_network_name` was removed from the `site_config` block since azurerm provider version 2.0.
I'm currently working in a previous version which allow me to use this field in an `ignore_changes` block like following (for both app_service and app_function):
```
lifecycle {
ignore_changes = [
site_config[0].vnetName,
app_settings
]
```
I need this to avoid breaking the vnet integration did by ARM templates.
How can I do the same in 2.0 version?
**New or Affected Resource(s)**
- azurerm_app_service
- azurerm_function_app
|
non_process
|
support for function app vnet integration community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the field virtual network name was removed from the site config block since azurerm provider version i m currently working in a previous version which allow me to use this field in an ignore changes block like following for both app service and app function lifecycle ignore changes site config vnetname app settings i need this to avoid breaking the vnet integration did by arm templates how can i do the same in version new or affected resource s azurerm app service azurerm function app
| 0
|
585,345
| 17,485,499,081
|
IssuesEvent
|
2021-08-09 10:28:04
|
nimblehq/nimble-medium-ios
|
https://api.github.com/repos/nimblehq/nimble-medium-ios
|
opened
|
As a user, I can login from the article details screen
|
type : feature type: integration priority: medium
|
## Why
For the existing users of the application that don't login yet, they should be able to see the `Login` screen for being able to login into the application, use more features and manage personal contents.
## Acceptance Criteria
- [ ] When the users tap on the `Login` highlighted text from the `Article Details` screen, present the common `Login ` screen modally.
- [ ] When the users tap on the `Close` button in the `Login ` screen, dismiss it modally.
## Resources
- Sample UX Flow:
https://user-images.githubusercontent.com/70877098/128692873-2d67b06d-bd97-40ef-b2ef-64c792ec478b.mov
|
1.0
|
As a user, I can login from the article details screen - ## Why
For the existing users of the application that don't login yet, they should be able to see the `Login` screen for being able to login into the application, use more features and manage personal contents.
## Acceptance Criteria
- [ ] When the users tap on the `Login` highlighted text from the `Article Details` screen, present the common `Login ` screen modally.
- [ ] When the users tap on the `Close` button in the `Login ` screen, dismiss it modally.
## Resources
- Sample UX Flow:
https://user-images.githubusercontent.com/70877098/128692873-2d67b06d-bd97-40ef-b2ef-64c792ec478b.mov
|
non_process
|
as a user i can login from the article details screen why for the existing users of the application that don t login yet they should be able to see the login screen for being able to login into the application use more features and manage personal contents acceptance criteria when the users tap on the login highlighted text from the article details screen present the common login screen modally when the users tap on the close button in the login screen dismiss it modally resources sample ux flow
| 0
|
21,948
| 30,449,508,104
|
IssuesEvent
|
2023-07-16 05:19:30
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Add support to attributes with array values in processors that manipulate attributes
|
enhancement Stale processor/attributes processor/resourcedetection processor/resource priority:needed closed as inactive
|
### Is your feature request related to a problem? Please describe.
[The spec states](https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/common/README.md#attribute) that attribute values can be "An array of primitive type values. The array MUST be homogeneous, i.e., it MUST NOT contain values of different types.".
The following processors manipulate attribute values:
* https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/attributesprocessor
* https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/resourceprocessor
* https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/resourcedetectionprocessor
However they don't support arrays of primitive type.
When I try to do something like:
```
resource:
attributes:
- key: "aws.log.group.names"
action: insert
value: [sample-app-trace-logs]
```
I get back the following from the collector stdout:
```
Error: cannot build pipelines: failed to create "resource" processor, in pipeline "traces": error creating "resource" processor: error unsupported value type "[]interface {}"
```
### Describe the solution you'd like
I'd like that any type or processor that manipulate attributes to support arrays of primitive types.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
|
3.0
|
Add support to attributes with array values in processors that manipulate attributes - ### Is your feature request related to a problem? Please describe.
[The spec states](https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/common/README.md#attribute) that attribute values can be "An array of primitive type values. The array MUST be homogeneous, i.e., it MUST NOT contain values of different types.".
The following processors manipulate attribute values:
* https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/attributesprocessor
* https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/resourceprocessor
* https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/resourcedetectionprocessor
However they don't support arrays of primitive type.
When I try to do something like:
```
resource:
attributes:
- key: "aws.log.group.names"
action: insert
value: [sample-app-trace-logs]
```
I get back the following from the collector stdout:
```
Error: cannot build pipelines: failed to create "resource" processor, in pipeline "traces": error creating "resource" processor: error unsupported value type "[]interface {}"
```
### Describe the solution you'd like
I'd like that any type or processor that manipulate attributes to support arrays of primitive types.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
|
process
|
add support to attributes with array values in processors that manipulate attributes is your feature request related to a problem please describe that attribute values can be an array of primitive type values the array must be homogeneous i e it must not contain values of different types the following processors manipulate attribute values however they don t support arrays of primitive type when i try to do something like resource attributes key aws log group names action insert value i get back the following from the collector stdout error cannot build pipelines failed to create resource processor in pipeline traces error creating resource processor error unsupported value type interface describe the solution you d like i d like that any type or processor that manipulate attributes to support arrays of primitive types describe alternatives you ve considered no response additional context no response
| 1
|
791,706
| 27,873,231,058
|
IssuesEvent
|
2023-03-21 14:39:03
|
labring/laf
|
https://api.github.com/repos/labring/laf
|
closed
|
[bug] Dependency load error after custom port
|
priority:medium good first issue
|
### Search before asking
- [X] I had searched in the [issues](https://github.com/labring/laf/issues?q=is%3Aissue) and found no similar issues.
### Environment
Linux (self-host)
### What happened
Dependency load error after custom port

### How to reproduce
1.修改laf/deploy/build/charts/apisix/values.yaml
servicePort: 24444
containerPort: 24444
2.修改/laf/deploy/build 下的start.sh
MINIO_EXTERNAL_ENDPOINT="${HTTP_SCHEMA}://${MINIO_DOMAIN}:24444"
CASDOOR_ENDPOINT="${HTTP_SCHEMA}://${CASDOOR_HOST}:24444"
CASDOOR_REDIRECT_URI="${HTTP_SCHEMA}://${DOMAIN}:24444/login_callback"
helm install server -n ${NAMESPACE} \
--set databaseUrl=${DATABASE_URL} \
--set jwt.secret=${SERVER_JWT_SECRET} \
--set apiServerHost=api.${DOMAIN} \
--set apiServerUrl=${HTTP_SCHEMA}://api.${DOMAIN}:24444 \
--set casdoor.endpoint=${CASDOOR_ENDPOINT} \
--set casdoor.client_id=${CASDOOR_CLIENT_ID} \
--set casdoor.client_secret=${CASDOOR_CLIENT_SECRET} \
--set casdoor.redirect_uri=${CASDOOR_REDIRECT_URI} \
--set default_region.minio_domain=${MINIO_DOMAIN} \
--set default_region.minio_external_endpoint=${MINIO_EXTERNAL_ENDPOINT} \
--set default_region.minio_internal_endpoint=${MINIO_INTERNAL_ENDPOINT} \
--set default_region.minio_root_access_key=${MINIO_ROOT_ACCESS_KEY} \
--set default_region.minio_root_secret_key=${MINIO_ROOT_SECRET_KEY} \
--set default_region.runtime_domain=${DOMAIN}:24444 \
--set default_region.website_domain=site.${DOMAIN}:24444 \
--set default_region.tls=false \
--set default_region.apisix_api_url=${APISIX_API_URL} \
--set default_region.apisix_api_key=${APISIX_API_KEY} \
--set default_region.apisix_public_port=24444 \
3.执行start.sh 创建应用后云函数无法加载依赖

4.laf-server 日志

5.进入apisix-dashboard 删除端口后恢复正常

### Are you willing to submit PR?
- [ ] Yes I am willing to submit a PR!
|
1.0
|
[bug] Dependency load error after custom port - ### Search before asking
- [X] I had searched in the [issues](https://github.com/labring/laf/issues?q=is%3Aissue) and found no similar issues.
### Environment
Linux (self-host)
### What happened
Dependency load error after custom port

### How to reproduce
1.修改laf/deploy/build/charts/apisix/values.yaml
servicePort: 24444
containerPort: 24444
2.修改/laf/deploy/build 下的start.sh
MINIO_EXTERNAL_ENDPOINT="${HTTP_SCHEMA}://${MINIO_DOMAIN}:24444"
CASDOOR_ENDPOINT="${HTTP_SCHEMA}://${CASDOOR_HOST}:24444"
CASDOOR_REDIRECT_URI="${HTTP_SCHEMA}://${DOMAIN}:24444/login_callback"
helm install server -n ${NAMESPACE} \
--set databaseUrl=${DATABASE_URL} \
--set jwt.secret=${SERVER_JWT_SECRET} \
--set apiServerHost=api.${DOMAIN} \
--set apiServerUrl=${HTTP_SCHEMA}://api.${DOMAIN}:24444 \
--set casdoor.endpoint=${CASDOOR_ENDPOINT} \
--set casdoor.client_id=${CASDOOR_CLIENT_ID} \
--set casdoor.client_secret=${CASDOOR_CLIENT_SECRET} \
--set casdoor.redirect_uri=${CASDOOR_REDIRECT_URI} \
--set default_region.minio_domain=${MINIO_DOMAIN} \
--set default_region.minio_external_endpoint=${MINIO_EXTERNAL_ENDPOINT} \
--set default_region.minio_internal_endpoint=${MINIO_INTERNAL_ENDPOINT} \
--set default_region.minio_root_access_key=${MINIO_ROOT_ACCESS_KEY} \
--set default_region.minio_root_secret_key=${MINIO_ROOT_SECRET_KEY} \
--set default_region.runtime_domain=${DOMAIN}:24444 \
--set default_region.website_domain=site.${DOMAIN}:24444 \
--set default_region.tls=false \
--set default_region.apisix_api_url=${APISIX_API_URL} \
--set default_region.apisix_api_key=${APISIX_API_KEY} \
--set default_region.apisix_public_port=24444 \
3.执行start.sh 创建应用后云函数无法加载依赖

4.laf-server 日志

5.进入apisix-dashboard 删除端口后恢复正常

### Are you willing to submit PR?
- [ ] Yes I am willing to submit a PR!
|
non_process
|
dependency load error after custom port search before asking i had searched in the and found no similar issues environment linux self host what happened dependency load error after custom port how to reproduce 修改laf deploy build charts apisix values yaml serviceport containerport 修改 laf deploy build 下的start sh minio external endpoint http schema minio domain casdoor endpoint http schema casdoor host casdoor redirect uri http schema domain login callback helm install server n namespace set databaseurl database url set jwt secret server jwt secret set apiserverhost api domain set apiserverurl http schema api domain set casdoor endpoint casdoor endpoint set casdoor client id casdoor client id set casdoor client secret casdoor client secret set casdoor redirect uri casdoor redirect uri set default region minio domain minio domain set default region minio external endpoint minio external endpoint set default region minio internal endpoint minio internal endpoint set default region minio root access key minio root access key set default region minio root secret key minio root secret key set default region runtime domain domain set default region website domain site domain set default region tls false set default region apisix api url apisix api url set default region apisix api key apisix api key set default region apisix public port 执行start sh 创建应用后云函数无法加载依赖 laf server 日志 进入apisix dashboard 删除端口后恢复正常 are you willing to submit pr yes i am willing to submit a pr
| 0
|
22,734
| 32,055,219,781
|
IssuesEvent
|
2023-09-24 02:00:10
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Fri, 22 Sep 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### Learning to Drive Anywhere
- **Authors:** Ruizhao Zhu, Peng Huang, Eshed Ohn-Bar, Venkatesh Saligrama
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Robotics (cs.RO)
- **Arxiv link:** https://arxiv.org/abs/2309.12295
- **Pdf link:** https://arxiv.org/pdf/2309.12295
- **Abstract**
Human drivers can seamlessly adapt their driving decisions across geographical locations with diverse conditions and rules of the road, e.g., left vs. right-hand traffic. In contrast, existing models for autonomous driving have been thus far only deployed within restricted operational domains, i.e., without accounting for varying driving behaviors across locations or model scalability. In this work, we propose AnyD, a single geographically-aware conditional imitation learning (CIL) model that can efficiently learn from heterogeneous and globally distributed data with dynamic environmental, traffic, and social characteristics. Our key insight is to introduce a high-capacity geo-location-based channel attention mechanism that effectively adapts to local nuances while also flexibly modeling similarities among regions in a data-driven manner. By optimizing a contrastive imitation objective, our proposed approach can efficiently scale across inherently imbalanced data distributions and location-dependent events. We demonstrate the benefits of our AnyD agent across multiple datasets, cities, and scalable deployment paradigms, i.e., centralized, semi-supervised, and distributed agent training. Specifically, AnyD outperforms CIL baselines by over 14% in open-loop evaluation and 30% in closed-loop testing on CARLA.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
### GenLayNeRF: Generalizable Layered Representations with 3D Model Alignment for Multi-Human View Synthesis
- **Authors:** Youssef Abdelkareem, Shady Shehata, Fakhri Karray
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.11627
- **Pdf link:** https://arxiv.org/pdf/2309.11627
- **Abstract**
Novel view synthesis (NVS) of multi-human scenes imposes challenges due to the complex inter-human occlusions. Layered representations handle the complexities by dividing the scene into multi-layered radiance fields, however, they are mainly constrained to per-scene optimization making them inefficient. Generalizable human view synthesis methods combine the pre-fitted 3D human meshes with image features to reach generalization, yet they are mainly designed to operate on single-human scenes. Another drawback is the reliance on multi-step optimization techniques for parametric pre-fitting of the 3D body models that suffer from misalignment with the images in sparse view settings causing hallucinations in synthesized views. In this work, we propose, GenLayNeRF, a generalizable layered scene representation for free-viewpoint rendering of multiple human subjects which requires no per-scene optimization and very sparse views as input. We divide the scene into multi-human layers anchored by the 3D body meshes. We then ensure pixel-level alignment of the body models with the input views through a novel end-to-end trainable module that carries out iterative parametric correction coupled with multi-view feature fusion to produce aligned 3D models. For NVS, we extract point-wise image-aligned and human-anchored features which are correlated and fused using self-attention and cross-attention modules. We augment low-level RGB values into the features with an attention-based RGB fusion module. To evaluate our approach, we construct two multi-human view synthesis datasets; DeepMultiSyn and ZJU-MultiHuman. The results indicate that our proposed approach outperforms generalizable and non-human per-scene NeRF methods while performing at par with layered per-scene methods without test time optimization.
### ORTexME: Occlusion-Robust Human Shape and Pose via Temporal Average Texture and Mesh Encoding
- **Authors:** Yu Cheng, Bo Wang, Robby T. Tan
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.12183
- **Pdf link:** https://arxiv.org/pdf/2309.12183
- **Abstract**
In 3D human shape and pose estimation from a monocular video, models trained with limited labeled data cannot generalize well to videos with occlusion, which is common in the wild videos. The recent human neural rendering approaches focusing on novel view synthesis initialized by the off-the-shelf human shape and pose methods have the potential to correct the initial human shape. However, the existing methods have some drawbacks such as, erroneous in handling occlusion, sensitive to inaccurate human segmentation, and ineffective loss computation due to the non-regularized opacity field. To address these problems, we introduce ORTexME, an occlusion-robust temporal method that utilizes temporal information from the input video to better regularize the occluded body parts. While our ORTexME is based on NeRF, to determine the reliable regions for the NeRF ray sampling, we utilize our novel average texture learning approach to learn the average appearance of a person, and to infer a mask based on the average texture. In addition, to guide the opacity-field updates in NeRF to suppress blur and noise, we propose the use of human body mesh. The quantitative evaluation demonstrates that our method achieves significant improvement on the challenging multi-person 3DPW dataset, where our method achieves 1.8 P-MPJPE error reduction. The SOTA rendering-based methods fail and enlarge the error up to 5.6 on the same dataset.
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Neural Image Compression Using Masked Sparse Visual Representation
- **Authors:** Wei Jiang, Wei Wang, Yue Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2309.11661
- **Pdf link:** https://arxiv.org/pdf/2309.11661
- **Abstract**
We study neural image compression based on the Sparse Visual Representation (SVR), where images are embedded into a discrete latent space spanned by learned visual codebooks. By sharing codebooks with the decoder, the encoder transfers integer codeword indices that are efficient and cross-platform robust, and the decoder retrieves the embedded latent feature using the indices for reconstruction. Previous SVR-based compression lacks effective mechanism for rate-distortion tradeoffs, where one can only pursue either high reconstruction quality or low transmission bitrate. We propose a Masked Adaptive Codebook learning (M-AdaCode) method that applies masks to the latent feature subspace to balance bitrate and reconstruction quality. A set of semantic-class-dependent basis codebooks are learned, which are weighted combined to generate a rich latent feature for high-quality reconstruction. The combining weights are adaptively derived from each input image, providing fidelity information with additional transmission costs. By masking out unimportant weights in the encoder and recovering them in the decoder, we can trade off reconstruction quality for transmission bits, and the masking rate controls the balance between bitrate and distortion. Experiments over the standard JPEG-AI dataset demonstrate the effectiveness of our M-AdaCode approach.
### CPR-Coach: Recognizing Composite Error Actions based on Single-class Training
- **Authors:** Shunli Wang, Qing Yu, Shuaibing Wang, Dingkang Yang, Liuzhen Su, Xiao Zhao, Haopeng Kuang, Peixuan Zhang, Peng Zhai, Lihua Zhang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.11718
- **Pdf link:** https://arxiv.org/pdf/2309.11718
- **Abstract**
The fine-grained medical action analysis task has received considerable attention from pattern recognition communities recently, but it faces the problems of data and algorithm shortage. Cardiopulmonary Resuscitation (CPR) is an essential skill in emergency treatment. Currently, the assessment of CPR skills mainly depends on dummies and trainers, leading to high training costs and low efficiency. For the first time, this paper constructs a vision-based system to complete error action recognition and skill assessment in CPR. Specifically, we define 13 types of single-error actions and 74 types of composite error actions during external cardiac compression and then develop a video dataset named CPR-Coach. By taking the CPR-Coach as a benchmark, this paper thoroughly investigates and compares the performance of existing action recognition models based on different data modalities. To solve the unavoidable Single-class Training & Multi-class Testing problem, we propose a humancognition-inspired framework named ImagineNet to improve the model's multierror recognition performance under restricted supervision. Extensive experiments verify the effectiveness of the framework. We hope this work could advance research toward fine-grained medical action analysis and skill assessment. The CPR-Coach dataset and the code of ImagineNet are publicly available on Github.
### TinyCLIP: CLIP Distillation via Affinity Mimicking and Weight Inheritance
- **Authors:** Kan Wu, Houwen Peng, Zhenghong Zhou, Bin Xiao, Mengchen Liu, Lu Yuan, Hong Xuan, Michael Valenzuela, Xi (Stephen)Chen, Xinggang Wang, Hongyang Chao, Han Hu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.12314
- **Pdf link:** https://arxiv.org/pdf/2309.12314
- **Abstract**
In this paper, we propose a novel cross-modal distillation method, called TinyCLIP, for large-scale language-image pre-trained models. The method introduces two core techniques: affinity mimicking and weight inheritance. Affinity mimicking explores the interaction between modalities during distillation, enabling student models to mimic teachers' behavior of learning cross-modal feature alignment in a visual-linguistic affinity space. Weight inheritance transmits the pre-trained weights from the teacher models to their student counterparts to improve distillation efficiency. Moreover, we extend the method into a multi-stage progressive distillation to mitigate the loss of informative weights during extreme compression. Comprehensive experiments demonstrate the efficacy of TinyCLIP, showing that it can reduce the size of the pre-trained CLIP ViT-B/32 by 50%, while maintaining comparable zero-shot performance. While aiming for comparable performance, distillation with weight inheritance can speed up the training by 1.4 - 7.8 $\times$ compared to training from scratch. Moreover, our TinyCLIP ViT-8M/16, trained on YFCC-15M, achieves an impressive zero-shot top-1 accuracy of 41.1% on ImageNet, surpassing the original CLIP ViT-B/16 by 3.5% while utilizing only 8.9% parameters. Finally, we demonstrate the good transferability of TinyCLIP in various downstream tasks. Code and models will be open-sourced at https://aka.ms/tinyclip.
## Keyword: RAW
### GenLayNeRF: Generalizable Layered Representations with 3D Model Alignment for Multi-Human View Synthesis
- **Authors:** Youssef Abdelkareem, Shady Shehata, Fakhri Karray
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.11627
- **Pdf link:** https://arxiv.org/pdf/2309.11627
- **Abstract**
Novel view synthesis (NVS) of multi-human scenes imposes challenges due to the complex inter-human occlusions. Layered representations handle the complexities by dividing the scene into multi-layered radiance fields, however, they are mainly constrained to per-scene optimization making them inefficient. Generalizable human view synthesis methods combine the pre-fitted 3D human meshes with image features to reach generalization, yet they are mainly designed to operate on single-human scenes. Another drawback is the reliance on multi-step optimization techniques for parametric pre-fitting of the 3D body models that suffer from misalignment with the images in sparse view settings causing hallucinations in synthesized views. In this work, we propose, GenLayNeRF, a generalizable layered scene representation for free-viewpoint rendering of multiple human subjects which requires no per-scene optimization and very sparse views as input. We divide the scene into multi-human layers anchored by the 3D body meshes. We then ensure pixel-level alignment of the body models with the input views through a novel end-to-end trainable module that carries out iterative parametric correction coupled with multi-view feature fusion to produce aligned 3D models. For NVS, we extract point-wise image-aligned and human-anchored features which are correlated and fused using self-attention and cross-attention modules. We augment low-level RGB values into the features with an attention-based RGB fusion module. To evaluate our approach, we construct two multi-human view synthesis datasets; DeepMultiSyn and ZJU-MultiHuman. The results indicate that our proposed approach outperforms generalizable and non-human per-scene NeRF methods while performing at par with layered per-scene methods without test time optimization.
### Exploiting CLIP-based Multi-modal Approach for Artwork Classification and Retrieval
- **Authors:** Alberto Baldrati, Marco Bertini, Tiberio Uricchio, Alberto Del Bimbo
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.12110
- **Pdf link:** https://arxiv.org/pdf/2309.12110
- **Abstract**
Given the recent advances in multimodal image pretraining where visual models trained with semantically dense textual supervision tend to have better generalization capabilities than those trained using categorical attributes or through unsupervised techniques, in this work we investigate how recent CLIP model can be applied in several tasks in artwork domain. We perform exhaustive experiments on the NoisyArt dataset which is a dataset of artwork images crawled from public resources on the web. On such dataset CLIP achieves impressive results on (zero-shot) classification and promising results in both artwork-to-artwork and description-to-artwork domain.
### ORTexME: Occlusion-Robust Human Shape and Pose via Temporal Average Texture and Mesh Encoding
- **Authors:** Yu Cheng, Bo Wang, Robby T. Tan
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.12183
- **Pdf link:** https://arxiv.org/pdf/2309.12183
- **Abstract**
In 3D human shape and pose estimation from a monocular video, models trained with limited labeled data cannot generalize well to videos with occlusion, which is common in the wild videos. The recent human neural rendering approaches focusing on novel view synthesis initialized by the off-the-shelf human shape and pose methods have the potential to correct the initial human shape. However, the existing methods have some drawbacks such as, erroneous in handling occlusion, sensitive to inaccurate human segmentation, and ineffective loss computation due to the non-regularized opacity field. To address these problems, we introduce ORTexME, an occlusion-robust temporal method that utilizes temporal information from the input video to better regularize the occluded body parts. While our ORTexME is based on NeRF, to determine the reliable regions for the NeRF ray sampling, we utilize our novel average texture learning approach to learn the average appearance of a person, and to infer a mask based on the average texture. In addition, to guide the opacity-field updates in NeRF to suppress blur and noise, we propose the use of human body mesh. The quantitative evaluation demonstrates that our method achieves significant improvement on the challenging multi-person 3DPW dataset, where our method achieves 1.8 P-MPJPE error reduction. The SOTA rendering-based methods fail and enlarge the error up to 5.6 on the same dataset.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Fri, 22 Sep 23 - ## Keyword: events
### Learning to Drive Anywhere
- **Authors:** Ruizhao Zhu, Peng Huang, Eshed Ohn-Bar, Venkatesh Saligrama
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Robotics (cs.RO)
- **Arxiv link:** https://arxiv.org/abs/2309.12295
- **Pdf link:** https://arxiv.org/pdf/2309.12295
- **Abstract**
Human drivers can seamlessly adapt their driving decisions across geographical locations with diverse conditions and rules of the road, e.g., left vs. right-hand traffic. In contrast, existing models for autonomous driving have been thus far only deployed within restricted operational domains, i.e., without accounting for varying driving behaviors across locations or model scalability. In this work, we propose AnyD, a single geographically-aware conditional imitation learning (CIL) model that can efficiently learn from heterogeneous and globally distributed data with dynamic environmental, traffic, and social characteristics. Our key insight is to introduce a high-capacity geo-location-based channel attention mechanism that effectively adapts to local nuances while also flexibly modeling similarities among regions in a data-driven manner. By optimizing a contrastive imitation objective, our proposed approach can efficiently scale across inherently imbalanced data distributions and location-dependent events. We demonstrate the benefits of our AnyD agent across multiple datasets, cities, and scalable deployment paradigms, i.e., centralized, semi-supervised, and distributed agent training. Specifically, AnyD outperforms CIL baselines by over 14% in open-loop evaluation and 30% in closed-loop testing on CARLA.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
### GenLayNeRF: Generalizable Layered Representations with 3D Model Alignment for Multi-Human View Synthesis
- **Authors:** Youssef Abdelkareem, Shady Shehata, Fakhri Karray
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.11627
- **Pdf link:** https://arxiv.org/pdf/2309.11627
- **Abstract**
Novel view synthesis (NVS) of multi-human scenes imposes challenges due to the complex inter-human occlusions. Layered representations handle the complexities by dividing the scene into multi-layered radiance fields, however, they are mainly constrained to per-scene optimization making them inefficient. Generalizable human view synthesis methods combine the pre-fitted 3D human meshes with image features to reach generalization, yet they are mainly designed to operate on single-human scenes. Another drawback is the reliance on multi-step optimization techniques for parametric pre-fitting of the 3D body models that suffer from misalignment with the images in sparse view settings causing hallucinations in synthesized views. In this work, we propose, GenLayNeRF, a generalizable layered scene representation for free-viewpoint rendering of multiple human subjects which requires no per-scene optimization and very sparse views as input. We divide the scene into multi-human layers anchored by the 3D body meshes. We then ensure pixel-level alignment of the body models with the input views through a novel end-to-end trainable module that carries out iterative parametric correction coupled with multi-view feature fusion to produce aligned 3D models. For NVS, we extract point-wise image-aligned and human-anchored features which are correlated and fused using self-attention and cross-attention modules. We augment low-level RGB values into the features with an attention-based RGB fusion module. To evaluate our approach, we construct two multi-human view synthesis datasets; DeepMultiSyn and ZJU-MultiHuman. The results indicate that our proposed approach outperforms generalizable and non-human per-scene NeRF methods while performing at par with layered per-scene methods without test time optimization.
### ORTexME: Occlusion-Robust Human Shape and Pose via Temporal Average Texture and Mesh Encoding
- **Authors:** Yu Cheng, Bo Wang, Robby T. Tan
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.12183
- **Pdf link:** https://arxiv.org/pdf/2309.12183
- **Abstract**
In 3D human shape and pose estimation from a monocular video, models trained with limited labeled data cannot generalize well to videos with occlusion, which is common in the wild videos. The recent human neural rendering approaches focusing on novel view synthesis initialized by the off-the-shelf human shape and pose methods have the potential to correct the initial human shape. However, the existing methods have some drawbacks such as, erroneous in handling occlusion, sensitive to inaccurate human segmentation, and ineffective loss computation due to the non-regularized opacity field. To address these problems, we introduce ORTexME, an occlusion-robust temporal method that utilizes temporal information from the input video to better regularize the occluded body parts. While our ORTexME is based on NeRF, to determine the reliable regions for the NeRF ray sampling, we utilize our novel average texture learning approach to learn the average appearance of a person, and to infer a mask based on the average texture. In addition, to guide the opacity-field updates in NeRF to suppress blur and noise, we propose the use of human body mesh. The quantitative evaluation demonstrates that our method achieves significant improvement on the challenging multi-person 3DPW dataset, where our method achieves 1.8 P-MPJPE error reduction. The SOTA rendering-based methods fail and enlarge the error up to 5.6 on the same dataset.
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Neural Image Compression Using Masked Sparse Visual Representation
- **Authors:** Wei Jiang, Wei Wang, Yue Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2309.11661
- **Pdf link:** https://arxiv.org/pdf/2309.11661
- **Abstract**
We study neural image compression based on the Sparse Visual Representation (SVR), where images are embedded into a discrete latent space spanned by learned visual codebooks. By sharing codebooks with the decoder, the encoder transfers integer codeword indices that are efficient and cross-platform robust, and the decoder retrieves the embedded latent feature using the indices for reconstruction. Previous SVR-based compression lacks effective mechanism for rate-distortion tradeoffs, where one can only pursue either high reconstruction quality or low transmission bitrate. We propose a Masked Adaptive Codebook learning (M-AdaCode) method that applies masks to the latent feature subspace to balance bitrate and reconstruction quality. A set of semantic-class-dependent basis codebooks are learned, which are weighted combined to generate a rich latent feature for high-quality reconstruction. The combining weights are adaptively derived from each input image, providing fidelity information with additional transmission costs. By masking out unimportant weights in the encoder and recovering them in the decoder, we can trade off reconstruction quality for transmission bits, and the masking rate controls the balance between bitrate and distortion. Experiments over the standard JPEG-AI dataset demonstrate the effectiveness of our M-AdaCode approach.
### CPR-Coach: Recognizing Composite Error Actions based on Single-class Training
- **Authors:** Shunli Wang, Qing Yu, Shuaibing Wang, Dingkang Yang, Liuzhen Su, Xiao Zhao, Haopeng Kuang, Peixuan Zhang, Peng Zhai, Lihua Zhang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.11718
- **Pdf link:** https://arxiv.org/pdf/2309.11718
- **Abstract**
The fine-grained medical action analysis task has received considerable attention from pattern recognition communities recently, but it faces the problems of data and algorithm shortage. Cardiopulmonary Resuscitation (CPR) is an essential skill in emergency treatment. Currently, the assessment of CPR skills mainly depends on dummies and trainers, leading to high training costs and low efficiency. For the first time, this paper constructs a vision-based system to complete error action recognition and skill assessment in CPR. Specifically, we define 13 types of single-error actions and 74 types of composite error actions during external cardiac compression and then develop a video dataset named CPR-Coach. By taking the CPR-Coach as a benchmark, this paper thoroughly investigates and compares the performance of existing action recognition models based on different data modalities. To solve the unavoidable Single-class Training & Multi-class Testing problem, we propose a humancognition-inspired framework named ImagineNet to improve the model's multierror recognition performance under restricted supervision. Extensive experiments verify the effectiveness of the framework. We hope this work could advance research toward fine-grained medical action analysis and skill assessment. The CPR-Coach dataset and the code of ImagineNet are publicly available on Github.
### TinyCLIP: CLIP Distillation via Affinity Mimicking and Weight Inheritance
- **Authors:** Kan Wu, Houwen Peng, Zhenghong Zhou, Bin Xiao, Mengchen Liu, Lu Yuan, Hong Xuan, Michael Valenzuela, Xi (Stephen)Chen, Xinggang Wang, Hongyang Chao, Han Hu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.12314
- **Pdf link:** https://arxiv.org/pdf/2309.12314
- **Abstract**
In this paper, we propose a novel cross-modal distillation method, called TinyCLIP, for large-scale language-image pre-trained models. The method introduces two core techniques: affinity mimicking and weight inheritance. Affinity mimicking explores the interaction between modalities during distillation, enabling student models to mimic teachers' behavior of learning cross-modal feature alignment in a visual-linguistic affinity space. Weight inheritance transmits the pre-trained weights from the teacher models to their student counterparts to improve distillation efficiency. Moreover, we extend the method into a multi-stage progressive distillation to mitigate the loss of informative weights during extreme compression. Comprehensive experiments demonstrate the efficacy of TinyCLIP, showing that it can reduce the size of the pre-trained CLIP ViT-B/32 by 50%, while maintaining comparable zero-shot performance. While aiming for comparable performance, distillation with weight inheritance can speed up the training by 1.4 - 7.8 $\times$ compared to training from scratch. Moreover, our TinyCLIP ViT-8M/16, trained on YFCC-15M, achieves an impressive zero-shot top-1 accuracy of 41.1% on ImageNet, surpassing the original CLIP ViT-B/16 by 3.5% while utilizing only 8.9% parameters. Finally, we demonstrate the good transferability of TinyCLIP in various downstream tasks. Code and models will be open-sourced at https://aka.ms/tinyclip.
## Keyword: RAW
### GenLayNeRF: Generalizable Layered Representations with 3D Model Alignment for Multi-Human View Synthesis
- **Authors:** Youssef Abdelkareem, Shady Shehata, Fakhri Karray
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.11627
- **Pdf link:** https://arxiv.org/pdf/2309.11627
- **Abstract**
Novel view synthesis (NVS) of multi-human scenes imposes challenges due to the complex inter-human occlusions. Layered representations handle the complexities by dividing the scene into multi-layered radiance fields, however, they are mainly constrained to per-scene optimization making them inefficient. Generalizable human view synthesis methods combine the pre-fitted 3D human meshes with image features to reach generalization, yet they are mainly designed to operate on single-human scenes. Another drawback is the reliance on multi-step optimization techniques for parametric pre-fitting of the 3D body models that suffer from misalignment with the images in sparse view settings causing hallucinations in synthesized views. In this work, we propose, GenLayNeRF, a generalizable layered scene representation for free-viewpoint rendering of multiple human subjects which requires no per-scene optimization and very sparse views as input. We divide the scene into multi-human layers anchored by the 3D body meshes. We then ensure pixel-level alignment of the body models with the input views through a novel end-to-end trainable module that carries out iterative parametric correction coupled with multi-view feature fusion to produce aligned 3D models. For NVS, we extract point-wise image-aligned and human-anchored features which are correlated and fused using self-attention and cross-attention modules. We augment low-level RGB values into the features with an attention-based RGB fusion module. To evaluate our approach, we construct two multi-human view synthesis datasets; DeepMultiSyn and ZJU-MultiHuman. The results indicate that our proposed approach outperforms generalizable and non-human per-scene NeRF methods while performing at par with layered per-scene methods without test time optimization.
### Exploiting CLIP-based Multi-modal Approach for Artwork Classification and Retrieval
- **Authors:** Alberto Baldrati, Marco Bertini, Tiberio Uricchio, Alberto Del Bimbo
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.12110
- **Pdf link:** https://arxiv.org/pdf/2309.12110
- **Abstract**
Given the recent advances in multimodal image pretraining where visual models trained with semantically dense textual supervision tend to have better generalization capabilities than those trained using categorical attributes or through unsupervised techniques, in this work we investigate how recent CLIP model can be applied in several tasks in artwork domain. We perform exhaustive experiments on the NoisyArt dataset which is a dataset of artwork images crawled from public resources on the web. On such dataset CLIP achieves impressive results on (zero-shot) classification and promising results in both artwork-to-artwork and description-to-artwork domain.
### ORTexME: Occlusion-Robust Human Shape and Pose via Temporal Average Texture and Mesh Encoding
- **Authors:** Yu Cheng, Bo Wang, Robby T. Tan
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.12183
- **Pdf link:** https://arxiv.org/pdf/2309.12183
- **Abstract**
In 3D human shape and pose estimation from a monocular video, models trained with limited labeled data cannot generalize well to videos with occlusion, which is common in the wild videos. The recent human neural rendering approaches focusing on novel view synthesis initialized by the off-the-shelf human shape and pose methods have the potential to correct the initial human shape. However, the existing methods have some drawbacks such as, erroneous in handling occlusion, sensitive to inaccurate human segmentation, and ineffective loss computation due to the non-regularized opacity field. To address these problems, we introduce ORTexME, an occlusion-robust temporal method that utilizes temporal information from the input video to better regularize the occluded body parts. While our ORTexME is based on NeRF, to determine the reliable regions for the NeRF ray sampling, we utilize our novel average texture learning approach to learn the average appearance of a person, and to infer a mask based on the average texture. In addition, to guide the opacity-field updates in NeRF to suppress blur and noise, we propose the use of human body mesh. The quantitative evaluation demonstrates that our method achieves significant improvement on the challenging multi-person 3DPW dataset, where our method achieves 1.8 P-MPJPE error reduction. The SOTA rendering-based methods fail and enlarge the error up to 5.6 on the same dataset.
## Keyword: raw image
There is no result
|
process
|
new submissions for fri sep keyword events learning to drive anywhere authors ruizhao zhu peng huang eshed ohn bar venkatesh saligrama subjects computer vision and pattern recognition cs cv artificial intelligence cs ai machine learning cs lg robotics cs ro arxiv link pdf link abstract human drivers can seamlessly adapt their driving decisions across geographical locations with diverse conditions and rules of the road e g left vs right hand traffic in contrast existing models for autonomous driving have been thus far only deployed within restricted operational domains i e without accounting for varying driving behaviors across locations or model scalability in this work we propose anyd a single geographically aware conditional imitation learning cil model that can efficiently learn from heterogeneous and globally distributed data with dynamic environmental traffic and social characteristics our key insight is to introduce a high capacity geo location based channel attention mechanism that effectively adapts to local nuances while also flexibly modeling similarities among regions in a data driven manner by optimizing a contrastive imitation objective our proposed approach can efficiently scale across inherently imbalanced data distributions and location dependent events we demonstrate the benefits of our anyd agent across multiple datasets cities and scalable deployment paradigms i e centralized semi supervised and distributed agent training specifically anyd outperforms cil baselines by over in open loop evaluation and in closed loop testing on carla keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb genlaynerf generalizable layered representations with model alignment for multi human view synthesis authors youssef abdelkareem shady shehata fakhri karray subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract novel view synthesis nvs of multi human scenes imposes challenges due to the complex inter human occlusions layered representations handle the complexities by dividing the scene into multi layered radiance fields however they are mainly constrained to per scene optimization making them inefficient generalizable human view synthesis methods combine the pre fitted human meshes with image features to reach generalization yet they are mainly designed to operate on single human scenes another drawback is the reliance on multi step optimization techniques for parametric pre fitting of the body models that suffer from misalignment with the images in sparse view settings causing hallucinations in synthesized views in this work we propose genlaynerf a generalizable layered scene representation for free viewpoint rendering of multiple human subjects which requires no per scene optimization and very sparse views as input we divide the scene into multi human layers anchored by the body meshes we then ensure pixel level alignment of the body models with the input views through a novel end to end trainable module that carries out iterative parametric correction coupled with multi view feature fusion to produce aligned models for nvs we extract point wise image aligned and human anchored features which are correlated and fused using self attention and cross attention modules we augment low level rgb values into the features with an attention based rgb fusion module to evaluate our approach we construct two multi human view synthesis datasets deepmultisyn and zju multihuman the results indicate that our proposed approach outperforms generalizable and non human per scene nerf methods while performing at par with layered per scene methods without test time optimization ortexme occlusion robust human shape and pose via temporal average texture and mesh encoding authors yu cheng bo wang robby t tan subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in human shape and pose estimation from a monocular video models trained with limited labeled data cannot generalize well to videos with occlusion which is common in the wild videos the recent human neural rendering approaches focusing on novel view synthesis initialized by the off the shelf human shape and pose methods have the potential to correct the initial human shape however the existing methods have some drawbacks such as erroneous in handling occlusion sensitive to inaccurate human segmentation and ineffective loss computation due to the non regularized opacity field to address these problems we introduce ortexme an occlusion robust temporal method that utilizes temporal information from the input video to better regularize the occluded body parts while our ortexme is based on nerf to determine the reliable regions for the nerf ray sampling we utilize our novel average texture learning approach to learn the average appearance of a person and to infer a mask based on the average texture in addition to guide the opacity field updates in nerf to suppress blur and noise we propose the use of human body mesh the quantitative evaluation demonstrates that our method achieves significant improvement on the challenging multi person dataset where our method achieves p mpjpe error reduction the sota rendering based methods fail and enlarge the error up to on the same dataset keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression neural image compression using masked sparse visual representation authors wei jiang wei wang yue chen subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract we study neural image compression based on the sparse visual representation svr where images are embedded into a discrete latent space spanned by learned visual codebooks by sharing codebooks with the decoder the encoder transfers integer codeword indices that are efficient and cross platform robust and the decoder retrieves the embedded latent feature using the indices for reconstruction previous svr based compression lacks effective mechanism for rate distortion tradeoffs where one can only pursue either high reconstruction quality or low transmission bitrate we propose a masked adaptive codebook learning m adacode method that applies masks to the latent feature subspace to balance bitrate and reconstruction quality a set of semantic class dependent basis codebooks are learned which are weighted combined to generate a rich latent feature for high quality reconstruction the combining weights are adaptively derived from each input image providing fidelity information with additional transmission costs by masking out unimportant weights in the encoder and recovering them in the decoder we can trade off reconstruction quality for transmission bits and the masking rate controls the balance between bitrate and distortion experiments over the standard jpeg ai dataset demonstrate the effectiveness of our m adacode approach cpr coach recognizing composite error actions based on single class training authors shunli wang qing yu shuaibing wang dingkang yang liuzhen su xiao zhao haopeng kuang peixuan zhang peng zhai lihua zhang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the fine grained medical action analysis task has received considerable attention from pattern recognition communities recently but it faces the problems of data and algorithm shortage cardiopulmonary resuscitation cpr is an essential skill in emergency treatment currently the assessment of cpr skills mainly depends on dummies and trainers leading to high training costs and low efficiency for the first time this paper constructs a vision based system to complete error action recognition and skill assessment in cpr specifically we define types of single error actions and types of composite error actions during external cardiac compression and then develop a video dataset named cpr coach by taking the cpr coach as a benchmark this paper thoroughly investigates and compares the performance of existing action recognition models based on different data modalities to solve the unavoidable single class training multi class testing problem we propose a humancognition inspired framework named imaginenet to improve the model s multierror recognition performance under restricted supervision extensive experiments verify the effectiveness of the framework we hope this work could advance research toward fine grained medical action analysis and skill assessment the cpr coach dataset and the code of imaginenet are publicly available on github tinyclip clip distillation via affinity mimicking and weight inheritance authors kan wu houwen peng zhenghong zhou bin xiao mengchen liu lu yuan hong xuan michael valenzuela xi stephen chen xinggang wang hongyang chao han hu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in this paper we propose a novel cross modal distillation method called tinyclip for large scale language image pre trained models the method introduces two core techniques affinity mimicking and weight inheritance affinity mimicking explores the interaction between modalities during distillation enabling student models to mimic teachers behavior of learning cross modal feature alignment in a visual linguistic affinity space weight inheritance transmits the pre trained weights from the teacher models to their student counterparts to improve distillation efficiency moreover we extend the method into a multi stage progressive distillation to mitigate the loss of informative weights during extreme compression comprehensive experiments demonstrate the efficacy of tinyclip showing that it can reduce the size of the pre trained clip vit b by while maintaining comparable zero shot performance while aiming for comparable performance distillation with weight inheritance can speed up the training by times compared to training from scratch moreover our tinyclip vit trained on yfcc achieves an impressive zero shot top accuracy of on imagenet surpassing the original clip vit b by while utilizing only parameters finally we demonstrate the good transferability of tinyclip in various downstream tasks code and models will be open sourced at keyword raw genlaynerf generalizable layered representations with model alignment for multi human view synthesis authors youssef abdelkareem shady shehata fakhri karray subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract novel view synthesis nvs of multi human scenes imposes challenges due to the complex inter human occlusions layered representations handle the complexities by dividing the scene into multi layered radiance fields however they are mainly constrained to per scene optimization making them inefficient generalizable human view synthesis methods combine the pre fitted human meshes with image features to reach generalization yet they are mainly designed to operate on single human scenes another drawback is the reliance on multi step optimization techniques for parametric pre fitting of the body models that suffer from misalignment with the images in sparse view settings causing hallucinations in synthesized views in this work we propose genlaynerf a generalizable layered scene representation for free viewpoint rendering of multiple human subjects which requires no per scene optimization and very sparse views as input we divide the scene into multi human layers anchored by the body meshes we then ensure pixel level alignment of the body models with the input views through a novel end to end trainable module that carries out iterative parametric correction coupled with multi view feature fusion to produce aligned models for nvs we extract point wise image aligned and human anchored features which are correlated and fused using self attention and cross attention modules we augment low level rgb values into the features with an attention based rgb fusion module to evaluate our approach we construct two multi human view synthesis datasets deepmultisyn and zju multihuman the results indicate that our proposed approach outperforms generalizable and non human per scene nerf methods while performing at par with layered per scene methods without test time optimization exploiting clip based multi modal approach for artwork classification and retrieval authors alberto baldrati marco bertini tiberio uricchio alberto del bimbo subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract given the recent advances in multimodal image pretraining where visual models trained with semantically dense textual supervision tend to have better generalization capabilities than those trained using categorical attributes or through unsupervised techniques in this work we investigate how recent clip model can be applied in several tasks in artwork domain we perform exhaustive experiments on the noisyart dataset which is a dataset of artwork images crawled from public resources on the web on such dataset clip achieves impressive results on zero shot classification and promising results in both artwork to artwork and description to artwork domain ortexme occlusion robust human shape and pose via temporal average texture and mesh encoding authors yu cheng bo wang robby t tan subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in human shape and pose estimation from a monocular video models trained with limited labeled data cannot generalize well to videos with occlusion which is common in the wild videos the recent human neural rendering approaches focusing on novel view synthesis initialized by the off the shelf human shape and pose methods have the potential to correct the initial human shape however the existing methods have some drawbacks such as erroneous in handling occlusion sensitive to inaccurate human segmentation and ineffective loss computation due to the non regularized opacity field to address these problems we introduce ortexme an occlusion robust temporal method that utilizes temporal information from the input video to better regularize the occluded body parts while our ortexme is based on nerf to determine the reliable regions for the nerf ray sampling we utilize our novel average texture learning approach to learn the average appearance of a person and to infer a mask based on the average texture in addition to guide the opacity field updates in nerf to suppress blur and noise we propose the use of human body mesh the quantitative evaluation demonstrates that our method achieves significant improvement on the challenging multi person dataset where our method achieves p mpjpe error reduction the sota rendering based methods fail and enlarge the error up to on the same dataset keyword raw image there is no result
| 1
|
6,088
| 8,949,629,179
|
IssuesEvent
|
2019-01-25 08:20:58
|
our-city-app/oca-backend
|
https://api.github.com/repos/our-city-app/oca-backend
|
closed
|
Failed to sync events
|
priority_major process_duplicate type_bug
|
```
2019-01-11 01:30:25.575 CET
process_cityapp_uitdatabank_events for service-c0e9e53e-0372-4c5a-9ae7-8a1fe788ace3@rogerth.at page 1 (/base/data/home/apps/e~rogerthat-server/20190109t140721.415280197299602092/solutions/common/cron/events/uitdatabank.py:73)
...
2019-01-11 01:30:26.478 CET
_get_uitdatabank_events_v2 failed with status_code 401 and content: (/base/data/home/apps/e~rogerthat-server/20190109t140721.415280197299602092/solutions/common/bizz/cityapp.py:159)
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>GlassFish Server Open Source Edition 3.1.2.2 - Error report</title><style type="text/css"><!--H1 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:22px;} H2 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:16px;} H3 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:14px;} BODY {font-family:Tahoma,Arial,sans-serif;color:black;background-color:white;} B {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;} P {font-family:Tahoma,Arial,sans-serif;background:white;color:black;font-size:12px;}A {color : black;}HR {color : #525D76;}--></style> </head><body><h1>HTTP Status 401 - Unauthorized</h1><hr/><p><b>type</b> Status report</p><p><b>message</b>Unauthorized</p><p><b>description</b>This request requires HTTP authentication (Unauthorized).</p><hr/><h3>GlassFish Server Open Source Edition 3.1.2.2</h3></body></html>
2019-01-11 01:30:26.478 CET
Make sure your credentials are correct. (/base/data/home/apps/e~rogerthat-server/20190109t140721.415280197299602092/add_1_monkey_patches.py:144)
None
```
|
1.0
|
Failed to sync events - ```
2019-01-11 01:30:25.575 CET
process_cityapp_uitdatabank_events for service-c0e9e53e-0372-4c5a-9ae7-8a1fe788ace3@rogerth.at page 1 (/base/data/home/apps/e~rogerthat-server/20190109t140721.415280197299602092/solutions/common/cron/events/uitdatabank.py:73)
...
2019-01-11 01:30:26.478 CET
_get_uitdatabank_events_v2 failed with status_code 401 and content: (/base/data/home/apps/e~rogerthat-server/20190109t140721.415280197299602092/solutions/common/bizz/cityapp.py:159)
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>GlassFish Server Open Source Edition 3.1.2.2 - Error report</title><style type="text/css"><!--H1 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:22px;} H2 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:16px;} H3 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:14px;} BODY {font-family:Tahoma,Arial,sans-serif;color:black;background-color:white;} B {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;} P {font-family:Tahoma,Arial,sans-serif;background:white;color:black;font-size:12px;}A {color : black;}HR {color : #525D76;}--></style> </head><body><h1>HTTP Status 401 - Unauthorized</h1><hr/><p><b>type</b> Status report</p><p><b>message</b>Unauthorized</p><p><b>description</b>This request requires HTTP authentication (Unauthorized).</p><hr/><h3>GlassFish Server Open Source Edition 3.1.2.2</h3></body></html>
2019-01-11 01:30:26.478 CET
Make sure your credentials are correct. (/base/data/home/apps/e~rogerthat-server/20190109t140721.415280197299602092/add_1_monkey_patches.py:144)
None
```
|
process
|
failed to sync events cet process cityapp uitdatabank events for service rogerth at page base data home apps e rogerthat server solutions common cron events uitdatabank py cet get uitdatabank events failed with status code and content base data home apps e rogerthat server solutions common bizz cityapp py http status unauthorized type status report message unauthorized description this request requires http authentication unauthorized glassfish server open source edition cet make sure your credentials are correct base data home apps e rogerthat server add monkey patches py none
| 1
|
93,059
| 8,392,803,221
|
IssuesEvent
|
2018-10-09 18:40:15
|
igsis/promac
|
https://api.github.com/repos/igsis/promac
|
closed
|
SMC - Label Adm - Erro Notas
|
Teste OK
|
Warning: mysqli_num_rows() expects parameter 1 to be mysqli_result, boolean given in C:\xampp\htdocs\promac\perfil\includes\label_smc_adm.php on line 144
|
1.0
|
SMC - Label Adm - Erro Notas - Warning: mysqli_num_rows() expects parameter 1 to be mysqli_result, boolean given in C:\xampp\htdocs\promac\perfil\includes\label_smc_adm.php on line 144
|
non_process
|
smc label adm erro notas warning mysqli num rows expects parameter to be mysqli result boolean given in c xampp htdocs promac perfil includes label smc adm php on line
| 0
|
42,539
| 12,894,381,276
|
IssuesEvent
|
2020-07-14 00:10:19
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Allow user to search saved objects across Spaces
|
Feature:Security/Spaces Team:Core UI Team:Security enhancement loe:multi-week
|
# Search across spaces
## Scope
MVP is only concerned with the `_find` operation for saved objects. All other APIs and operations will continue to be scoped to the user's current space.
## Approach 1: Expose `namespaces` as a query parameter
- /api/saved_objects/_find?type=index-pattern&namespaces=*
| `/_find` | OSS | Spaces | Security | Spaces + Security
| ------- | -- | -- | -- | -- |
| No namespaces specified | no changes (search within `default` namespace)| search within user's active space | no changes (authorize operation globally: * resource) |no changes: Spaces: search within user's active space Security: authorize operation against user's active space |
| namespaces=foo | search within `foo` namespace. This isn't something we should normally do, however.| search within `foo` space, instead of user's active space. |Authorize operation against foo resource |Spaces: search within foo space Security: authorize operation against foo resource |
| namespaces=foo,bar | search within both `foo` and `bar` namespaces.| search within `foo` and `bar` spaces, instead of user's active space |Authorize operation against both foo and bar resources |Spaces: search within foo and bar spaces Security: authorize operation against both foo and bar resources |
| namespaces=* | Options: <ul><li>reject operation</li><li>ignore parameter, search as if it wasn't specified</li><li>search across all namespaces</li></ul>| Steps: <ol><li>query for all spaces</li><li>search within all available spaces instead of user's active space.</li></ol> |Authorize operation globally: * resource |1) query for all spaces where authorized to find the requested types 2) search within all authorized spaces instead of user's active space |
## Approach 2: Denote which APIs can support a multi-space operation:
- /s/*/api/saved_objects/_find?foo=bar
**Not currently exploring this option.** We would require a way for routes to opt-into this behavior in a way that makes sense for OSS too. The only current mechanism is the `tags` route option, which isn't ideal.
## Complications
- Different saved objects can share an ID across spaces. How to resolve this?
- If we can wait for Phase 3? of sharing saved objects between spaces, then this problem goes away
- If not, then we can return these "duplicate" objects, but denote which `namespaces` they exist in.
|
True
|
Allow user to search saved objects across Spaces - # Search across spaces
## Scope
MVP is only concerned with the `_find` operation for saved objects. All other APIs and operations will continue to be scoped to the user's current space.
## Approach 1: Expose `namespaces` as a query parameter
- /api/saved_objects/_find?type=index-pattern&namespaces=*
| `/_find` | OSS | Spaces | Security | Spaces + Security
| ------- | -- | -- | -- | -- |
| No namespaces specified | no changes (search within `default` namespace)| search within user's active space | no changes (authorize operation globally: * resource) |no changes: Spaces: search within user's active space Security: authorize operation against user's active space |
| namespaces=foo | search within `foo` namespace. This isn't something we should normally do, however.| search within `foo` space, instead of user's active space. |Authorize operation against foo resource |Spaces: search within foo space Security: authorize operation against foo resource |
| namespaces=foo,bar | search within both `foo` and `bar` namespaces.| search within `foo` and `bar` spaces, instead of user's active space |Authorize operation against both foo and bar resources |Spaces: search within foo and bar spaces Security: authorize operation against both foo and bar resources |
| namespaces=* | Options: <ul><li>reject operation</li><li>ignore parameter, search as if it wasn't specified</li><li>search across all namespaces</li></ul>| Steps: <ol><li>query for all spaces</li><li>search within all available spaces instead of user's active space.</li></ol> |Authorize operation globally: * resource |1) query for all spaces where authorized to find the requested types 2) search within all authorized spaces instead of user's active space |
## Approach 2: Denote which APIs can support a multi-space operation:
- /s/*/api/saved_objects/_find?foo=bar
**Not currently exploring this option.** We would require a way for routes to opt-into this behavior in a way that makes sense for OSS too. The only current mechanism is the `tags` route option, which isn't ideal.
## Complications
- Different saved objects can share an ID across spaces. How to resolve this?
- If we can wait for Phase 3? of sharing saved objects between spaces, then this problem goes away
- If not, then we can return these "duplicate" objects, but denote which `namespaces` they exist in.
|
non_process
|
allow user to search saved objects across spaces search across spaces scope mvp is only concerned with the find operation for saved objects all other apis and operations will continue to be scoped to the user s current space approach expose namespaces as a query parameter api saved objects find type index pattern namespaces find oss spaces security spaces security no namespaces specified no changes search within default namespace search within user s active space no changes authorize operation globally resource no changes spaces search within user s active space security authorize operation against user s active space namespaces foo search within foo namespace this isn t something we should normally do however search within foo space instead of user s active space authorize operation against foo resource spaces search within foo space security authorize operation against foo resource namespaces foo bar search within both foo and bar namespaces search within foo and bar spaces instead of user s active space authorize operation against both foo and bar resources spaces search within foo and bar spaces security authorize operation against both foo and bar resources namespaces options reject operation ignore parameter search as if it wasn t specified search across all namespaces steps query for all spaces search within all available spaces instead of user s active space authorize operation globally resource query for all spaces where authorized to find the requested types search within all authorized spaces instead of user s active space approach denote which apis can support a multi space operation s api saved objects find foo bar not currently exploring this option we would require a way for routes to opt into this behavior in a way that makes sense for oss too the only current mechanism is the tags route option which isn t ideal complications different saved objects can share an id across spaces how to resolve this if we can wait for phase of sharing saved objects between spaces then this problem goes away if not then we can return these duplicate objects but denote which namespaces they exist in
| 0
|
15,272
| 19,251,918,792
|
IssuesEvent
|
2021-12-09 06:48:41
|
ppy/osu-web
|
https://api.github.com/repos/ppy/osu-web
|
closed
|
Graveyarded map has missing metadata and difficulty information
|
area:beatmap-processing
|
Originally spotted at https://github.com/ppy/osu/discussions/15908.
See the following beatmap set: https://osu.ppy.sh/beatmapsets/711461
Two difficulties on this beatmap set (https://osu.ppy.sh/beatmapsets/711461#osu/3104419 and https://osu.ppy.sh/beatmapsets/711461#osu/3104450) have no difficulty name, zero star rating, zero duration, no BPM/object count info and no difficulty information:

<details>
<summary>Extracted json for one of the offending beatmaps</summary>
```json
{
"beatmapset_id": 711461,
"difficulty_rating": 0,
"id": 3104419,
"mode": "osu",
"status": "graveyard",
"total_length": 0,
"user_id": 10652548,
"version": "",
"accuracy": 0,
"ar": 0,
"bpm": null,
"convert": false,
"count_circles": 0,
"count_sliders": 0,
"count_spinners": 0,
"cs": 0,
"deleted_at": null,
"drain": 0,
"hit_length": 0,
"is_scoreable": false,
"last_updated": "2021-07-11T00:07:04+00:00",
"mode_int": 0,
"passcount": 0,
"playcount": 0,
"ranked": -2,
"url": "https://osu.ppy.sh/beatmaps/3104419",
"checksum": null,
"failtimes": {
"exit": [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
],
"fail": [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
]
},
"max_combo": null
}
```
</details>
Particularly of note is the `null` bpm which is causing client-side parsing failures. Not sure whether that is something that should be handled client-side or an actual failure, which is why I'm opening this as an issue.
|
1.0
|
Graveyarded map has missing metadata and difficulty information - Originally spotted at https://github.com/ppy/osu/discussions/15908.
See the following beatmap set: https://osu.ppy.sh/beatmapsets/711461
Two difficulties on this beatmap set (https://osu.ppy.sh/beatmapsets/711461#osu/3104419 and https://osu.ppy.sh/beatmapsets/711461#osu/3104450) have no difficulty name, zero star rating, zero duration, no BPM/object count info and no difficulty information:

<details>
<summary>Extracted json for one of the offending beatmaps</summary>
```json
{
"beatmapset_id": 711461,
"difficulty_rating": 0,
"id": 3104419,
"mode": "osu",
"status": "graveyard",
"total_length": 0,
"user_id": 10652548,
"version": "",
"accuracy": 0,
"ar": 0,
"bpm": null,
"convert": false,
"count_circles": 0,
"count_sliders": 0,
"count_spinners": 0,
"cs": 0,
"deleted_at": null,
"drain": 0,
"hit_length": 0,
"is_scoreable": false,
"last_updated": "2021-07-11T00:07:04+00:00",
"mode_int": 0,
"passcount": 0,
"playcount": 0,
"ranked": -2,
"url": "https://osu.ppy.sh/beatmaps/3104419",
"checksum": null,
"failtimes": {
"exit": [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
],
"fail": [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
]
},
"max_combo": null
}
```
</details>
Particularly of note is the `null` bpm which is causing client-side parsing failures. Not sure whether that is something that should be handled client-side or an actual failure, which is why I'm opening this as an issue.
|
process
|
graveyarded map has missing metadata and difficulty information originally spotted at see the following beatmap set two difficulties on this beatmap set and have no difficulty name zero star rating zero duration no bpm object count info and no difficulty information extracted json for one of the offending beatmaps json beatmapset id difficulty rating id mode osu status graveyard total length user id version accuracy ar bpm null convert false count circles count sliders count spinners cs deleted at null drain hit length is scoreable false last updated mode int passcount playcount ranked url checksum null failtimes exit fail max combo null particularly of note is the null bpm which is causing client side parsing failures not sure whether that is something that should be handled client side or an actual failure which is why i m opening this as an issue
| 1
|
15,380
| 19,562,578,332
|
IssuesEvent
|
2022-01-03 18:21:42
|
2i2c-org/infrastructure
|
https://api.github.com/repos/2i2c-org/infrastructure
|
opened
|
[discuss] Should we change our policy for self-merging `infrastructure/` PRs?
|
:label: team-process
|
In our Review / Merge policy, we note that PRs should be merged *by the PR author* (https://infrastructure.2i2c.org/en/latest/contributing/code-review.html#changing-active-infrastructure-in-general). This is because we assume the PR author has the most context for what infrastructure might change.
However, this also puts up a barrier towards getting PRs merged in general, and we have had a few instances where a PR had 1+ approvals, but was not merged. In general, we probably want to encourage merges if they have approvals, so here are a few thoughts:
### Two ideas
- Loosen the restriction around who can merge a PR in `infrastructure/` to define a more strict subset that requires a self-merge. E.g., "If a PR will be deployed via CI/CD, anybody can merge it, but if a PR requires a manual deploy, the PR author must merge it"
- Come up with a team practice that encourages that PR author to merge things in more quickly after there are approvals, to discourage the existence of dangling PRs that are just waiting for someone to press the button.
ref #908
|
1.0
|
[discuss] Should we change our policy for self-merging `infrastructure/` PRs? - In our Review / Merge policy, we note that PRs should be merged *by the PR author* (https://infrastructure.2i2c.org/en/latest/contributing/code-review.html#changing-active-infrastructure-in-general). This is because we assume the PR author has the most context for what infrastructure might change.
However, this also puts up a barrier towards getting PRs merged in general, and we have had a few instances where a PR had 1+ approvals, but was not merged. In general, we probably want to encourage merges if they have approvals, so here are a few thoughts:
### Two ideas
- Loosen the restriction around who can merge a PR in `infrastructure/` to define a more strict subset that requires a self-merge. E.g., "If a PR will be deployed via CI/CD, anybody can merge it, but if a PR requires a manual deploy, the PR author must merge it"
- Come up with a team practice that encourages that PR author to merge things in more quickly after there are approvals, to discourage the existence of dangling PRs that are just waiting for someone to press the button.
ref #908
|
process
|
should we change our policy for self merging infrastructure prs in our review merge policy we note that prs should be merged by the pr author this is because we assume the pr author has the most context for what infrastructure might change however this also puts up a barrier towards getting prs merged in general and we have had a few instances where a pr had approvals but was not merged in general we probably want to encourage merges if they have approvals so here are a few thoughts two ideas loosen the restriction around who can merge a pr in infrastructure to define a more strict subset that requires a self merge e g if a pr will be deployed via ci cd anybody can merge it but if a pr requires a manual deploy the pr author must merge it come up with a team practice that encourages that pr author to merge things in more quickly after there are approvals to discourage the existence of dangling prs that are just waiting for someone to press the button ref
| 1
|
256,093
| 27,552,618,171
|
IssuesEvent
|
2023-03-07 15:51:31
|
BrianMcDonaldWS/genie
|
https://api.github.com/repos/BrianMcDonaldWS/genie
|
opened
|
CVE-2022-38751 (Medium) detected in snakeyaml-1.25.jar
|
security vulnerability
|
## CVE-2022-38751 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.25.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /genie-agent/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.25/8b6e01ef661d8378ae6dd7b511a7f2a33fae1421/snakeyaml-1.25.jar,/root/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.25/8b6e01ef661d8378ae6dd7b511a7f2a33fae1421/snakeyaml-1.25.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-2.2.2.RELEASE.jar (Root Library)
- spring-boot-starter-2.2.2.RELEASE.jar
- :x: **snakeyaml-1.25.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38751>CVE-2022-38751</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039</a></p>
<p>Release Date: 2022-09-05</p>
<p>Fix Resolution (org.yaml:snakeyaml): 1.31</p>
<p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-actuator): 2.6.9</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
True
|
CVE-2022-38751 (Medium) detected in snakeyaml-1.25.jar - ## CVE-2022-38751 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.25.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /genie-agent/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.25/8b6e01ef661d8378ae6dd7b511a7f2a33fae1421/snakeyaml-1.25.jar,/root/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.25/8b6e01ef661d8378ae6dd7b511a7f2a33fae1421/snakeyaml-1.25.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-2.2.2.RELEASE.jar (Root Library)
- spring-boot-starter-2.2.2.RELEASE.jar
- :x: **snakeyaml-1.25.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38751>CVE-2022-38751</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039</a></p>
<p>Release Date: 2022-09-05</p>
<p>Fix Resolution (org.yaml:snakeyaml): 1.31</p>
<p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-actuator): 2.6.9</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
non_process
|
cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file genie agent build gradle path to vulnerable library root gradle caches modules files org yaml snakeyaml snakeyaml jar root gradle caches modules files org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter actuator release jar root library spring boot starter release jar x snakeyaml jar vulnerable library vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml direct dependency fix resolution org springframework boot spring boot starter actuator check this box to open an automated fix pr
| 0
|
124,490
| 10,312,994,546
|
IssuesEvent
|
2019-08-29 21:19:00
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
[helm] [helm-controller] failed with : helm template failed. Error: stat template-dir: no such file or directory
|
[zube]: To Test area/catalog internal kind/bug team/az
|
<!--
Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue
For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase.
-->
**What kind of request is this (question/bug/enhancement/feature request):**
bug
**Steps to reproduce (least amount of steps as possible):**
1. add helm repo via rancher UI
repo: https://raw.githubusercontent.com/jasonsoft/helm/master/
name: jasonsoft
level: global
2. lauch app from category. Search "whitemos" and click "view detail" button
3. just click "lauch" button
**Result:**
got an error message on the page.
`Helm template failed. Error: stat template-dir: no such file or directory : exit status 1`
**Other details that may be helpful:**
If we use helm cli to install the package via console, it works.
error message from rancher
```
2019/03/03 06:19:02 [ERROR] AppController p-xrqw7/whitemos-md4q2 [helm-controller] failed with : helm template failed. Error: stat template-dir: no such file or directory
```
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI):
- Installation option (single install/HA):
`Rancher/Rancher: v2.2.0-rc2, single node`
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported):
- Machine type (cloud/VM/metal) and specifications (CPU/memory):
- Kubernetes version (use `kubectl version`):
Cluster type: imported ( cluster created via kubeadm)
Machine type: vm
```
Client Version: version.Info{Major:"1", Minor:"12", GitVersion:"v1.12.6", GitCommit:"ab91afd7062d4240e95e51ac00a18bd58fddd365", GitTreeState:"clean", BuildDate:"2019-02-26T12:59:46Z", GoVersion:"go1.10.8", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"12", GitVersion:"v1.12.6", GitCommit:"ab91afd7062d4240e95e51ac00a18bd58fddd365", GitTreeState:"clean", BuildDate:"2019-02-26T12:49:28Z", GoVersion:"go1.10.8", Compiler:"gc", Platform:"linux/amd64"}
```
- Docker version (use `docker version`):
```
Client:
Version: 18.06.3-ce
API version: 1.38
Go version: go1.10.3
Git commit: d7080c1
Built: Wed Feb 20 02:26:51 2019
OS/Arch: linux/amd64
Experimental: false
Server:
Engine:
Version: 18.06.3-ce
API version: 1.38 (minimum version 1.12)
Go version: go1.10.3
Git commit: d7080c1
Built: Wed Feb 20 02:28:17 2019
OS/Arch: linux/amd64
Experimental: false
```
|
1.0
|
[helm] [helm-controller] failed with : helm template failed. Error: stat template-dir: no such file or directory - <!--
Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue
For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase.
-->
**What kind of request is this (question/bug/enhancement/feature request):**
bug
**Steps to reproduce (least amount of steps as possible):**
1. add helm repo via rancher UI
repo: https://raw.githubusercontent.com/jasonsoft/helm/master/
name: jasonsoft
level: global
2. lauch app from category. Search "whitemos" and click "view detail" button
3. just click "lauch" button
**Result:**
got an error message on the page.
`Helm template failed. Error: stat template-dir: no such file or directory : exit status 1`
**Other details that may be helpful:**
If we use helm cli to install the package via console, it works.
error message from rancher
```
2019/03/03 06:19:02 [ERROR] AppController p-xrqw7/whitemos-md4q2 [helm-controller] failed with : helm template failed. Error: stat template-dir: no such file or directory
```
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI):
- Installation option (single install/HA):
`Rancher/Rancher: v2.2.0-rc2, single node`
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported):
- Machine type (cloud/VM/metal) and specifications (CPU/memory):
- Kubernetes version (use `kubectl version`):
Cluster type: imported ( cluster created via kubeadm)
Machine type: vm
```
Client Version: version.Info{Major:"1", Minor:"12", GitVersion:"v1.12.6", GitCommit:"ab91afd7062d4240e95e51ac00a18bd58fddd365", GitTreeState:"clean", BuildDate:"2019-02-26T12:59:46Z", GoVersion:"go1.10.8", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"12", GitVersion:"v1.12.6", GitCommit:"ab91afd7062d4240e95e51ac00a18bd58fddd365", GitTreeState:"clean", BuildDate:"2019-02-26T12:49:28Z", GoVersion:"go1.10.8", Compiler:"gc", Platform:"linux/amd64"}
```
- Docker version (use `docker version`):
```
Client:
Version: 18.06.3-ce
API version: 1.38
Go version: go1.10.3
Git commit: d7080c1
Built: Wed Feb 20 02:26:51 2019
OS/Arch: linux/amd64
Experimental: false
Server:
Engine:
Version: 18.06.3-ce
API version: 1.38 (minimum version 1.12)
Go version: go1.10.3
Git commit: d7080c1
Built: Wed Feb 20 02:28:17 2019
OS/Arch: linux/amd64
Experimental: false
```
|
non_process
|
failed with helm template failed error stat template dir no such file or directory please search for existing issues first then read to see what we expect in an issue for security issues please email security rancher com instead of posting a public issue in github you may but are not required to use the gpg key located on keybase what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible add helm repo via rancher ui repo name jasonsoft level global lauch app from category search whitemos and click view detail button just click lauch button result got an error message on the page helm template failed error stat template dir no such file or directory exit status other details that may be helpful if we use helm cli to install the package via console it works error message from rancher appcontroller p whitemos failed with helm template failed error stat template dir no such file or directory environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui installation option single install ha rancher rancher single node if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported machine type cloud vm metal and specifications cpu memory kubernetes version use kubectl version cluster type imported cluster created via kubeadm machine type vm client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux docker version use docker version client version ce api version go version git commit built wed feb os arch linux experimental false server engine version ce api version minimum version go version git commit built wed feb os arch linux experimental false
| 0
|
292,927
| 8,970,743,426
|
IssuesEvent
|
2019-01-29 14:23:50
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
thegoodcaster.com - site is not usable
|
browser-firefox priority-normal
|
<!-- @browser: Firefox 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:66.0) Gecko/20100101 Firefox/66.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://thegoodcaster.com/redirect/57a764d042bf8
**Browser / Version**: Firefox 66.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: block it
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190124173137</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
thegoodcaster.com - site is not usable - <!-- @browser: Firefox 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:66.0) Gecko/20100101 Firefox/66.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://thegoodcaster.com/redirect/57a764d042bf8
**Browser / Version**: Firefox 66.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: block it
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190124173137</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
thegoodcaster com site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description block it steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel aurora from with ❤️
| 0
|
9,117
| 12,195,377,000
|
IssuesEvent
|
2020-04-29 17:16:20
|
pacificclimate/climate-explorer-data-prep
|
https://api.github.com/repos/pacificclimate/climate-explorer-data-prep
|
opened
|
Incorrectly calculated frost free day data
|
process new data update existing data
|
Frost free day data (`ffd`) was calculated from frost day (`fdETCCDI`) data via (365 - `fdETCCDI`), which is correct for annual data.
Unfortunately, monthly and seasonal data was also calculated this way, and it is _quite_ incorrect. It's not being used for anything at present. Remove or replace before it is.
|
1.0
|
Incorrectly calculated frost free day data - Frost free day data (`ffd`) was calculated from frost day (`fdETCCDI`) data via (365 - `fdETCCDI`), which is correct for annual data.
Unfortunately, monthly and seasonal data was also calculated this way, and it is _quite_ incorrect. It's not being used for anything at present. Remove or replace before it is.
|
process
|
incorrectly calculated frost free day data frost free day data ffd was calculated from frost day fdetccdi data via fdetccdi which is correct for annual data unfortunately monthly and seasonal data was also calculated this way and it is quite incorrect it s not being used for anything at present remove or replace before it is
| 1
|
175,835
| 13,612,531,853
|
IssuesEvent
|
2020-09-23 10:26:16
|
softmatterlab/Braph-2.0-Matlab
|
https://api.github.com/repos/softmatterlab/Braph-2.0-Matlab
|
closed
|
Implement Average Overlapping Out Degree
|
measure test
|
- [x] Implement OverlappingOutDegreeAv.m
- [x] Implement test_OverlappingOutDegreeAv.m
Branch from Develop
Look how is structured OverlappingDegreeAv.m and OverlappingDegree.m
|
1.0
|
Implement Average Overlapping Out Degree - - [x] Implement OverlappingOutDegreeAv.m
- [x] Implement test_OverlappingOutDegreeAv.m
Branch from Develop
Look how is structured OverlappingDegreeAv.m and OverlappingDegree.m
|
non_process
|
implement average overlapping out degree implement overlappingoutdegreeav m implement test overlappingoutdegreeav m branch from develop look how is structured overlappingdegreeav m and overlappingdegree m
| 0
|
8,500
| 11,683,012,026
|
IssuesEvent
|
2020-03-05 01:59:33
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `Left` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `Left` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `Left` from TiDB -
## Description
Port the scalar function `Left` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function left from tidb description port the scalar function left from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
2,754
| 5,672,158,687
|
IssuesEvent
|
2017-04-12 00:11:17
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
opened
|
ENOENT error message inconsistencies
|
errors fs libuv process
|
* **Version**: 8.0.0-rc.0
* **Platform**: Windows 7 x64
* **Subsystem**: errors, fs, libuv, process
Currently, most of the ENOENT error messages have a similar signature:
Error: ENOENT: no such file or directory, [call] '/full/path/to/filename'
For example:
```js
fs.accessSync('does-not-exists'); // etc...
```
```
Error: ENOENT: no such file or directory, access '...does-not-exists'
Error: ENOENT: no such file or directory, chmod '...does-not-exists'
Error: ENOENT: no such file or directory, link '...does-not-exists' -> '...does-not-exists2'
Error: ENOENT: no such file or directory, lstat '...does-not-exists'
Error: ENOENT: no such file or directory, open '...does-not-exists'
Error: ENOENT: no such file or directory, readlink '...does-not-exists'
Error: ENOENT: no such file or directory, rename '...does-not-exists' -> '...does-not-exists2'
Error: ENOENT: no such file or directory, rmdir '...does-not-exists'
Error: ENOENT: no such file or directory, scandir '...does-not-exists'
Error: ENOENT: no such file or directory, stat '...does-not-exists'
Error: ENOENT: no such file or directory, unlink '...does-not-exists'
Error: ENOENT: no such file or directory, utime '...does-not-exists'
```
I've stumbled upon 2 cases that have different and somehow confusing signatures:
1. Different order, no description, filename without quotes, no full path:
```js
fs.watch('does-not-exists');
// Error: watch does-not-exists ENOENT
```
2. No file name / path at all (the call name can be wrongly considered as the filename):
```js
process.chdir('does-not-exists');
// Error: ENOENT: no such file or directory, uv_chdir
```
Are these cases worth unification? Can they be addressed in Node.js or they are libuv features?
|
1.0
|
ENOENT error message inconsistencies - * **Version**: 8.0.0-rc.0
* **Platform**: Windows 7 x64
* **Subsystem**: errors, fs, libuv, process
Currently, most of the ENOENT error messages have a similar signature:
Error: ENOENT: no such file or directory, [call] '/full/path/to/filename'
For example:
```js
fs.accessSync('does-not-exists'); // etc...
```
```
Error: ENOENT: no such file or directory, access '...does-not-exists'
Error: ENOENT: no such file or directory, chmod '...does-not-exists'
Error: ENOENT: no such file or directory, link '...does-not-exists' -> '...does-not-exists2'
Error: ENOENT: no such file or directory, lstat '...does-not-exists'
Error: ENOENT: no such file or directory, open '...does-not-exists'
Error: ENOENT: no such file or directory, readlink '...does-not-exists'
Error: ENOENT: no such file or directory, rename '...does-not-exists' -> '...does-not-exists2'
Error: ENOENT: no such file or directory, rmdir '...does-not-exists'
Error: ENOENT: no such file or directory, scandir '...does-not-exists'
Error: ENOENT: no such file or directory, stat '...does-not-exists'
Error: ENOENT: no such file or directory, unlink '...does-not-exists'
Error: ENOENT: no such file or directory, utime '...does-not-exists'
```
I've stumbled upon 2 cases that have different and somehow confusing signatures:
1. Different order, no description, filename without quotes, no full path:
```js
fs.watch('does-not-exists');
// Error: watch does-not-exists ENOENT
```
2. No file name / path at all (the call name can be wrongly considered as the filename):
```js
process.chdir('does-not-exists');
// Error: ENOENT: no such file or directory, uv_chdir
```
Are these cases worth unification? Can they be addressed in Node.js or they are libuv features?
|
process
|
enoent error message inconsistencies version rc platform windows subsystem errors fs libuv process currently most of the enoent error messages have a similar signature error enoent no such file or directory full path to filename for example js fs accesssync does not exists etc error enoent no such file or directory access does not exists error enoent no such file or directory chmod does not exists error enoent no such file or directory link does not exists does not error enoent no such file or directory lstat does not exists error enoent no such file or directory open does not exists error enoent no such file or directory readlink does not exists error enoent no such file or directory rename does not exists does not error enoent no such file or directory rmdir does not exists error enoent no such file or directory scandir does not exists error enoent no such file or directory stat does not exists error enoent no such file or directory unlink does not exists error enoent no such file or directory utime does not exists i ve stumbled upon cases that have different and somehow confusing signatures different order no description filename without quotes no full path js fs watch does not exists error watch does not exists enoent no file name path at all the call name can be wrongly considered as the filename js process chdir does not exists error enoent no such file or directory uv chdir are these cases worth unification can they be addressed in node js or they are libuv features
| 1
|
21,585
| 29,953,993,617
|
IssuesEvent
|
2023-06-23 05:34:55
|
oasis-tcs/sarif-spec
|
https://api.github.com/repos/oasis-tcs/sarif-spec
|
closed
|
Errata01 20230619 Appendix M links to section 3 and 4 broken
|
process 2.1.0-erratum editorial
|
When accepting all changes the links in the rightmost column of the table in appendix M appera to be broken.
When I replace them with links they look identical (field values on/off toggling) but then they work (jumping to section start of 3 and 4 respectively).
Could be an application specific problem.
Proposal: If possible make them work.
Also, there is no link to the reference entry of the RFC in:
> The following is a list of MIME types and file extensions for files that conform to this specification, registered according to [RFC2048].
Reference bundle <https://www.oasis-open.org/committees/document.php?document_id=71131&wg_abbrev=sarif> and file `sarif-v2.1.0-errata01-csd01-redlined.docx` (at the very end)
|
1.0
|
Errata01 20230619 Appendix M links to section 3 and 4 broken - When accepting all changes the links in the rightmost column of the table in appendix M appera to be broken.
When I replace them with links they look identical (field values on/off toggling) but then they work (jumping to section start of 3 and 4 respectively).
Could be an application specific problem.
Proposal: If possible make them work.
Also, there is no link to the reference entry of the RFC in:
> The following is a list of MIME types and file extensions for files that conform to this specification, registered according to [RFC2048].
Reference bundle <https://www.oasis-open.org/committees/document.php?document_id=71131&wg_abbrev=sarif> and file `sarif-v2.1.0-errata01-csd01-redlined.docx` (at the very end)
|
process
|
appendix m links to section and broken when accepting all changes the links in the rightmost column of the table in appendix m appera to be broken when i replace them with links they look identical field values on off toggling but then they work jumping to section start of and respectively could be an application specific problem proposal if possible make them work also there is no link to the reference entry of the rfc in the following is a list of mime types and file extensions for files that conform to this specification registered according to reference bundle and file sarif redlined docx at the very end
| 1
|
315,577
| 27,086,646,252
|
IssuesEvent
|
2023-02-14 17:28:07
|
eclipse-openj9/openj9
|
https://api.github.com/repos/eclipse-openj9/openj9
|
opened
|
CryptoTests timeout in SeedGenerator reading /dev/random
|
comp:test test failure comp:infra
|
https://openj9-jenkins.osuosl.org/job/Test_openjdk11_j9_extended.functional_ppc64le_linux_Nightly_testList_1/457/ - [cent7-ppcle-2](https://openj9-jenkins.osuosl.org/computer/cent7-ppcle-2)
https://openj9-jenkins.osuosl.org/job/Test_openjdk11_j9_extended.functional_x86-64_linux_Nightly_testList_0/451/ - [ub18-x86-2](https://openj9-jenkins.osuosl.org/computer/ub18-x86-2)
cryptotest/tests/SecureRandomTests.java
cryptotest/CryptoTest.java
Seems a problem reading from the default securerandom.source `/dev/random` on some machines.
Not sure if it's a valid test, but if I do `cat /dev/random` it either hangs right away, or prints a little bit and then hangs. Unless somebody can figure out how to fix the machines, it would be better for the test to use `/dev/urandom`, which never blocks, by having the test set the system property `-Djava.security.egd=file:/dev/urandom`.
```
ACTION: main -- Error. Program `/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_ppc64le_linux_Nightly_testList_1/openjdkbinary/j2sdk-image/bin/java' timed out (timeout set to 120000ms, elapsed time including timeout handling was 120772ms).
"MainThread" prio=5 Id=26 RUNNABLE
at java.base@11.0.18-internal/java.io.FileInputStream.readBytes(Native Method)
at java.base@11.0.18-internal/java.io.FileInputStream.read(FileInputStream.java:279)
at java.base@11.0.18-internal/java.io.FilterInputStream.read(FilterInputStream.java:133)
at java.base@11.0.18-internal/sun.security.provider.SeedGenerator$URLSeedGenerator.getSeedBytes(SeedGenerator.java:541)
at java.base@11.0.18-internal/sun.security.provider.SeedGenerator.generateSeed(SeedGenerator.java:144)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.lambda$static$0(AbstractDrbg.java:524)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.getEntropyInput(AbstractDrbg.java:507)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.getEntropyInput(AbstractDrbg.java:494)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.instantiateIfNecessary(AbstractDrbg.java:696)
- locked sun.security.provider.HashDrbg@37bf7209
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.engineNextBytes(AbstractDrbg.java:378)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.engineNextBytes(AbstractDrbg.java:334)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg$SeederHolder.lambda$static$0(AbstractDrbg.java:558)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.getEntropyInput(AbstractDrbg.java:507)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.getEntropyInput(AbstractDrbg.java:494)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.instantiateIfNecessary(AbstractDrbg.java:696)
- locked sun.security.provider.HashDrbg@70b36660
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.engineNextBytes(AbstractDrbg.java:378)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.engineNextBytes(AbstractDrbg.java:334)
at java.base@11.0.18-internal/sun.security.provider.DRBG.engineNextBytes(DRBG.java:233)
at java.base@11.0.18-internal/java.security.SecureRandom.nextBytes(SecureRandom.java:768)
at java.base@11.0.18-internal/java.security.SecureRandom.next(SecureRandom.java:825)
at java.base@11.0.18-internal/java.util.Random.nextInt(Random.java:329)
at app//cryptotest.tests.SecureRandomTests.checkAlgorithm(SecureRandomTests.java:73)
at app//cryptotest.utils.AlgorithmTest.mainLoop(AlgorithmTest.java:79)
at app//cryptotest.tests.SecureRandomTests.main(SecureRandomTests.java:60)
```
@llxia @AdamBrousseau
|
2.0
|
CryptoTests timeout in SeedGenerator reading /dev/random - https://openj9-jenkins.osuosl.org/job/Test_openjdk11_j9_extended.functional_ppc64le_linux_Nightly_testList_1/457/ - [cent7-ppcle-2](https://openj9-jenkins.osuosl.org/computer/cent7-ppcle-2)
https://openj9-jenkins.osuosl.org/job/Test_openjdk11_j9_extended.functional_x86-64_linux_Nightly_testList_0/451/ - [ub18-x86-2](https://openj9-jenkins.osuosl.org/computer/ub18-x86-2)
cryptotest/tests/SecureRandomTests.java
cryptotest/CryptoTest.java
Seems a problem reading from the default securerandom.source `/dev/random` on some machines.
Not sure if it's a valid test, but if I do `cat /dev/random` it either hangs right away, or prints a little bit and then hangs. Unless somebody can figure out how to fix the machines, it would be better for the test to use `/dev/urandom`, which never blocks, by having the test set the system property `-Djava.security.egd=file:/dev/urandom`.
```
ACTION: main -- Error. Program `/home/jenkins/workspace/Test_openjdk11_j9_extended.functional_ppc64le_linux_Nightly_testList_1/openjdkbinary/j2sdk-image/bin/java' timed out (timeout set to 120000ms, elapsed time including timeout handling was 120772ms).
"MainThread" prio=5 Id=26 RUNNABLE
at java.base@11.0.18-internal/java.io.FileInputStream.readBytes(Native Method)
at java.base@11.0.18-internal/java.io.FileInputStream.read(FileInputStream.java:279)
at java.base@11.0.18-internal/java.io.FilterInputStream.read(FilterInputStream.java:133)
at java.base@11.0.18-internal/sun.security.provider.SeedGenerator$URLSeedGenerator.getSeedBytes(SeedGenerator.java:541)
at java.base@11.0.18-internal/sun.security.provider.SeedGenerator.generateSeed(SeedGenerator.java:144)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.lambda$static$0(AbstractDrbg.java:524)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.getEntropyInput(AbstractDrbg.java:507)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.getEntropyInput(AbstractDrbg.java:494)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.instantiateIfNecessary(AbstractDrbg.java:696)
- locked sun.security.provider.HashDrbg@37bf7209
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.engineNextBytes(AbstractDrbg.java:378)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.engineNextBytes(AbstractDrbg.java:334)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg$SeederHolder.lambda$static$0(AbstractDrbg.java:558)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.getEntropyInput(AbstractDrbg.java:507)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.getEntropyInput(AbstractDrbg.java:494)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.instantiateIfNecessary(AbstractDrbg.java:696)
- locked sun.security.provider.HashDrbg@70b36660
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.engineNextBytes(AbstractDrbg.java:378)
at java.base@11.0.18-internal/sun.security.provider.AbstractDrbg.engineNextBytes(AbstractDrbg.java:334)
at java.base@11.0.18-internal/sun.security.provider.DRBG.engineNextBytes(DRBG.java:233)
at java.base@11.0.18-internal/java.security.SecureRandom.nextBytes(SecureRandom.java:768)
at java.base@11.0.18-internal/java.security.SecureRandom.next(SecureRandom.java:825)
at java.base@11.0.18-internal/java.util.Random.nextInt(Random.java:329)
at app//cryptotest.tests.SecureRandomTests.checkAlgorithm(SecureRandomTests.java:73)
at app//cryptotest.utils.AlgorithmTest.mainLoop(AlgorithmTest.java:79)
at app//cryptotest.tests.SecureRandomTests.main(SecureRandomTests.java:60)
```
@llxia @AdamBrousseau
|
non_process
|
cryptotests timeout in seedgenerator reading dev random cryptotest tests securerandomtests java cryptotest cryptotest java seems a problem reading from the default securerandom source dev random on some machines not sure if it s a valid test but if i do cat dev random it either hangs right away or prints a little bit and then hangs unless somebody can figure out how to fix the machines it would be better for the test to use dev urandom which never blocks by having the test set the system property djava security egd file dev urandom action main error program home jenkins workspace test extended functional linux nightly testlist openjdkbinary image bin java timed out timeout set to elapsed time including timeout handling was mainthread prio id runnable at java base internal java io fileinputstream readbytes native method at java base internal java io fileinputstream read fileinputstream java at java base internal java io filterinputstream read filterinputstream java at java base internal sun security provider seedgenerator urlseedgenerator getseedbytes seedgenerator java at java base internal sun security provider seedgenerator generateseed seedgenerator java at java base internal sun security provider abstractdrbg lambda static abstractdrbg java at java base internal sun security provider abstractdrbg getentropyinput abstractdrbg java at java base internal sun security provider abstractdrbg getentropyinput abstractdrbg java at java base internal sun security provider abstractdrbg instantiateifnecessary abstractdrbg java locked sun security provider hashdrbg at java base internal sun security provider abstractdrbg enginenextbytes abstractdrbg java at java base internal sun security provider abstractdrbg enginenextbytes abstractdrbg java at java base internal sun security provider abstractdrbg seederholder lambda static abstractdrbg java at java base internal sun security provider abstractdrbg getentropyinput abstractdrbg java at java base internal sun security provider abstractdrbg getentropyinput abstractdrbg java at java base internal sun security provider abstractdrbg instantiateifnecessary abstractdrbg java locked sun security provider hashdrbg at java base internal sun security provider abstractdrbg enginenextbytes abstractdrbg java at java base internal sun security provider abstractdrbg enginenextbytes abstractdrbg java at java base internal sun security provider drbg enginenextbytes drbg java at java base internal java security securerandom nextbytes securerandom java at java base internal java security securerandom next securerandom java at java base internal java util random nextint random java at app cryptotest tests securerandomtests checkalgorithm securerandomtests java at app cryptotest utils algorithmtest mainloop algorithmtest java at app cryptotest tests securerandomtests main securerandomtests java llxia adambrousseau
| 0
|
5,224
| 8,029,187,910
|
IssuesEvent
|
2018-07-27 15:13:37
|
threefoldfoundation/tf_app
|
https://api.github.com/repos/threefoldfoundation/tf_app
|
closed
|
Delete 'ThreeFold Foundation' under TF Foundation
|
priority_major process_duplicate state_inprogress
|
TF Foundation is now split up under
Decentralized exchange
ThreeFold Grid
ThreeFold Token
No others needed here
|
1.0
|
Delete 'ThreeFold Foundation' under TF Foundation - TF Foundation is now split up under
Decentralized exchange
ThreeFold Grid
ThreeFold Token
No others needed here
|
process
|
delete threefold foundation under tf foundation tf foundation is now split up under decentralized exchange threefold grid threefold token no others needed here
| 1
|
20,858
| 27,637,756,996
|
IssuesEvent
|
2023-03-10 15:39:47
|
bitfocus/companion-module-requests
|
https://api.github.com/repos/bitfocus/companion-module-requests
|
opened
|
Local Keyboard
|
NOT YET PROCESSED
|
- [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested**
The name of the device, hardware, or software you would like to control:
Local keyboard
What you would like to be able to make it do from Companion:
Keyboard shortcuts without running xdotool or equivalent
Direct links or attachments to the ethernet control protocol or API:
This is already implemented by Enerccio: https://github.com/Enerccio/companion-modules-os-keyboard
Is there a reason not to just add this?
|
1.0
|
Local Keyboard - - [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested**
The name of the device, hardware, or software you would like to control:
Local keyboard
What you would like to be able to make it do from Companion:
Keyboard shortcuts without running xdotool or equivalent
Direct links or attachments to the ethernet control protocol or API:
This is already implemented by Enerccio: https://github.com/Enerccio/companion-modules-os-keyboard
Is there a reason not to just add this?
|
process
|
local keyboard i have researched the list of existing companion modules and requests and have determined this has not yet been requested the name of the device hardware or software you would like to control local keyboard what you would like to be able to make it do from companion keyboard shortcuts without running xdotool or equivalent direct links or attachments to the ethernet control protocol or api this is already implemented by enerccio is there a reason not to just add this
| 1
|
12,607
| 15,008,152,841
|
IssuesEvent
|
2021-01-31 08:44:54
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
closed
|
Uncaught TypeError: Cannot read property 'naturalHeight' of undefined - On VDC list
|
process_duplicate process_wontfix type_bug
|
1) I have one deployed VDC Right now
2) Browse to VDC list with console open
results in
Uncaught TypeError: Cannot read property 'naturalHeight' of undefined
Every new browse throws the error again.
Issue was logged and closed in the past for a different page.
See https://github.com/threefoldtech/js-sdk/issues/1029
|
2.0
|
Uncaught TypeError: Cannot read property 'naturalHeight' of undefined - On VDC list - 1) I have one deployed VDC Right now
2) Browse to VDC list with console open
results in
Uncaught TypeError: Cannot read property 'naturalHeight' of undefined
Every new browse throws the error again.
Issue was logged and closed in the past for a different page.
See https://github.com/threefoldtech/js-sdk/issues/1029
|
process
|
uncaught typeerror cannot read property naturalheight of undefined on vdc list i have one deployed vdc right now browse to vdc list with console open results in uncaught typeerror cannot read property naturalheight of undefined every new browse throws the error again issue was logged and closed in the past for a different page see
| 1
|
115,547
| 11,881,731,095
|
IssuesEvent
|
2020-03-27 13:14:51
|
brinkflew/odev
|
https://api.github.com/repos/brinkflew/odev
|
closed
|
Requirements have no installation candidate during
|
bug documentation
|
## Issue
Some requirements have no installation candidates because they are already standard to Python. This leads to errors while following the setup instructions as defined in the README.
## Ideas
- Remove standard libraries from the setup part of the README
|
1.0
|
Requirements have no installation candidate during - ## Issue
Some requirements have no installation candidates because they are already standard to Python. This leads to errors while following the setup instructions as defined in the README.
## Ideas
- Remove standard libraries from the setup part of the README
|
non_process
|
requirements have no installation candidate during issue some requirements have no installation candidates because they are already standard to python this leads to errors while following the setup instructions as defined in the readme ideas remove standard libraries from the setup part of the readme
| 0
|
4,945
| 7,795,905,530
|
IssuesEvent
|
2018-06-08 09:41:04
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
Document Plotgen 3 (Trac #189)
|
Migrated from Trac post_processing task
|
Documentation needs to be created for Plotgen 3, in particular:
1. Usage
1. Implementing Cases
1. Code Design
I started a TWiki topic [http://larson-group.com/twiki/bin/view.pl/Documentation/CarsonDoc/Plotgen3 here].
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/189
```json
{
"status": "closed",
"changetime": "2009-08-26T15:10:55",
"description": "Documentation needs to be created for Plotgen 3, in particular:\n 1. Usage\n 1. Implementing Cases\n 1. Code Design\n\nI started a TWiki topic [http://larson-group.com/twiki/bin/view.pl/Documentation/CarsonDoc/Plotgen3 here].",
"reporter": "senkbeil@uwm.edu",
"cc": "nielsenb@uwm.edu",
"resolution": "fixed",
"_ts": "1251299455000000",
"component": "post_processing",
"summary": "Document Plotgen 3",
"priority": "major",
"keywords": "",
"time": "2009-08-14T18:06:59",
"milestone": "Plotgen 3.0",
"owner": "",
"type": "task"
}
```
|
1.0
|
Document Plotgen 3 (Trac #189) - Documentation needs to be created for Plotgen 3, in particular:
1. Usage
1. Implementing Cases
1. Code Design
I started a TWiki topic [http://larson-group.com/twiki/bin/view.pl/Documentation/CarsonDoc/Plotgen3 here].
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/189
```json
{
"status": "closed",
"changetime": "2009-08-26T15:10:55",
"description": "Documentation needs to be created for Plotgen 3, in particular:\n 1. Usage\n 1. Implementing Cases\n 1. Code Design\n\nI started a TWiki topic [http://larson-group.com/twiki/bin/view.pl/Documentation/CarsonDoc/Plotgen3 here].",
"reporter": "senkbeil@uwm.edu",
"cc": "nielsenb@uwm.edu",
"resolution": "fixed",
"_ts": "1251299455000000",
"component": "post_processing",
"summary": "Document Plotgen 3",
"priority": "major",
"keywords": "",
"time": "2009-08-14T18:06:59",
"milestone": "Plotgen 3.0",
"owner": "",
"type": "task"
}
```
|
process
|
document plotgen trac documentation needs to be created for plotgen in particular usage implementing cases code design i started a twiki topic attachments migrated from json status closed changetime description documentation needs to be created for plotgen in particular n usage n implementing cases n code design n ni started a twiki topic reporter senkbeil uwm edu cc nielsenb uwm edu resolution fixed ts component post processing summary document plotgen priority major keywords time milestone plotgen owner type task
| 1
|
96
| 2,536,373,208
|
IssuesEvent
|
2015-01-26 13:33:03
|
appsgate2015/appsgate
|
https://api.github.com/repos/appsgate2015/appsgate
|
closed
|
Message dans la bulle d'état
|
appsgate-client P1 PROCESSING
|
Ce message n'est pas toujours le bon, par exemple quand le programme est vide
|
1.0
|
Message dans la bulle d'état - Ce message n'est pas toujours le bon, par exemple quand le programme est vide
|
process
|
message dans la bulle d état ce message n est pas toujours le bon par exemple quand le programme est vide
| 1
|
5,965
| 8,786,835,367
|
IssuesEvent
|
2018-12-20 16:46:35
|
home-assistant/home-assistant
|
https://api.github.com/repos/home-assistant/home-assistant
|
closed
|
Seven Segments cannot have multiple instances because of ocr.png
|
platform: image_processing.seven_segments
|
Per this discussion:
https://community.home-assistant.io/t/multiple-seven-segments/67166/4
Seven Segments can not have multiple instances because the component hard codes the png file name fed to ssocr as home-assistant-config-dir/ocr.png.
Attempting to have multiple instances results in a race condition where instances overwrite each other's data.
|
1.0
|
Seven Segments cannot have multiple instances because of ocr.png - Per this discussion:
https://community.home-assistant.io/t/multiple-seven-segments/67166/4
Seven Segments can not have multiple instances because the component hard codes the png file name fed to ssocr as home-assistant-config-dir/ocr.png.
Attempting to have multiple instances results in a race condition where instances overwrite each other's data.
|
process
|
seven segments cannot have multiple instances because of ocr png per this discussion seven segments can not have multiple instances because the component hard codes the png file name fed to ssocr as home assistant config dir ocr png attempting to have multiple instances results in a race condition where instances overwrite each other s data
| 1
|
22,703
| 32,023,907,208
|
IssuesEvent
|
2023-09-22 07:26:30
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Hiting `Test` in QgsProcessingPointCloudExpressionDialog with QgsProcessingParameterMultipleLayers as inputs cause crash
|
Processing Bug Crash/Data Corruption Point Clouds
|
### What is the bug or the crash?
## Report Details
**Python Stack Trace**
```
Windows fatal exception: code 0xc0000139
Current thread 0x0000239c (most recent call first):
<no Python frame>
Windows fatal exception: code 0xc0000139
Current thread 0x0000239c (most recent call first):
<no Python frame>
Windows fatal exception: access violation
Current thread 0x0000239c (most recent call first):
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\ProcessingPlugin.py", line 432 in executeAlgorithm
dlg.exec_()
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\ProcessingToolbox.py", line 232 in executeAlgorithm
self.executeWithGui.emit(alg.id(), self, self.in_place_mode, False)
```
**Stack Trace**
```
QgsProcessingPointCloudExpressionDialog::test :
QObject::qt_static_metacall :
QAbstractButton::clicked :
QAbstractButton::click :
QAbstractButton::mouseReleaseEvent :
QWidget::event :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QApplicationPrivate::sendMouseEvent :
QSizePolicy::QSizePolicy :
QSizePolicy::QSizePolicy :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QGuiApplicationPrivate::processMouseEvent :
QWindowSystemInterface::sendWindowSystemEvents :
QEventDispatcherWin32::processEvents :
qt_plugin_query_metadata :
QEventLoop::exec :
QDialog::exec :
QgsProcessingPointCloudExpressionLineEdit::editExpression :
QObject::qt_static_metacall :
QAbstractButton::clicked :
QAbstractButton::click :
QAbstractButton::mouseReleaseEvent :
QToolButton::mouseReleaseEvent :
QWidget::event :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QApplicationPrivate::sendMouseEvent :
QSizePolicy::QSizePolicy :
QSizePolicy::QSizePolicy :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QGuiApplicationPrivate::processMouseEvent :
QWindowSystemInterface::sendWindowSystemEvents :
QEventDispatcherWin32::processEvents :
qt_plugin_query_metadata :
QEventLoop::exec :
QDialog::exec :
PyInit_QtWidgets :
PyArg_ParseTuple_SizeT :
PyEval_EvalFrameDefault :
PyObject_GC_Del :
PyFunction_Vectorcall :
PyFloat_FromDouble :
PyVectorcall_Call :
PyObject_Call :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
QObject::qt_static_metacall :
QMetaObject::activate :
PyInit_QtCore :
PyInit_QtCore :
PyType_GenericNew :
PyEval_EvalFrameDefault :
PyFunction_Vectorcall :
PyFloat_FromDouble :
PyVectorcall_Call :
PyObject_Call :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
QObject::qt_static_metacall :
QAbstractItemView::doubleClicked :
QTreeView::mouseDoubleClickEvent :
PyInit__gui :
QWidget::event :
QFrame::event :
QAbstractItemView::viewportEvent :
PyInit__gui :
QCoreApplicationPrivate::sendThroughObjectEventFilters :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QApplicationPrivate::sendMouseEvent :
QSizePolicy::QSizePolicy :
QSizePolicy::QSizePolicy :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QGuiApplicationPrivate::processMouseEvent :
QWindowSystemInterface::sendWindowSystemEvents :
QEventDispatcherWin32::processEvents :
qt_plugin_query_metadata :
QEventLoop::exec :
QCoreApplication::exec :
main :
BaseThreadInitThunk :
```
**QGIS Info**
QGIS Version: 3.32.3-Lima
QGIS code revision: 67d46100b5
Compiled against Qt: 5.15.3
Running against Qt: 5.15.3
Compiled against GDAL: 3.7.2
Running against GDAL: 3.7.2
**System Info**
CPU Type: x86_64
Kernel Type: winnt
Kernel Version: 10.0.19045
### Steps to reproduce the issue
1. Open Merge point cloud layers alg
2. Add multiple layers
3. Open expression dialog
4. Set "Classification = 2" and hit Test
5. Crash
### Versions
3.32.3 OSGeo4w
### Supported QGIS version
- [X] I'm running a supported QGIS version according to [the roadmap](https://www.qgis.org/en/site/getinvolved/development/roadmap.html#release-schedule).
### New profile
- [X] I tried with a new [QGIS profile](https://docs.qgis.org/latest/en/docs/user_manual/introduction/qgis_configuration.html#working-with-user-profiles)
### Additional context
_No response_
|
1.0
|
Hiting `Test` in QgsProcessingPointCloudExpressionDialog with QgsProcessingParameterMultipleLayers as inputs cause crash - ### What is the bug or the crash?
## Report Details
**Python Stack Trace**
```
Windows fatal exception: code 0xc0000139
Current thread 0x0000239c (most recent call first):
<no Python frame>
Windows fatal exception: code 0xc0000139
Current thread 0x0000239c (most recent call first):
<no Python frame>
Windows fatal exception: access violation
Current thread 0x0000239c (most recent call first):
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\ProcessingPlugin.py", line 432 in executeAlgorithm
dlg.exec_()
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\ProcessingToolbox.py", line 232 in executeAlgorithm
self.executeWithGui.emit(alg.id(), self, self.in_place_mode, False)
```
**Stack Trace**
```
QgsProcessingPointCloudExpressionDialog::test :
QObject::qt_static_metacall :
QAbstractButton::clicked :
QAbstractButton::click :
QAbstractButton::mouseReleaseEvent :
QWidget::event :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QApplicationPrivate::sendMouseEvent :
QSizePolicy::QSizePolicy :
QSizePolicy::QSizePolicy :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QGuiApplicationPrivate::processMouseEvent :
QWindowSystemInterface::sendWindowSystemEvents :
QEventDispatcherWin32::processEvents :
qt_plugin_query_metadata :
QEventLoop::exec :
QDialog::exec :
QgsProcessingPointCloudExpressionLineEdit::editExpression :
QObject::qt_static_metacall :
QAbstractButton::clicked :
QAbstractButton::click :
QAbstractButton::mouseReleaseEvent :
QToolButton::mouseReleaseEvent :
QWidget::event :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QApplicationPrivate::sendMouseEvent :
QSizePolicy::QSizePolicy :
QSizePolicy::QSizePolicy :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QGuiApplicationPrivate::processMouseEvent :
QWindowSystemInterface::sendWindowSystemEvents :
QEventDispatcherWin32::processEvents :
qt_plugin_query_metadata :
QEventLoop::exec :
QDialog::exec :
PyInit_QtWidgets :
PyArg_ParseTuple_SizeT :
PyEval_EvalFrameDefault :
PyObject_GC_Del :
PyFunction_Vectorcall :
PyFloat_FromDouble :
PyVectorcall_Call :
PyObject_Call :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
QObject::qt_static_metacall :
QMetaObject::activate :
PyInit_QtCore :
PyInit_QtCore :
PyType_GenericNew :
PyEval_EvalFrameDefault :
PyFunction_Vectorcall :
PyFloat_FromDouble :
PyVectorcall_Call :
PyObject_Call :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
PyInit_QtCore :
QObject::qt_static_metacall :
QAbstractItemView::doubleClicked :
QTreeView::mouseDoubleClickEvent :
PyInit__gui :
QWidget::event :
QFrame::event :
QAbstractItemView::viewportEvent :
PyInit__gui :
QCoreApplicationPrivate::sendThroughObjectEventFilters :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QApplicationPrivate::sendMouseEvent :
QSizePolicy::QSizePolicy :
QSizePolicy::QSizePolicy :
QApplicationPrivate::notify_helper :
QApplication::notify :
QgsApplication::notify :
QCoreApplication::notifyInternal2 :
QGuiApplicationPrivate::processMouseEvent :
QWindowSystemInterface::sendWindowSystemEvents :
QEventDispatcherWin32::processEvents :
qt_plugin_query_metadata :
QEventLoop::exec :
QCoreApplication::exec :
main :
BaseThreadInitThunk :
```
**QGIS Info**
QGIS Version: 3.32.3-Lima
QGIS code revision: 67d46100b5
Compiled against Qt: 5.15.3
Running against Qt: 5.15.3
Compiled against GDAL: 3.7.2
Running against GDAL: 3.7.2
**System Info**
CPU Type: x86_64
Kernel Type: winnt
Kernel Version: 10.0.19045
### Steps to reproduce the issue
1. Open Merge point cloud layers alg
2. Add multiple layers
3. Open expression dialog
4. Set "Classification = 2" and hit Test
5. Crash
### Versions
3.32.3 OSGeo4w
### Supported QGIS version
- [X] I'm running a supported QGIS version according to [the roadmap](https://www.qgis.org/en/site/getinvolved/development/roadmap.html#release-schedule).
### New profile
- [X] I tried with a new [QGIS profile](https://docs.qgis.org/latest/en/docs/user_manual/introduction/qgis_configuration.html#working-with-user-profiles)
### Additional context
_No response_
|
process
|
hiting test in qgsprocessingpointcloudexpressiondialog with qgsprocessingparametermultiplelayers as inputs cause crash what is the bug or the crash report details python stack trace windows fatal exception code current thread most recent call first windows fatal exception code current thread most recent call first windows fatal exception access violation current thread most recent call first file c apps qgis python plugins processing processingplugin py line in executealgorithm dlg exec file c apps qgis python plugins processing gui processingtoolbox py line in executealgorithm self executewithgui emit alg id self self in place mode false stack trace qgsprocessingpointcloudexpressiondialog test qobject qt static metacall qabstractbutton clicked qabstractbutton click qabstractbutton mousereleaseevent qwidget event qapplicationprivate notify helper qapplication notify qgsapplication notify qcoreapplication qapplicationprivate sendmouseevent qsizepolicy qsizepolicy qsizepolicy qsizepolicy qapplicationprivate notify helper qapplication notify qgsapplication notify qcoreapplication qguiapplicationprivate processmouseevent qwindowsysteminterface sendwindowsystemevents processevents qt plugin query metadata qeventloop exec qdialog exec qgsprocessingpointcloudexpressionlineedit editexpression qobject qt static metacall qabstractbutton clicked qabstractbutton click qabstractbutton mousereleaseevent qtoolbutton mousereleaseevent qwidget event qapplicationprivate notify helper qapplication notify qgsapplication notify qcoreapplication qapplicationprivate sendmouseevent qsizepolicy qsizepolicy qsizepolicy qsizepolicy qapplicationprivate notify helper qapplication notify qgsapplication notify qcoreapplication qguiapplicationprivate processmouseevent qwindowsysteminterface sendwindowsystemevents processevents qt plugin query metadata qeventloop exec qdialog exec pyinit qtwidgets pyarg parsetuple sizet pyeval evalframedefault pyobject gc del pyfunction vectorcall pyfloat fromdouble pyvectorcall call pyobject call pyinit qtcore pyinit qtcore pyinit qtcore pyinit qtcore pyinit qtcore pyinit qtcore qobject qt static metacall qmetaobject activate pyinit qtcore pyinit qtcore pytype genericnew pyeval evalframedefault pyfunction vectorcall pyfloat fromdouble pyvectorcall call pyobject call pyinit qtcore pyinit qtcore pyinit qtcore pyinit qtcore qobject qt static metacall qabstractitemview doubleclicked qtreeview mousedoubleclickevent pyinit gui qwidget event qframe event qabstractitemview viewportevent pyinit gui qcoreapplicationprivate sendthroughobjecteventfilters qapplicationprivate notify helper qapplication notify qgsapplication notify qcoreapplication qapplicationprivate sendmouseevent qsizepolicy qsizepolicy qsizepolicy qsizepolicy qapplicationprivate notify helper qapplication notify qgsapplication notify qcoreapplication qguiapplicationprivate processmouseevent qwindowsysteminterface sendwindowsystemevents processevents qt plugin query metadata qeventloop exec qcoreapplication exec main basethreadinitthunk qgis info qgis version lima qgis code revision compiled against qt running against qt compiled against gdal running against gdal system info cpu type kernel type winnt kernel version steps to reproduce the issue open merge point cloud layers alg add multiple layers open expression dialog set classification and hit test crash versions supported qgis version i m running a supported qgis version according to new profile i tried with a new additional context no response
| 1
|
1,850
| 3,402,790,196
|
IssuesEvent
|
2015-12-03 02:14:06
|
elmsln/elmsln
|
https://api.github.com/repos/elmsln/elmsln
|
closed
|
wrapper drush-create-site so that it can process 1 network at a time
|
future infrastructure scale / performance
|
right now we have a lock file that gets touched to ensure that multiple drush si builds don't happen at the same time. Possible way of ensuring this script doesn't step on itself while still processing multiple targets
- [ ] Lock on 1 network at a time; but we can hit all aspects of this network at the same time
- [ ] this involves kicking each off to its own disconnected thread that takes 2 arguments, one for the coursename and the other for the system to process
- [ ] each thread creates a file and destroys it when done
- [ ] higher level script notices that all progress files within that network are finished and destroys the lock
|
1.0
|
wrapper drush-create-site so that it can process 1 network at a time - right now we have a lock file that gets touched to ensure that multiple drush si builds don't happen at the same time. Possible way of ensuring this script doesn't step on itself while still processing multiple targets
- [ ] Lock on 1 network at a time; but we can hit all aspects of this network at the same time
- [ ] this involves kicking each off to its own disconnected thread that takes 2 arguments, one for the coursename and the other for the system to process
- [ ] each thread creates a file and destroys it when done
- [ ] higher level script notices that all progress files within that network are finished and destroys the lock
|
non_process
|
wrapper drush create site so that it can process network at a time right now we have a lock file that gets touched to ensure that multiple drush si builds don t happen at the same time possible way of ensuring this script doesn t step on itself while still processing multiple targets lock on network at a time but we can hit all aspects of this network at the same time this involves kicking each off to its own disconnected thread that takes arguments one for the coursename and the other for the system to process each thread creates a file and destroys it when done higher level script notices that all progress files within that network are finished and destroys the lock
| 0
|
13,679
| 16,431,942,443
|
IssuesEvent
|
2021-05-20 03:39:17
|
gfx-rs/naga
|
https://api.github.com/repos/gfx-rs/naga
|
closed
|
Receiving buffer structure size wrong with min_binding_size of unbound array, with no unbound_array
|
area: processing
|
Hello, I create a minimum example here: https://github.com/Dispersia/naga_example
Error:
```
wgpu error: Validation Error
Caused by:
In Device::create_render_pipeline
error matching VERTEX shader requirements against the pipeline
shader global ResourceBinding { group: 0, binding: 0 } is not available in the layout pipeline layout
buffer structure size 80, added to one element of an unbound array, if it's the last field, ended up greater than the given `min_binding_size`
```
If it matters, it only triggered when adding * globals.scale in the shader, but maybe naga is just removing the entire uniform since it's not used otherwise.
wgpu: 0.8.1, naga 0.4.1.
Specs: windows 10, using vulkan
|
1.0
|
Receiving buffer structure size wrong with min_binding_size of unbound array, with no unbound_array - Hello, I create a minimum example here: https://github.com/Dispersia/naga_example
Error:
```
wgpu error: Validation Error
Caused by:
In Device::create_render_pipeline
error matching VERTEX shader requirements against the pipeline
shader global ResourceBinding { group: 0, binding: 0 } is not available in the layout pipeline layout
buffer structure size 80, added to one element of an unbound array, if it's the last field, ended up greater than the given `min_binding_size`
```
If it matters, it only triggered when adding * globals.scale in the shader, but maybe naga is just removing the entire uniform since it's not used otherwise.
wgpu: 0.8.1, naga 0.4.1.
Specs: windows 10, using vulkan
|
process
|
receiving buffer structure size wrong with min binding size of unbound array with no unbound array hello i create a minimum example here error wgpu error validation error caused by in device create render pipeline error matching vertex shader requirements against the pipeline shader global resourcebinding group binding is not available in the layout pipeline layout buffer structure size added to one element of an unbound array if it s the last field ended up greater than the given min binding size if it matters it only triggered when adding globals scale in the shader but maybe naga is just removing the entire uniform since it s not used otherwise wgpu naga specs windows using vulkan
| 1
|
20,418
| 27,078,780,559
|
IssuesEvent
|
2023-02-14 12:34:42
|
island-is/island.is
|
https://api.github.com/repos/island-is/island.is
|
opened
|
New process - project/service lifecycle docs
|
sync 2022-Q1 process
|
What happens when a service is "done" and goes into maintenance and the team is no longer around.
Suggestion is to assign it to "Core" team in codeowners.
|
1.0
|
New process - project/service lifecycle docs - What happens when a service is "done" and goes into maintenance and the team is no longer around.
Suggestion is to assign it to "Core" team in codeowners.
|
process
|
new process project service lifecycle docs what happens when a service is done and goes into maintenance and the team is no longer around suggestion is to assign it to core team in codeowners
| 1
|
15,674
| 19,847,423,306
|
IssuesEvent
|
2022-01-21 08:27:07
|
ooi-data/RS01SBPD-DP01A-06-DOSTAD104-recovered_inst-dpc_optode_instrument_recovered
|
https://api.github.com/repos/ooi-data/RS01SBPD-DP01A-06-DOSTAD104-recovered_inst-dpc_optode_instrument_recovered
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T08:27:06.519118.
## Details
Flow name: `RS01SBPD-DP01A-06-DOSTAD104-recovered_inst-dpc_optode_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T08:27:06.519118.
## Details
Flow name: `RS01SBPD-DP01A-06-DOSTAD104-recovered_inst-dpc_optode_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered inst dpc optode instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
285,128
| 8,755,097,713
|
IssuesEvent
|
2018-12-14 13:53:23
|
bio-tools/biotoolsRegistry
|
https://api.github.com/repos/bio-tools/biotoolsRegistry
|
closed
|
HELP: rare disease domain is down
|
bug content critical priority fix verified
|
Hi,
We had settle up our domain some weeks ago (rare diseases) but it doesn't seems to work now; indeed we had more than 100 resources and now none of them are listed and I can't edit the domain either. Can you check what happened? and why all those resources have disappeared?
Thanks,
Leslie
|
1.0
|
HELP: rare disease domain is down - Hi,
We had settle up our domain some weeks ago (rare diseases) but it doesn't seems to work now; indeed we had more than 100 resources and now none of them are listed and I can't edit the domain either. Can you check what happened? and why all those resources have disappeared?
Thanks,
Leslie
|
non_process
|
help rare disease domain is down hi we had settle up our domain some weeks ago rare diseases but it doesn t seems to work now indeed we had more than resources and now none of them are listed and i can t edit the domain either can you check what happened and why all those resources have disappeared thanks leslie
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.