Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
29,202
| 23,796,487,306
|
IssuesEvent
|
2022-09-02 20:23:51
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
opened
|
Mono Bionic Android arm64 CI failure - Failed to get device's property SupportedArchitectures / Device unauthorized / $ADB_VENDOR_KEYS is not set
|
area-Infrastructure-mono test-failure arm64
|
Found in a release/7.0 PR: https://github.com/dotnet/runtime/pull/74926
- Queue: `net7.0-Linux-Release-arm64-Mono_Release_LinuxBionic-Windows.10.Amd64.Android.Open`
- Artifacts: https://dev.azure.com/dnceng-public/public/_build/results?buildId=1392&view=ms.vss-test-web.build-test-results-tab&runId=19052&paneView=dotnet-dnceng.dnceng-anon-build-release-tasks.helix-anon-test-information-tab&resultId=180803
- Affected the `System.Diagnostics.Contracts.Tests` work item.
Log file: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-74926-merge-0416eef15c6a45a1a6/System.Diagnostics.Contracts.Tests/1/console.fa330ef5.log?helixlogtype=result
Callstack:
```
[11:56:41] fail: Failed to get device's property SupportedArchitectures. Check if a device is attached / emulator is started
error: device unauthorized.
This adb server's $ADB_VENDOR_KEYS is not set
Try 'adb kill-server' if that seems wrong.
Otherwise check for a confirmation dialog on your device.
[11:56:41] fail: No attached device supports one of required architectures arm64-v8a
[11:56:41] dbug: No suitable devices found
[11:56:41] crit: Failed to find compatible device: arm64-v8a
```
|
1.0
|
Mono Bionic Android arm64 CI failure - Failed to get device's property SupportedArchitectures / Device unauthorized / $ADB_VENDOR_KEYS is not set - Found in a release/7.0 PR: https://github.com/dotnet/runtime/pull/74926
- Queue: `net7.0-Linux-Release-arm64-Mono_Release_LinuxBionic-Windows.10.Amd64.Android.Open`
- Artifacts: https://dev.azure.com/dnceng-public/public/_build/results?buildId=1392&view=ms.vss-test-web.build-test-results-tab&runId=19052&paneView=dotnet-dnceng.dnceng-anon-build-release-tasks.helix-anon-test-information-tab&resultId=180803
- Affected the `System.Diagnostics.Contracts.Tests` work item.
Log file: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-74926-merge-0416eef15c6a45a1a6/System.Diagnostics.Contracts.Tests/1/console.fa330ef5.log?helixlogtype=result
Callstack:
```
[11:56:41] fail: Failed to get device's property SupportedArchitectures. Check if a device is attached / emulator is started
error: device unauthorized.
This adb server's $ADB_VENDOR_KEYS is not set
Try 'adb kill-server' if that seems wrong.
Otherwise check for a confirmation dialog on your device.
[11:56:41] fail: No attached device supports one of required architectures arm64-v8a
[11:56:41] dbug: No suitable devices found
[11:56:41] crit: Failed to find compatible device: arm64-v8a
```
|
non_process
|
mono bionic android ci failure failed to get device s property supportedarchitectures device unauthorized adb vendor keys is not set found in a release pr queue linux release mono release linuxbionic windows android open artifacts affected the system diagnostics contracts tests work item log file callstack fail failed to get device s property supportedarchitectures check if a device is attached emulator is started error device unauthorized this adb server s adb vendor keys is not set try adb kill server if that seems wrong otherwise check for a confirmation dialog on your device fail no attached device supports one of required architectures dbug no suitable devices found crit failed to find compatible device
| 0
|
8,342
| 11,497,807,619
|
IssuesEvent
|
2020-02-12 10:43:11
|
18F/tts-tech-portfolio
|
https://api.github.com/repos/18F/tts-tech-portfolio
|
closed
|
TTS Tech Portfolio agile approach -- in depth discussion
|
epic: internal workflow/procedures needs grooming workflow: process
|
## Background information
https://github.com/18F/tts-tech-portfolio/issues/282
## User stories
<!-- one or more -->
- As a ..., I want ... so that ...
- As a ..., I want ... so that ...
## Implementation
- [ ] [first small task]
- [ ] [another small task]
## Acceptance criteria
- [ ] size labeling
- [ ] do we need Entrance Criteria and Exit Criteria for columns, or could we consolidate to one?
- [ ] want deadlines or specifying a finish-by date?
- [ ] Consider making a theme of the week/sprint/month
|
1.0
|
TTS Tech Portfolio agile approach -- in depth discussion - ## Background information
https://github.com/18F/tts-tech-portfolio/issues/282
## User stories
<!-- one or more -->
- As a ..., I want ... so that ...
- As a ..., I want ... so that ...
## Implementation
- [ ] [first small task]
- [ ] [another small task]
## Acceptance criteria
- [ ] size labeling
- [ ] do we need Entrance Criteria and Exit Criteria for columns, or could we consolidate to one?
- [ ] want deadlines or specifying a finish-by date?
- [ ] Consider making a theme of the week/sprint/month
|
process
|
tts tech portfolio agile approach in depth discussion background information user stories as a i want so that as a i want so that implementation acceptance criteria size labeling do we need entrance criteria and exit criteria for columns or could we consolidate to one want deadlines or specifying a finish by date consider making a theme of the week sprint month
| 1
|
7,673
| 10,760,662,277
|
IssuesEvent
|
2019-10-31 19:04:35
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
opened
|
Testing: rationalize / normalize VPCSC environment detection in systests
|
api: automl api: cloudasset api: dlp api: monitoring api: storage api: translation api: videointelligence testing type: process
|
We need a clear pattern for how to test whether the appropriate environment variables are set for VPCSC, and ways to skip tests when they are missing. I believe the constraints should be:
- Never set the environment variables in `noxfile.py`.
- If the inside / outside project variables are missing, skip the affected testcases cleanly.
I'm proposing to create a new module, `test_utils/test_utils/vpcsc_config.py`, which centralizes all this policy. Usage from systests would look like:
```python
from test_utils.vpcsc_config import vpcsc_config
@vpcsc_config.skip_if_no_inside_project
def test_requiring_inside_project():
do_something_with(vpcsc_config.project_inside)
@vpcsc_config.skip_if_no_outside_project
def test_requiring_outside_project():
do_something_with(vpcsc_config.project_outside)
@vpcsc_config.skip_if_no_inside_project
@vpcsc_config.skip_if_no_outside_project
def test_requiring_inside_and_outside_projects():
if vpcsc_config.inside_vpcsc:
do_something_with(vpcsc_config.project_inside, vpcsc_config.project_outside)
```
|
1.0
|
Testing: rationalize / normalize VPCSC environment detection in systests - We need a clear pattern for how to test whether the appropriate environment variables are set for VPCSC, and ways to skip tests when they are missing. I believe the constraints should be:
- Never set the environment variables in `noxfile.py`.
- If the inside / outside project variables are missing, skip the affected testcases cleanly.
I'm proposing to create a new module, `test_utils/test_utils/vpcsc_config.py`, which centralizes all this policy. Usage from systests would look like:
```python
from test_utils.vpcsc_config import vpcsc_config
@vpcsc_config.skip_if_no_inside_project
def test_requiring_inside_project():
do_something_with(vpcsc_config.project_inside)
@vpcsc_config.skip_if_no_outside_project
def test_requiring_outside_project():
do_something_with(vpcsc_config.project_outside)
@vpcsc_config.skip_if_no_inside_project
@vpcsc_config.skip_if_no_outside_project
def test_requiring_inside_and_outside_projects():
if vpcsc_config.inside_vpcsc:
do_something_with(vpcsc_config.project_inside, vpcsc_config.project_outside)
```
|
process
|
testing rationalize normalize vpcsc environment detection in systests we need a clear pattern for how to test whether the appropriate environment variables are set for vpcsc and ways to skip tests when they are missing i believe the constraints should be never set the environment variables in noxfile py if the inside outside project variables are missing skip the affected testcases cleanly i m proposing to create a new module test utils test utils vpcsc config py which centralizes all this policy usage from systests would look like python from test utils vpcsc config import vpcsc config vpcsc config skip if no inside project def test requiring inside project do something with vpcsc config project inside vpcsc config skip if no outside project def test requiring outside project do something with vpcsc config project outside vpcsc config skip if no inside project vpcsc config skip if no outside project def test requiring inside and outside projects if vpcsc config inside vpcsc do something with vpcsc config project inside vpcsc config project outside
| 1
|
7,517
| 10,596,006,065
|
IssuesEvent
|
2019-10-09 20:15:41
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
translocation into host/entry into host
|
multi-species process
|
responding to
https://github.com/geneontology/go-ontology/issues/17045
made me realise there are some unnecessary and confusing grouping terms under
GO:0044403 symbiont process
~For example
GO:0051824 recognition of other organism involved in symbiotic interaction
with a symbiont and host split~ see
https://github.com/geneontology/go-ontology/issues/17977
GO:0051836 translocation of molecules into other organism involved in symbiotic interaction
with a symbiont and host split
(the processes under here represent movement of the symbiont into the host, and movement of nutrients into the symbiont. The grouping term is biologically meaningless here. these processes aren't related at all.
~These terms often give pairs of descendants which are not biologically related at all, or any more than similar processes in no symbionts, they would be better grouped under~
~1 "host colonization" or similar
and
2. a suitable grouping term for the symbiont if required for terms which do not fit under such a term~
|
1.0
|
translocation into host/entry into host - responding to
https://github.com/geneontology/go-ontology/issues/17045
made me realise there are some unnecessary and confusing grouping terms under
GO:0044403 symbiont process
~For example
GO:0051824 recognition of other organism involved in symbiotic interaction
with a symbiont and host split~ see
https://github.com/geneontology/go-ontology/issues/17977
GO:0051836 translocation of molecules into other organism involved in symbiotic interaction
with a symbiont and host split
(the processes under here represent movement of the symbiont into the host, and movement of nutrients into the symbiont. The grouping term is biologically meaningless here. these processes aren't related at all.
~These terms often give pairs of descendants which are not biologically related at all, or any more than similar processes in no symbionts, they would be better grouped under~
~1 "host colonization" or similar
and
2. a suitable grouping term for the symbiont if required for terms which do not fit under such a term~
|
process
|
translocation into host entry into host responding to made me realise there are some unnecessary and confusing grouping terms under go symbiont process for example go recognition of other organism involved in symbiotic interaction with a symbiont and host split see go translocation of molecules into other organism involved in symbiotic interaction with a symbiont and host split the processes under here represent movement of the symbiont into the host and movement of nutrients into the symbiont the grouping term is biologically meaningless here these processes aren t related at all these terms often give pairs of descendants which are not biologically related at all or any more than similar processes in no symbionts they would be better grouped under host colonization or similar and a suitable grouping term for the symbiont if required for terms which do not fit under such a term
| 1
|
20,840
| 27,610,490,456
|
IssuesEvent
|
2023-03-09 15:40:41
|
Sebastian009w/hyper-burguer
|
https://api.github.com/repos/Sebastian009w/hyper-burguer
|
opened
|
Components
|
process
|
- [ ] Form
- [ ] Button
- [ ] Card
- [ ] TableContent
- [ ] Logo
- [ ] Title
- [ ] Hours
- [ ] Booking
|
1.0
|
Components - - [ ] Form
- [ ] Button
- [ ] Card
- [ ] TableContent
- [ ] Logo
- [ ] Title
- [ ] Hours
- [ ] Booking
|
process
|
components form button card tablecontent logo title hours booking
| 1
|
16,705
| 10,554,634,366
|
IssuesEvent
|
2019-10-03 19:55:48
|
microsoft/BotBuilder-Samples
|
https://api.github.com/repos/microsoft/BotBuilder-Samples
|
closed
|
Runing nodeJs npl-with-dispatch BotBuilder sample behind corporate proxy
|
Bot Services customer-replied-to customer-reported
|
I'm trying to run the sample "14.nlp-with-dispatch" on my local machine and test it with the emulator. I have a luis and Qna azure services. I have modified some code under node_module to add proxy configuration (just to run the exemple on my local machine) : this works for my luis service but I didn't find where to add proxy configuration for qna service knowing that I have tried to use npm packages "global-tunnel" and "global-tunnel-ng" but this didn't work for me because my node version is 12.9.1.
Has anyone experienced this?
Is there any other way to add proxy configuration for this two services ?
|
1.0
|
Runing nodeJs npl-with-dispatch BotBuilder sample behind corporate proxy - I'm trying to run the sample "14.nlp-with-dispatch" on my local machine and test it with the emulator. I have a luis and Qna azure services. I have modified some code under node_module to add proxy configuration (just to run the exemple on my local machine) : this works for my luis service but I didn't find where to add proxy configuration for qna service knowing that I have tried to use npm packages "global-tunnel" and "global-tunnel-ng" but this didn't work for me because my node version is 12.9.1.
Has anyone experienced this?
Is there any other way to add proxy configuration for this two services ?
|
non_process
|
runing nodejs npl with dispatch botbuilder sample behind corporate proxy i m trying to run the sample nlp with dispatch on my local machine and test it with the emulator i have a luis and qna azure services i have modified some code under node module to add proxy configuration just to run the exemple on my local machine this works for my luis service but i didn t find where to add proxy configuration for qna service knowing that i have tried to use npm packages global tunnel and global tunnel ng but this didn t work for me because my node version is has anyone experienced this is there any other way to add proxy configuration for this two services
| 0
|
3,685
| 6,715,898,957
|
IssuesEvent
|
2017-10-14 00:09:55
|
HelpyTeam/HelpyWeb
|
https://api.github.com/repos/HelpyTeam/HelpyWeb
|
closed
|
Update Conversation View
|
Front-end In Process priority/1
|
# Overview
Update front-end task
# Target
- [x] Set height for conversation view
- [x] Create scroll-bar for for conversation view _(Ref: https://www.w3schools.com/howto/howto_css_menu_horizontal_scroll.asp)_
- ~Find textarea format and apply to the front-end _(Ref: https://github.com/sstur/react-rte)_~
- ~Change current "Loading..." text into loading button. _(Ref: http://www.material-ui.com/#/components/refresh-indicator)_~
|
1.0
|
Update Conversation View - # Overview
Update front-end task
# Target
- [x] Set height for conversation view
- [x] Create scroll-bar for for conversation view _(Ref: https://www.w3schools.com/howto/howto_css_menu_horizontal_scroll.asp)_
- ~Find textarea format and apply to the front-end _(Ref: https://github.com/sstur/react-rte)_~
- ~Change current "Loading..." text into loading button. _(Ref: http://www.material-ui.com/#/components/refresh-indicator)_~
|
process
|
update conversation view overview update front end task target set height for conversation view create scroll bar for for conversation view ref find textarea format and apply to the front end ref change current loading text into loading button ref
| 1
|
183,266
| 31,240,191,585
|
IssuesEvent
|
2023-08-20 19:19:09
|
HUSTLE-UMC/HUSTLE_web
|
https://api.github.com/repos/HUSTLE-UMC/HUSTLE_web
|
closed
|
[Refactor] 로그인 화면 수정
|
Refactor 데이브/김원준 🎨Design
|
### Issue
> 로그인 화면 수정
### ToDoList
- [ ] 카카오 로그인 버튼 수정
### 추가 선택사항(Optional)
|
1.0
|
[Refactor] 로그인 화면 수정 - ### Issue
> 로그인 화면 수정
### ToDoList
- [ ] 카카오 로그인 버튼 수정
### 추가 선택사항(Optional)
|
non_process
|
로그인 화면 수정 issue 로그인 화면 수정 todolist 카카오 로그인 버튼 수정 추가 선택사항 optional
| 0
|
249,125
| 26,887,011,761
|
IssuesEvent
|
2023-02-06 04:46:04
|
MendDemo/PoiMakia
|
https://api.github.com/repos/MendDemo/PoiMakia
|
opened
|
jsoup-1.9.2.jar: 2 vulnerabilities (highest severity is: 7.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsoup-1.9.2.jar</b></p></summary>
<p>jsoup HTML parser</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar,/lib/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/MendDemo/PoiMakia/commit/47d070b80133c8fe80f3796674688596a612e8dd">47d070b80133c8fe80f3796674688596a612e8dd</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jsoup version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-37714](https://www.mend.io/vulnerability-database/CVE-2021-37714) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | jsoup-1.9.2.jar | Direct | 1.14.2 | ❌ |
| [CVE-2022-36033](https://www.mend.io/vulnerability-database/CVE-2022-36033) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jsoup-1.9.2.jar | Direct | 1.15.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-37714</summary>
### Vulnerable Library - <b>jsoup-1.9.2.jar</b></p>
<p>jsoup HTML parser</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar,/lib/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jsoup-1.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MendDemo/PoiMakia/commit/47d070b80133c8fe80f3796674688596a612e8dd">47d070b80133c8fe80f3796674688596a612e8dd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jsoup is a Java library for working with HTML. Those using jsoup versions prior to 1.14.2 to parse untrusted HTML or XML may be vulnerable to DOS attacks. If the parser is run on user supplied input, an attacker may supply content that causes the parser to get stuck (loop indefinitely until cancelled), to complete more slowly than usual, or to throw an unexpected exception. This effect may support a denial of service attack. The issue is patched in version 1.14.2. There are a few available workarounds. Users may rate limit input parsing, limit the size of inputs based on system resources, and/or implement thread watchdogs to cap and timeout parse runtimes.
<p>Publish Date: 2021-08-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-37714>CVE-2021-37714</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://jsoup.org/news/release-1.14.2">https://jsoup.org/news/release-1.14.2</a></p>
<p>Release Date: 2021-08-18</p>
<p>Fix Resolution: 1.14.2</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-36033</summary>
### Vulnerable Library - <b>jsoup-1.9.2.jar</b></p>
<p>jsoup HTML parser</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar,/lib/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jsoup-1.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MendDemo/PoiMakia/commit/47d070b80133c8fe80f3796674688596a612e8dd">47d070b80133c8fe80f3796674688596a612e8dd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jsoup is a Java HTML parser, built for HTML editing, cleaning, scraping, and cross-site scripting (XSS) safety. jsoup may incorrectly sanitize HTML including `javascript:` URL expressions, which could allow XSS attacks when a reader subsequently clicks that link. If the non-default `SafeList.preserveRelativeLinks` option is enabled, HTML including `javascript:` URLs that have been crafted with control characters will not be sanitized. If the site that this HTML is published on does not set a Content Security Policy, an XSS attack is then possible. This issue is patched in jsoup 1.15.3. Users should upgrade to this version. Additionally, as the unsanitized input may have been persisted, old content should be cleaned again using the updated version. To remediate this issue without immediately upgrading: - disable `SafeList.preserveRelativeLinks`, which will rewrite input URLs as absolute URLs - ensure an appropriate [Content Security Policy](https://developer.mozilla.org/en-US/docs/Web/HTTP/CSP) is defined. (This should be used regardless of upgrading, as a defence-in-depth best practice.)
<p>Publish Date: 2022-08-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-36033>CVE-2022-36033</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jhy/jsoup/security/advisories/GHSA-gp7f-rwcx-9369">https://github.com/jhy/jsoup/security/advisories/GHSA-gp7f-rwcx-9369</a></p>
<p>Release Date: 2022-08-29</p>
<p>Fix Resolution: 1.15.3</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
jsoup-1.9.2.jar: 2 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsoup-1.9.2.jar</b></p></summary>
<p>jsoup HTML parser</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar,/lib/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/MendDemo/PoiMakia/commit/47d070b80133c8fe80f3796674688596a612e8dd">47d070b80133c8fe80f3796674688596a612e8dd</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jsoup version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-37714](https://www.mend.io/vulnerability-database/CVE-2021-37714) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | jsoup-1.9.2.jar | Direct | 1.14.2 | ❌ |
| [CVE-2022-36033](https://www.mend.io/vulnerability-database/CVE-2022-36033) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jsoup-1.9.2.jar | Direct | 1.15.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-37714</summary>
### Vulnerable Library - <b>jsoup-1.9.2.jar</b></p>
<p>jsoup HTML parser</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar,/lib/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jsoup-1.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MendDemo/PoiMakia/commit/47d070b80133c8fe80f3796674688596a612e8dd">47d070b80133c8fe80f3796674688596a612e8dd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jsoup is a Java library for working with HTML. Those using jsoup versions prior to 1.14.2 to parse untrusted HTML or XML may be vulnerable to DOS attacks. If the parser is run on user supplied input, an attacker may supply content that causes the parser to get stuck (loop indefinitely until cancelled), to complete more slowly than usual, or to throw an unexpected exception. This effect may support a denial of service attack. The issue is patched in version 1.14.2. There are a few available workarounds. Users may rate limit input parsing, limit the size of inputs based on system resources, and/or implement thread watchdogs to cap and timeout parse runtimes.
<p>Publish Date: 2021-08-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-37714>CVE-2021-37714</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://jsoup.org/news/release-1.14.2">https://jsoup.org/news/release-1.14.2</a></p>
<p>Release Date: 2021-08-18</p>
<p>Fix Resolution: 1.14.2</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-36033</summary>
### Vulnerable Library - <b>jsoup-1.9.2.jar</b></p>
<p>jsoup HTML parser</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar,/lib/org/jsoup/jsoup/1.9.2/jsoup-1.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jsoup-1.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MendDemo/PoiMakia/commit/47d070b80133c8fe80f3796674688596a612e8dd">47d070b80133c8fe80f3796674688596a612e8dd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jsoup is a Java HTML parser, built for HTML editing, cleaning, scraping, and cross-site scripting (XSS) safety. jsoup may incorrectly sanitize HTML including `javascript:` URL expressions, which could allow XSS attacks when a reader subsequently clicks that link. If the non-default `SafeList.preserveRelativeLinks` option is enabled, HTML including `javascript:` URLs that have been crafted with control characters will not be sanitized. If the site that this HTML is published on does not set a Content Security Policy, an XSS attack is then possible. This issue is patched in jsoup 1.15.3. Users should upgrade to this version. Additionally, as the unsanitized input may have been persisted, old content should be cleaned again using the updated version. To remediate this issue without immediately upgrading: - disable `SafeList.preserveRelativeLinks`, which will rewrite input URLs as absolute URLs - ensure an appropriate [Content Security Policy](https://developer.mozilla.org/en-US/docs/Web/HTTP/CSP) is defined. (This should be used regardless of upgrading, as a defence-in-depth best practice.)
<p>Publish Date: 2022-08-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-36033>CVE-2022-36033</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jhy/jsoup/security/advisories/GHSA-gp7f-rwcx-9369">https://github.com/jhy/jsoup/security/advisories/GHSA-gp7f-rwcx-9369</a></p>
<p>Release Date: 2022-08-29</p>
<p>Fix Resolution: 1.15.3</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_process
|
jsoup jar vulnerabilities highest severity is vulnerable library jsoup jar jsoup html parser path to dependency file pom xml path to vulnerable library home wss scanner repository org jsoup jsoup jsoup jar lib org jsoup jsoup jsoup jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in jsoup version remediation available high jsoup jar direct medium jsoup jar direct details cve vulnerable library jsoup jar jsoup html parser path to dependency file pom xml path to vulnerable library home wss scanner repository org jsoup jsoup jsoup jar lib org jsoup jsoup jsoup jar dependency hierarchy x jsoup jar vulnerable library found in head commit a href found in base branch master vulnerability details jsoup is a java library for working with html those using jsoup versions prior to to parse untrusted html or xml may be vulnerable to dos attacks if the parser is run on user supplied input an attacker may supply content that causes the parser to get stuck loop indefinitely until cancelled to complete more slowly than usual or to throw an unexpected exception this effect may support a denial of service attack the issue is patched in version there are a few available workarounds users may rate limit input parsing limit the size of inputs based on system resources and or implement thread watchdogs to cap and timeout parse runtimes publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend cve vulnerable library jsoup jar jsoup html parser path to dependency file pom xml path to vulnerable library home wss scanner repository org jsoup jsoup jsoup jar lib org jsoup jsoup jsoup jar dependency hierarchy x jsoup jar vulnerable library found in head commit a href found in base branch master vulnerability details jsoup is a java html parser built for html editing cleaning scraping and cross site scripting xss safety jsoup may incorrectly sanitize html including javascript url expressions which could allow xss attacks when a reader subsequently clicks that link if the non default safelist preserverelativelinks option is enabled html including javascript urls that have been crafted with control characters will not be sanitized if the site that this html is published on does not set a content security policy an xss attack is then possible this issue is patched in jsoup users should upgrade to this version additionally as the unsanitized input may have been persisted old content should be cleaned again using the updated version to remediate this issue without immediately upgrading disable safelist preserverelativelinks which will rewrite input urls as absolute urls ensure an appropriate is defined this should be used regardless of upgrading as a defence in depth best practice publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
18,564
| 24,555,759,759
|
IssuesEvent
|
2022-10-12 15:44:44
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Mobile apps] Study activities screen > UI issue
|
Bug P2 iOS Android Process: Fixed Process: Tested QA Process: Tested dev
|
Study activities screen > UI issue > There should be a gap between time and time unit (AM or PM)


|
3.0
|
[Mobile apps] Study activities screen > UI issue - Study activities screen > UI issue > There should be a gap between time and time unit (AM or PM)


|
process
|
study activities screen ui issue study activities screen ui issue there should be a gap between time and time unit am or pm
| 1
|
2,607
| 5,367,273,329
|
IssuesEvent
|
2017-02-22 03:23:44
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
closed
|
[subtitles] [fr] « La France est le pays européen qui utilise le plus de pesticides »
|
Language: French Process: [6] Approved
|
# Video title
« La France est le pays européen qui utilise le plus de pesticides »
# URL
https://www.youtube.com/watch?v=Ju3vITxmjQs&t=4s
# Youtube subtitles language
French
# Duration
1:09
# Subtitles URL
https://www.youtube.com/timedtext_editor?bl=vmp&lang=fr&ref=player&action_mde_edit_form=1&ui=hd&v=Ju3vITxmjQs&tab=captions
|
1.0
|
[subtitles] [fr] « La France est le pays européen qui utilise le plus de pesticides » - # Video title
« La France est le pays européen qui utilise le plus de pesticides »
# URL
https://www.youtube.com/watch?v=Ju3vITxmjQs&t=4s
# Youtube subtitles language
French
# Duration
1:09
# Subtitles URL
https://www.youtube.com/timedtext_editor?bl=vmp&lang=fr&ref=player&action_mde_edit_form=1&ui=hd&v=Ju3vITxmjQs&tab=captions
|
process
|
« la france est le pays européen qui utilise le plus de pesticides » video title « la france est le pays européen qui utilise le plus de pesticides » url youtube subtitles language french duration subtitles url
| 1
|
27,660
| 5,075,252,152
|
IssuesEvent
|
2016-12-27 18:36:14
|
phingofficial/phing
|
https://api.github.com/repos/phingofficial/phing
|
closed
|
Repeated prompt when Input task is used within Then block (Trac #1006)
|
defect Migrated from Trac phing-core
|
Using an input task within the `then` block of an `if` statement causes the text or message to be printed four times instead of one.
Some exploration indicates that the addText method is being called four times, so it appends the text to itself.
---
build.xml:
``` text
<?xml version="1.0"?>
<project name="test" default="test">
<target name="test">
<if>
<equals arg1="true" arg2="true" />
<then>
<input propertyname="test" defaultValue="no" validargs="yes,no">Phing?</input>
</then>
</if>
</target>
</project>
```
Expected:
``` text
Phing?(yes/no) [no]
```
Actual:
``` text
Phing?Phing?Phing?Phing?(yes/no) [no]
```
Migrated from https://www.phing.info/trac/ticket/1006
``` json
{
"status": "new",
"changetime": "2013-06-14T23:43:47",
"description": "Using an input task within the `then` block of an `if` statement causes the text or message to be printed four times instead of one.\n\nSome exploration indicates that the addText method is being called four times, so it appends the text to itself.\n\n----\nbuild.xml:\n{{{\n<?xml version=\"1.0\"?>\n<project name=\"test\" default=\"test\">\n\n\t<target name=\"test\">\n\t\t<if>\n\t\t\t<equals arg1=\"true\" arg2=\"true\" />\n\t\t\t<then>\n\t\t\t\t<input propertyname=\"test\" defaultValue=\"no\" validargs=\"yes,no\">Phing?</input>\n\t\t\t</then>\n\t\t</if>\n\t</target>\n\n</project>\n}}}\n\nExpected:\n{{{\nPhing?(yes/no) [no]\n}}}\n\nActual:\n{{{\nPhing?Phing?Phing?Phing?(yes/no) [no]\n}}}",
"reporter": "doug.fitzmaurice@ents24.com",
"cc": "",
"resolution": "",
"_ts": "1371253427358917",
"component": "phing-core",
"summary": "Repeated prompt when Input task is used within Then block",
"priority": "tbd",
"keywords": "",
"version": "2.5.0",
"time": "2013-04-19T12:42:56",
"milestone": "",
"owner": "mrook",
"type": "defect"
}
```
|
1.0
|
Repeated prompt when Input task is used within Then block (Trac #1006) - Using an input task within the `then` block of an `if` statement causes the text or message to be printed four times instead of one.
Some exploration indicates that the addText method is being called four times, so it appends the text to itself.
---
build.xml:
``` text
<?xml version="1.0"?>
<project name="test" default="test">
<target name="test">
<if>
<equals arg1="true" arg2="true" />
<then>
<input propertyname="test" defaultValue="no" validargs="yes,no">Phing?</input>
</then>
</if>
</target>
</project>
```
Expected:
``` text
Phing?(yes/no) [no]
```
Actual:
``` text
Phing?Phing?Phing?Phing?(yes/no) [no]
```
Migrated from https://www.phing.info/trac/ticket/1006
``` json
{
"status": "new",
"changetime": "2013-06-14T23:43:47",
"description": "Using an input task within the `then` block of an `if` statement causes the text or message to be printed four times instead of one.\n\nSome exploration indicates that the addText method is being called four times, so it appends the text to itself.\n\n----\nbuild.xml:\n{{{\n<?xml version=\"1.0\"?>\n<project name=\"test\" default=\"test\">\n\n\t<target name=\"test\">\n\t\t<if>\n\t\t\t<equals arg1=\"true\" arg2=\"true\" />\n\t\t\t<then>\n\t\t\t\t<input propertyname=\"test\" defaultValue=\"no\" validargs=\"yes,no\">Phing?</input>\n\t\t\t</then>\n\t\t</if>\n\t</target>\n\n</project>\n}}}\n\nExpected:\n{{{\nPhing?(yes/no) [no]\n}}}\n\nActual:\n{{{\nPhing?Phing?Phing?Phing?(yes/no) [no]\n}}}",
"reporter": "doug.fitzmaurice@ents24.com",
"cc": "",
"resolution": "",
"_ts": "1371253427358917",
"component": "phing-core",
"summary": "Repeated prompt when Input task is used within Then block",
"priority": "tbd",
"keywords": "",
"version": "2.5.0",
"time": "2013-04-19T12:42:56",
"milestone": "",
"owner": "mrook",
"type": "defect"
}
```
|
non_process
|
repeated prompt when input task is used within then block trac using an input task within the then block of an if statement causes the text or message to be printed four times instead of one some exploration indicates that the addtext method is being called four times so it appends the text to itself build xml text phing expected text phing yes no actual text phing phing phing phing yes no migrated from json status new changetime description using an input task within the then block of an if statement causes the text or message to be printed four times instead of one n nsome exploration indicates that the addtext method is being called four times so it appends the text to itself n n nbuild xml n n n n n t n t t n t t t n t t t n t t t t phing n t t t n t t n t n n n n nexpected n nphing yes no n n nactual n nphing phing phing phing yes no n reporter doug fitzmaurice com cc resolution ts component phing core summary repeated prompt when input task is used within then block priority tbd keywords version time milestone owner mrook type defect
| 0
|
233,485
| 7,698,238,764
|
IssuesEvent
|
2018-05-18 22:06:29
|
trailofbits/echidna
|
https://api.github.com/repos/trailofbits/echidna
|
opened
|
Inconsistent test results in a simple contract (2)
|
bug help wanted high-priority
|
In this simple contract:
```solidity
contract A {}
contract B {
function f() {
return;
}
}
contract TEST {
A[] private xs;
B b = new B();
function g() public {
return;
}
//function add() {
// xs.push(new A());
//}
function echidna_true() returns (bool) {
b.f();
return true;
}
}
```
the echidna_true test cannot fail and Echida works as expected. However it reports a failed test **when you uncomment the add() function**. For instance:
```bash
$ echidna-test test.sol test.sol:TEST
━━━ test.sol ━━━
✗ "echidna_true" failed after 2 tests.
│ Call sequence: add();
✗ 1 failed.
```
|
1.0
|
Inconsistent test results in a simple contract (2) - In this simple contract:
```solidity
contract A {}
contract B {
function f() {
return;
}
}
contract TEST {
A[] private xs;
B b = new B();
function g() public {
return;
}
//function add() {
// xs.push(new A());
//}
function echidna_true() returns (bool) {
b.f();
return true;
}
}
```
the echidna_true test cannot fail and Echida works as expected. However it reports a failed test **when you uncomment the add() function**. For instance:
```bash
$ echidna-test test.sol test.sol:TEST
━━━ test.sol ━━━
✗ "echidna_true" failed after 2 tests.
│ Call sequence: add();
✗ 1 failed.
```
|
non_process
|
inconsistent test results in a simple contract in this simple contract solidity contract a contract b function f return contract test a private xs b b new b function g public return function add xs push new a function echidna true returns bool b f return true the echidna true test cannot fail and echida works as expected however it reports a failed test when you uncomment the add function for instance bash echidna test test sol test sol test ━━━ test sol ━━━ ✗ echidna true failed after tests │ call sequence add ✗ failed
| 0
|
17,080
| 12,219,420,058
|
IssuesEvent
|
2020-05-01 21:41:28
|
enarx/enarx
|
https://api.github.com/repos/enarx/enarx
|
opened
|
Investigate a 'cargo-make' feature that tests all combinations of features
|
infrastructure research
|
Some of our crates conditionally compile code depending on if a feature flag is enabled.
`cargo-make` can enable all features or certain features, but can it test all different combinations of features automatically?
|
1.0
|
Investigate a 'cargo-make' feature that tests all combinations of features - Some of our crates conditionally compile code depending on if a feature flag is enabled.
`cargo-make` can enable all features or certain features, but can it test all different combinations of features automatically?
|
non_process
|
investigate a cargo make feature that tests all combinations of features some of our crates conditionally compile code depending on if a feature flag is enabled cargo make can enable all features or certain features but can it test all different combinations of features automatically
| 0
|
9,489
| 12,483,529,912
|
IssuesEvent
|
2020-05-30 09:52:46
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
export size not correct
|
bug: pending difficulty: hard priority: high reproduce: confirmed scope: image processing
|
<!-- IMPORTANT
Bug reports that do not make an effort to help the developers will be closed without notice.
Make sure that this bug has not already been opened and/or closed by searching the issues on GitHub, as duplicate bug reports will be closed.
A bug report simply stating that Darktable crashes is unhelpful, so please fill in most of the items below and provide detailed information.
-->
**Describe the bug**
An attempt to fix this issue has been made in #5022. But this has introduced a regression.
**To Reproduce**
$ cd src/tests/integration
$ darktable-cli --width 2048 --height 2048 --hq true images/mire1.cr2 0001-exposure/exposure.xmp output.png
$ exiv2 output.png | grep -i "Image size"
Image size : 2049 x 1365
**Expected behavior**
Reverting : c12e3d3 and d881b85
The export size is correct and 2048 x 1364
$ exiv2 output.png | grep -i "Image size"
Image size : 2048 x 1364
**Platform (please complete the following information):**
- Darktable Version: master
- OS: GNU/Linux
**Additional context**
We should either revert commits c12e3d3 and d881b85 or have a fix for 3.2.
|
1.0
|
export size not correct - <!-- IMPORTANT
Bug reports that do not make an effort to help the developers will be closed without notice.
Make sure that this bug has not already been opened and/or closed by searching the issues on GitHub, as duplicate bug reports will be closed.
A bug report simply stating that Darktable crashes is unhelpful, so please fill in most of the items below and provide detailed information.
-->
**Describe the bug**
An attempt to fix this issue has been made in #5022. But this has introduced a regression.
**To Reproduce**
$ cd src/tests/integration
$ darktable-cli --width 2048 --height 2048 --hq true images/mire1.cr2 0001-exposure/exposure.xmp output.png
$ exiv2 output.png | grep -i "Image size"
Image size : 2049 x 1365
**Expected behavior**
Reverting : c12e3d3 and d881b85
The export size is correct and 2048 x 1364
$ exiv2 output.png | grep -i "Image size"
Image size : 2048 x 1364
**Platform (please complete the following information):**
- Darktable Version: master
- OS: GNU/Linux
**Additional context**
We should either revert commits c12e3d3 and d881b85 or have a fix for 3.2.
|
process
|
export size not correct important bug reports that do not make an effort to help the developers will be closed without notice make sure that this bug has not already been opened and or closed by searching the issues on github as duplicate bug reports will be closed a bug report simply stating that darktable crashes is unhelpful so please fill in most of the items below and provide detailed information describe the bug an attempt to fix this issue has been made in but this has introduced a regression to reproduce cd src tests integration darktable cli width height hq true images exposure exposure xmp output png output png grep i image size image size x expected behavior reverting and the export size is correct and x output png grep i image size image size x platform please complete the following information darktable version master os gnu linux additional context we should either revert commits and or have a fix for
| 1
|
1,418
| 2,514,194,332
|
IssuesEvent
|
2015-01-15 09:10:15
|
georchestra/georchestra
|
https://api.github.com/repos/georchestra/georchestra
|
opened
|
doc - native libs
|
1 - Ready enhancement priority-top
|
Quoting gaston:
```
10:06 < gaston> finalement entre https://github.com/georchestra/georchestra/blob/master/geoserver/NATIVE_LIBS.md et
https://github.com/georchestra/georchestra/blob/master/doc/setup/native_libs.md...
10:07 < gaston> faut suivre lequel ? :)
```
|
1.0
|
doc - native libs - Quoting gaston:
```
10:06 < gaston> finalement entre https://github.com/georchestra/georchestra/blob/master/geoserver/NATIVE_LIBS.md et
https://github.com/georchestra/georchestra/blob/master/doc/setup/native_libs.md...
10:07 < gaston> faut suivre lequel ? :)
```
|
non_process
|
doc native libs quoting gaston finalement entre et faut suivre lequel
| 0
|
5,962
| 8,785,529,783
|
IssuesEvent
|
2018-12-20 13:17:37
|
lutraconsulting/qgis-crayfish-plugin
|
https://api.github.com/repos/lutraconsulting/qgis-crayfish-plugin
|
closed
|
Export mesh&datasets to vector format
|
enhancement processing
|
We need to add processing algorithm in crayfish plugin to be able to export raw MDAL data to geopackage.
MDAL supports data defined on vertices and faces (https://github.com/qgis/QGIS-Enhancement-Proposals/issues/119). Also it defines scalar data and vector data (x,y).
relevant docs: https://github.com/qgis/QGIS/blob/master/src/core/mesh/qgsmeshlayer.h
https://github.com/qgis/QGIS/blob/master/src/core/mesh/qgsmeshdataprovider.h
test data: https://github.com/lutraconsulting/MDAL/tree/master/tests (2dm + ascii_dat)
a) For face datasets, one feature in the vector layer will be one face. Attributes of the face will be dataset values in particular timestep, e.g. let say we have dataset "Depth" and "Velocity"
so attributes will be
Depth = 1
Velocity_x = 1.0
Velocity_y = 2.0
....
b) for vertex datasets, the attributes will be the same, but the exported features will be nodes (vertices)
note: user will be able to export just one (selected timestep!). If one wants to export more/all, best to run this algorithm in batch mode. Note that some datasets have just 1 value (e.g. Depth/Maximums). in this case it does not matter which timestep user selects, it is always exported
-------
As for graphical side, one will be able to select
- INPUT: mesh layer
- INPUT: dataset groups (multichoice, ideally "all/none" button)
- INPUT: a timestep (combo box)
and
- OUTPUT: geopackage filename where to put export or "in-memory" option
the input/output UX should be in-line with most used processing algorithms so user knows where to click.
-----
```
m = iface.activeLayer()
dp = m.dataProvider()
dp
<qgis._core.QgsMeshDataProvider object at 0x12d194678>
mesh = QgsMesh()
dp.populateMesh(mesh)
mesh.faceCount()
1600
mesh.face(0)
[34, 33, 0, 1]
mesh.vertexCount()
1683
mesh.vertex(0)
<QgsPoint: PointZ (20 10 0)>
dp.datasetGroupCount()
1
groupMeta = dp.datasetGroupMetadata(0)
groupMeta.isScalar()
False
dp.datasetCount(QgsMeshDatasetIndex(0))
27
values = dp.datasetValues(QgsMeshDatasetIndex(0, 2), 0, 300)
value = values.value(0)
value
<qgis._core.QgsMeshDatasetValue object at 0x12d194948>
value.x()
nan
value.y()
nan
```
|
1.0
|
Export mesh&datasets to vector format - We need to add processing algorithm in crayfish plugin to be able to export raw MDAL data to geopackage.
MDAL supports data defined on vertices and faces (https://github.com/qgis/QGIS-Enhancement-Proposals/issues/119). Also it defines scalar data and vector data (x,y).
relevant docs: https://github.com/qgis/QGIS/blob/master/src/core/mesh/qgsmeshlayer.h
https://github.com/qgis/QGIS/blob/master/src/core/mesh/qgsmeshdataprovider.h
test data: https://github.com/lutraconsulting/MDAL/tree/master/tests (2dm + ascii_dat)
a) For face datasets, one feature in the vector layer will be one face. Attributes of the face will be dataset values in particular timestep, e.g. let say we have dataset "Depth" and "Velocity"
so attributes will be
Depth = 1
Velocity_x = 1.0
Velocity_y = 2.0
....
b) for vertex datasets, the attributes will be the same, but the exported features will be nodes (vertices)
note: user will be able to export just one (selected timestep!). If one wants to export more/all, best to run this algorithm in batch mode. Note that some datasets have just 1 value (e.g. Depth/Maximums). in this case it does not matter which timestep user selects, it is always exported
-------
As for graphical side, one will be able to select
- INPUT: mesh layer
- INPUT: dataset groups (multichoice, ideally "all/none" button)
- INPUT: a timestep (combo box)
and
- OUTPUT: geopackage filename where to put export or "in-memory" option
the input/output UX should be in-line with most used processing algorithms so user knows where to click.
-----
```
m = iface.activeLayer()
dp = m.dataProvider()
dp
<qgis._core.QgsMeshDataProvider object at 0x12d194678>
mesh = QgsMesh()
dp.populateMesh(mesh)
mesh.faceCount()
1600
mesh.face(0)
[34, 33, 0, 1]
mesh.vertexCount()
1683
mesh.vertex(0)
<QgsPoint: PointZ (20 10 0)>
dp.datasetGroupCount()
1
groupMeta = dp.datasetGroupMetadata(0)
groupMeta.isScalar()
False
dp.datasetCount(QgsMeshDatasetIndex(0))
27
values = dp.datasetValues(QgsMeshDatasetIndex(0, 2), 0, 300)
value = values.value(0)
value
<qgis._core.QgsMeshDatasetValue object at 0x12d194948>
value.x()
nan
value.y()
nan
```
|
process
|
export mesh datasets to vector format we need to add processing algorithm in crayfish plugin to be able to export raw mdal data to geopackage mdal supports data defined on vertices and faces also it defines scalar data and vector data x y relevant docs test data ascii dat a for face datasets one feature in the vector layer will be one face attributes of the face will be dataset values in particular timestep e g let say we have dataset depth and velocity so attributes will be depth velocity x velocity y b for vertex datasets the attributes will be the same but the exported features will be nodes vertices note user will be able to export just one selected timestep if one wants to export more all best to run this algorithm in batch mode note that some datasets have just value e g depth maximums in this case it does not matter which timestep user selects it is always exported as for graphical side one will be able to select input mesh layer input dataset groups multichoice ideally all none button input a timestep combo box and output geopackage filename where to put export or in memory option the input output ux should be in line with most used processing algorithms so user knows where to click m iface activelayer dp m dataprovider dp mesh qgsmesh dp populatemesh mesh mesh facecount mesh face mesh vertexcount mesh vertex dp datasetgroupcount groupmeta dp datasetgroupmetadata groupmeta isscalar false dp datasetcount qgsmeshdatasetindex values dp datasetvalues qgsmeshdatasetindex value values value value value x nan value y nan
| 1
|
313,420
| 9,561,330,726
|
IssuesEvent
|
2019-05-03 22:45:07
|
AugurProject/augur
|
https://api.github.com/repos/AugurProject/augur
|
closed
|
Update v1 docs with recent Augur Node changes
|
Chore Priority: Medium
|
This is mostly for changes made to `getUserTradingPositions` and `getReportingFees`:
https://github.com/AugurProject/augur-node/pull/821
https://github.com/AugurProject/augur-node/pull/834
https://github.com/AugurProject/augur-node/pull/840
https://github.com/AugurProject/augur-node/pull/846
https://github.com/AugurProject/augur-node/pull/854
Also includes a few other miscellaneous recent changes to Augur Node.
|
1.0
|
Update v1 docs with recent Augur Node changes - This is mostly for changes made to `getUserTradingPositions` and `getReportingFees`:
https://github.com/AugurProject/augur-node/pull/821
https://github.com/AugurProject/augur-node/pull/834
https://github.com/AugurProject/augur-node/pull/840
https://github.com/AugurProject/augur-node/pull/846
https://github.com/AugurProject/augur-node/pull/854
Also includes a few other miscellaneous recent changes to Augur Node.
|
non_process
|
update docs with recent augur node changes this is mostly for changes made to getusertradingpositions and getreportingfees also includes a few other miscellaneous recent changes to augur node
| 0
|
59,680
| 24,849,392,240
|
IssuesEvent
|
2022-10-26 18:38:50
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
[Enhancement] Address UI shortcomings with status updates
|
Type: Bug Report Service: Dev Need: 2-Should Have Type: Enhancement Product: Moped Project: Moped v2.0
|
- [x] add linebreaks to project types and project partners
- [x] add border radius to highlighted div
- [x] autofocus inputs
- [x] Set cursor to "pointer" when hovering over editable field
- [x] fix up status update highlight to not highlight the label
- [x] Display ` - ` instead of `None` when a field is blank
<img width="1451" alt="Screen Shot 2022-10-06 at 2 55 23 PM" src="https://user-images.githubusercontent.com/14793120/194395770-4951a601-0aa4-4f46-8464-037653125c2c.png">
|
1.0
|
[Enhancement] Address UI shortcomings with status updates - - [x] add linebreaks to project types and project partners
- [x] add border radius to highlighted div
- [x] autofocus inputs
- [x] Set cursor to "pointer" when hovering over editable field
- [x] fix up status update highlight to not highlight the label
- [x] Display ` - ` instead of `None` when a field is blank
<img width="1451" alt="Screen Shot 2022-10-06 at 2 55 23 PM" src="https://user-images.githubusercontent.com/14793120/194395770-4951a601-0aa4-4f46-8464-037653125c2c.png">
|
non_process
|
address ui shortcomings with status updates add linebreaks to project types and project partners add border radius to highlighted div autofocus inputs set cursor to pointer when hovering over editable field fix up status update highlight to not highlight the label display instead of none when a field is blank img width alt screen shot at pm src
| 0
|
825,212
| 31,279,090,152
|
IssuesEvent
|
2023-08-22 08:26:44
|
bitcoin-dev-project/sim-ln
|
https://api.github.com/repos/bitcoin-dev-project/sim-ln
|
opened
|
Feature: Generate Activity From Graph Topology
|
feature Medium Priority
|
Rather than require a user to provide a description of activity, provide a mode that will randomly generate activity based on the graph + some heuristics, eg:
* Number of channels a node has
* Size of channels
This is primarily aimed to serve the "regtest" case where the person running the simulation has access to all of the nodes, and can just easily spin up the simulation to run random activity.
|
1.0
|
Feature: Generate Activity From Graph Topology - Rather than require a user to provide a description of activity, provide a mode that will randomly generate activity based on the graph + some heuristics, eg:
* Number of channels a node has
* Size of channels
This is primarily aimed to serve the "regtest" case where the person running the simulation has access to all of the nodes, and can just easily spin up the simulation to run random activity.
|
non_process
|
feature generate activity from graph topology rather than require a user to provide a description of activity provide a mode that will randomly generate activity based on the graph some heuristics eg number of channels a node has size of channels this is primarily aimed to serve the regtest case where the person running the simulation has access to all of the nodes and can just easily spin up the simulation to run random activity
| 0
|
2,805
| 5,738,493,411
|
IssuesEvent
|
2017-04-23 04:47:19
|
g8os/core0
|
https://api.github.com/repos/g8os/core0
|
closed
|
Installer for G8OS
|
process_wontfix
|
We need an installer that prepare the disks and filesystem on the host machine.
Reason for this is the G8OS FS needs to reserve some disk space to store its backend files.
Also now that we will have to include extra tools then the one from the initramfs (qemu/kvm, libvirt), having a dedicated space for the core0 on a disk to store these file is also required.
|
1.0
|
Installer for G8OS - We need an installer that prepare the disks and filesystem on the host machine.
Reason for this is the G8OS FS needs to reserve some disk space to store its backend files.
Also now that we will have to include extra tools then the one from the initramfs (qemu/kvm, libvirt), having a dedicated space for the core0 on a disk to store these file is also required.
|
process
|
installer for we need an installer that prepare the disks and filesystem on the host machine reason for this is the fs needs to reserve some disk space to store its backend files also now that we will have to include extra tools then the one from the initramfs qemu kvm libvirt having a dedicated space for the on a disk to store these file is also required
| 1
|
168,524
| 14,163,644,763
|
IssuesEvent
|
2020-11-12 02:54:20
|
pi-fatec-bd/semaforo-consumidor
|
https://api.github.com/repos/pi-fatec-bd/semaforo-consumidor
|
closed
|
Como é o Calculo do Score
|
documentation front-end
|
O usuário deve entender como é calculado o score através da página inicial.
**Requisito relacionado:** R19
**Nome da branch:** explicar-score
|
1.0
|
Como é o Calculo do Score - O usuário deve entender como é calculado o score através da página inicial.
**Requisito relacionado:** R19
**Nome da branch:** explicar-score
|
non_process
|
como é o calculo do score o usuário deve entender como é calculado o score através da página inicial requisito relacionado nome da branch explicar score
| 0
|
445,763
| 31,280,890,117
|
IssuesEvent
|
2023-08-22 09:30:40
|
openstack-k8s-operators/install_yamls
|
https://api.github.com/repos/openstack-k8s-operators/install_yamls
|
closed
|
Getting kubectl: command not found error
|
bug documentation
|
Getting below issue when running make openstack command using crc,
~/install_yamls/out/operator/baremetal-operator
~/install_yamls/out/operator/baremetal-operator ~/install_yamls/out/operator/baremetal-operator
tools/deploy.sh: line 213: kubectl: command not found
make: *** [Makefile:384: crc_bmo_setup] Error 127
|
1.0
|
Getting kubectl: command not found error - Getting below issue when running make openstack command using crc,
~/install_yamls/out/operator/baremetal-operator
~/install_yamls/out/operator/baremetal-operator ~/install_yamls/out/operator/baremetal-operator
tools/deploy.sh: line 213: kubectl: command not found
make: *** [Makefile:384: crc_bmo_setup] Error 127
|
non_process
|
getting kubectl command not found error getting below issue when running make openstack command using crc install yamls out operator baremetal operator install yamls out operator baremetal operator install yamls out operator baremetal operator tools deploy sh line kubectl command not found make error
| 0
|
13,635
| 16,255,150,816
|
IssuesEvent
|
2021-05-08 03:06:57
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
copr: Roadmap to chunk-based Enum/Set support in TiKV
|
sig/coprocessor type/enhancement
|
## Goal
Implement RFC https://github.com/tikv/rfcs/pull/57
## Roadmap
- [x] Add Enum/Set type #8849
- [x] Add ChunkedVecEnum/ChunkedVecSet type #8948 #8988
- [x] Add enum/set into ScalarValue and VectorValue #9021
- [x] Implement enum/set related copr functions #9133
- [x] Implement enum/set related aggr functions
- [x] Count #9143
- [x] First #9135
- [x] Sum #9148 #9184
- [x] Avg #9186
- [x] Max/Min #9146 #9184
---
- [ ] Add `elems` in FieldType (tipb)
- [ ] Export `elems` from praser to tipb (tidb)
- [ ] Extrace `elems` from FieldType to construct ChunkedVecEnum/ChunkedVecSet
|
1.0
|
copr: Roadmap to chunk-based Enum/Set support in TiKV - ## Goal
Implement RFC https://github.com/tikv/rfcs/pull/57
## Roadmap
- [x] Add Enum/Set type #8849
- [x] Add ChunkedVecEnum/ChunkedVecSet type #8948 #8988
- [x] Add enum/set into ScalarValue and VectorValue #9021
- [x] Implement enum/set related copr functions #9133
- [x] Implement enum/set related aggr functions
- [x] Count #9143
- [x] First #9135
- [x] Sum #9148 #9184
- [x] Avg #9186
- [x] Max/Min #9146 #9184
---
- [ ] Add `elems` in FieldType (tipb)
- [ ] Export `elems` from praser to tipb (tidb)
- [ ] Extrace `elems` from FieldType to construct ChunkedVecEnum/ChunkedVecSet
|
process
|
copr roadmap to chunk based enum set support in tikv goal implement rfc roadmap add enum set type add chunkedvecenum chunkedvecset type add enum set into scalarvalue and vectorvalue implement enum set related copr functions implement enum set related aggr functions count first sum avg max min add elems in fieldtype tipb export elems from praser to tipb tidb extrace elems from fieldtype to construct chunkedvecenum chunkedvecset
| 1
|
4,836
| 7,726,701,323
|
IssuesEvent
|
2018-05-24 22:14:46
|
nion-software/nionswift
|
https://api.github.com/repos/nion-software/nionswift
|
opened
|
Improve error messages when operations are applied to invalid inputs.
|
f - processing level - easy p2 - high type - bug w4 - ready
|
The "error message" right now is to do nothing.
An alert? Or a notification? Something in the output window?
|
1.0
|
Improve error messages when operations are applied to invalid inputs. - The "error message" right now is to do nothing.
An alert? Or a notification? Something in the output window?
|
process
|
improve error messages when operations are applied to invalid inputs the error message right now is to do nothing an alert or a notification something in the output window
| 1
|
16,490
| 5,240,878,633
|
IssuesEvent
|
2017-01-31 14:22:54
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Oracle types in package are not generated
|
C: Code Generation C: DB: Oracle P: High R: Worksforme T: Defect
|
The jOOQ generator does not generate classes for types inside a package if only the package name is given.
Given a package with the name "per_per". If I tell jOOQ to include this package in the code-generation with
```xml
<includes>per_per</includes>
```
It will create classes for all Stored Procedures inside the package. But the types in the package are ignored.
This behavior stays exactly the same if I use this:
```xml
<includes>per_per(\..*)?</includes>
```
Only if I add each type with name then jOOQ will generate classes.
```xml
<includes>per_per
| TYPE_NAME1
| TYPE_NAME2
|...
</includes>
```
I think it would be a great improvement if I only had to specifiy the package name and the jooq generator includes the types and the procedures inside the package.
|
1.0
|
Oracle types in package are not generated - The jOOQ generator does not generate classes for types inside a package if only the package name is given.
Given a package with the name "per_per". If I tell jOOQ to include this package in the code-generation with
```xml
<includes>per_per</includes>
```
It will create classes for all Stored Procedures inside the package. But the types in the package are ignored.
This behavior stays exactly the same if I use this:
```xml
<includes>per_per(\..*)?</includes>
```
Only if I add each type with name then jOOQ will generate classes.
```xml
<includes>per_per
| TYPE_NAME1
| TYPE_NAME2
|...
</includes>
```
I think it would be a great improvement if I only had to specifiy the package name and the jooq generator includes the types and the procedures inside the package.
|
non_process
|
oracle types in package are not generated the jooq generator does not generate classes for types inside a package if only the package name is given given a package with the name per per if i tell jooq to include this package in the code generation with xml per per it will create classes for all stored procedures inside the package but the types in the package are ignored this behavior stays exactly the same if i use this xml per per only if i add each type with name then jooq will generate classes xml per per type type i think it would be a great improvement if i only had to specifiy the package name and the jooq generator includes the types and the procedures inside the package
| 0
|
108,627
| 16,796,206,980
|
IssuesEvent
|
2021-06-16 04:10:41
|
Techini/WebGoat
|
https://api.github.com/repos/Techini/WebGoat
|
opened
|
WS-2020-0293 (Medium) detected in spring-security-web-5.2.0.RELEASE.jar
|
security vulnerability
|
## WS-2020-0293 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.2.0.RELEASE.jar</b></p></summary>
<p>spring-security-web</p>
<p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p>
<p>Path to dependency file: WebGoat/webgoat-integration-tests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.2.0.RELEASE/spring-security-web-5.2.0.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- webwolf-v8.0.0-SNAPSHOT.jar (Root Library)
- spring-boot-starter-security-2.2.0.RELEASE.jar
- :x: **spring-security-web-5.2.0.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Techini/WebGoat/commit/d33cc0e32a0d1b949ff1b85af16890cd452276f8">d33cc0e32a0d1b949ff1b85af16890cd452276f8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Security before 5.2.9, 5.3.7, and 5.4.3 vulnerable to side-channel attacks. Vulnerable versions of Spring Security don't use constant time comparisons for CSRF tokens.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://github.com/spring-projects/spring-security/commit/40e027c56d11b9b4c5071360bfc718165c937784>WS-2020-0293</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/spring-projects/spring-security/issues/9291">https://github.com/spring-projects/spring-security/issues/9291</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: org.springframework.security:spring-security-web:5.2.9,5.3.7,5.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2020-0293 (Medium) detected in spring-security-web-5.2.0.RELEASE.jar - ## WS-2020-0293 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.2.0.RELEASE.jar</b></p></summary>
<p>spring-security-web</p>
<p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p>
<p>Path to dependency file: WebGoat/webgoat-integration-tests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.2.0.RELEASE/spring-security-web-5.2.0.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- webwolf-v8.0.0-SNAPSHOT.jar (Root Library)
- spring-boot-starter-security-2.2.0.RELEASE.jar
- :x: **spring-security-web-5.2.0.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Techini/WebGoat/commit/d33cc0e32a0d1b949ff1b85af16890cd452276f8">d33cc0e32a0d1b949ff1b85af16890cd452276f8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Security before 5.2.9, 5.3.7, and 5.4.3 vulnerable to side-channel attacks. Vulnerable versions of Spring Security don't use constant time comparisons for CSRF tokens.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://github.com/spring-projects/spring-security/commit/40e027c56d11b9b4c5071360bfc718165c937784>WS-2020-0293</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/spring-projects/spring-security/issues/9291">https://github.com/spring-projects/spring-security/issues/9291</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: org.springframework.security:spring-security-web:5.2.9,5.3.7,5.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in spring security web release jar ws medium severity vulnerability vulnerable library spring security web release jar spring security web library home page a href path to dependency file webgoat webgoat integration tests pom xml path to vulnerable library home wss scanner repository org springframework security spring security web release spring security web release jar dependency hierarchy webwolf snapshot jar root library spring boot starter security release jar x spring security web release jar vulnerable library found in head commit a href vulnerability details spring security before and vulnerable to side channel attacks vulnerable versions of spring security don t use constant time comparisons for csrf tokens publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework security spring security web step up your open source security game with whitesource
| 0
|
11,647
| 14,501,161,222
|
IssuesEvent
|
2020-12-11 19:05:50
|
GoogleCloudPlatform/cloud-opensource-java
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-opensource-java
|
closed
|
Shall we stop the Circle CI build?
|
process
|
Shall we remove Circle CI build, for the following reasons?
1. This repository has used Circle CI for Linux (Java 8) build. With the recently-introduced Github Actions that run for Java 8 (and Java 11) on Linux, I don't think we need Circle CI build.
2. Circle CI requires permission to see logs.
<img width="1345" alt="Screen Shot 2020-12-04 at 15 11 08" src="https://user-images.githubusercontent.com/28604/101221098-fcb7f880-3654-11eb-8c37-9fa2f67552c4.png">
3. CI with Github Actions is in line with other Google Cloud Java repositories maintained by Yoshi team.
# Memo on disabling Circle CI
https://support.circleci.com/hc/en-us/articles/360021666393-How-to-stop-building-by-manually-removing-the-CircleCI-webhook-and-deploy-key-from-your-GitHub-repository
- Updated https://circleci.com/hooks/github webhook as inactive.
- Updated protected branch
- not to require Circle CI
- added "units (8)" as required check.
|
1.0
|
Shall we stop the Circle CI build? - Shall we remove Circle CI build, for the following reasons?
1. This repository has used Circle CI for Linux (Java 8) build. With the recently-introduced Github Actions that run for Java 8 (and Java 11) on Linux, I don't think we need Circle CI build.
2. Circle CI requires permission to see logs.
<img width="1345" alt="Screen Shot 2020-12-04 at 15 11 08" src="https://user-images.githubusercontent.com/28604/101221098-fcb7f880-3654-11eb-8c37-9fa2f67552c4.png">
3. CI with Github Actions is in line with other Google Cloud Java repositories maintained by Yoshi team.
# Memo on disabling Circle CI
https://support.circleci.com/hc/en-us/articles/360021666393-How-to-stop-building-by-manually-removing-the-CircleCI-webhook-and-deploy-key-from-your-GitHub-repository
- Updated https://circleci.com/hooks/github webhook as inactive.
- Updated protected branch
- not to require Circle CI
- added "units (8)" as required check.
|
process
|
shall we stop the circle ci build shall we remove circle ci build for the following reasons this repository has used circle ci for linux java build with the recently introduced github actions that run for java and java on linux i don t think we need circle ci build circle ci requires permission to see logs img width alt screen shot at src ci with github actions is in line with other google cloud java repositories maintained by yoshi team memo on disabling circle ci updated webhook as inactive updated protected branch not to require circle ci added units as required check
| 1
|
39,610
| 2,857,393,450
|
IssuesEvent
|
2015-06-02 19:30:40
|
HPI-SWA-Teaching/SWT15-Project-13
|
https://api.github.com/repos/HPI-SWA-Teaching/SWT15-Project-13
|
closed
|
Definition zusätzlicher Methoden
|
priority: important type: feature
|
| | | |
| :---------: | :-------------------------------------------------: | :------------------: |
| _8_ | | _5_ |
| | Als **Anwender** möchte ich **zusätzliche Methoden vermöge der Auswertung eines einprägsamen Ausdrucks definieren**, um **nicht auf die grafische Benutzerschnittstelle der Entwicklungsumgebung angewiesen zu sein**. | |
| _7_ | | _?_ |
|
1.0
|
Definition zusätzlicher Methoden - | | | |
| :---------: | :-------------------------------------------------: | :------------------: |
| _8_ | | _5_ |
| | Als **Anwender** möchte ich **zusätzliche Methoden vermöge der Auswertung eines einprägsamen Ausdrucks definieren**, um **nicht auf die grafische Benutzerschnittstelle der Entwicklungsumgebung angewiesen zu sein**. | |
| _7_ | | _?_ |
|
non_process
|
definition zusätzlicher methoden als anwender möchte ich zusätzliche methoden vermöge der auswertung eines einprägsamen ausdrucks definieren um nicht auf die grafische benutzerschnittstelle der entwicklungsumgebung angewiesen zu sein
| 0
|
121,772
| 26,031,150,466
|
IssuesEvent
|
2022-12-21 21:25:33
|
llvm/llvm-project
|
https://api.github.com/repos/llvm/llvm-project
|
closed
|
`amdgpu-isel` appears to be the name for SelectionDAGISel for some targets?
|
test-suite backend:AMDGPU llvm:codegen
|
I'm trying to write a test for some changes I'm making to SelectionDAGBuilder. The `RUN:` lines of my `FileCheck` tests on LLVM IR have `llc -mtriple=aarch64-linux-gnu /tmp/x.ll -stop-after=finalize-isel -o - -global-isel=0 -fast-isel=0`.
To isolate my test to testing isel, if I add: `-start-before=finalize-isel`, I no longer get any codegen from `llc`. It still prints the input IR and yaml for MIR, but the MIR output is gone.
If I replace `-start-before=finalize-isel` with `-print-before-all`, I see:
`# *** IR Dump Before AArch64 Instruction Selection (amdgpu-isel) ***:`
(huh, why is `amdgpu-isel` in there?)
On a hunch, testing `-start-before=amdgpu-isel` does what I'd have expected from `-start-before=finalize-isel`.
Expected:
`llc -mtriple=aarch64-linux-gnu /tmp/x.ll -start-before=finalize-isel -stop-after=finalize-isel -o -`
Actual:
`llc -mtriple=aarch64-linux-gnu /tmp/x.ll -start-before=amdgpu-isel -stop-after=finalize-isel -o -`
(Though, adding `-print-before-all` to Expected above shows what looks like `amdgpu-isel`, then `aarch64-local-dynamic-tls-cleanup`, then `finalize-isel` being run).
I see the same issue if I change the target triple to `-mtriple=x86_64-linux-gnu` as well, so not specific to aarch64 I think. So not tagging this as related to globalisel (since I don't want to test globalisel anyways).
cc @arsenm @aeubanks
|
1.0
|
`amdgpu-isel` appears to be the name for SelectionDAGISel for some targets? - I'm trying to write a test for some changes I'm making to SelectionDAGBuilder. The `RUN:` lines of my `FileCheck` tests on LLVM IR have `llc -mtriple=aarch64-linux-gnu /tmp/x.ll -stop-after=finalize-isel -o - -global-isel=0 -fast-isel=0`.
To isolate my test to testing isel, if I add: `-start-before=finalize-isel`, I no longer get any codegen from `llc`. It still prints the input IR and yaml for MIR, but the MIR output is gone.
If I replace `-start-before=finalize-isel` with `-print-before-all`, I see:
`# *** IR Dump Before AArch64 Instruction Selection (amdgpu-isel) ***:`
(huh, why is `amdgpu-isel` in there?)
On a hunch, testing `-start-before=amdgpu-isel` does what I'd have expected from `-start-before=finalize-isel`.
Expected:
`llc -mtriple=aarch64-linux-gnu /tmp/x.ll -start-before=finalize-isel -stop-after=finalize-isel -o -`
Actual:
`llc -mtriple=aarch64-linux-gnu /tmp/x.ll -start-before=amdgpu-isel -stop-after=finalize-isel -o -`
(Though, adding `-print-before-all` to Expected above shows what looks like `amdgpu-isel`, then `aarch64-local-dynamic-tls-cleanup`, then `finalize-isel` being run).
I see the same issue if I change the target triple to `-mtriple=x86_64-linux-gnu` as well, so not specific to aarch64 I think. So not tagging this as related to globalisel (since I don't want to test globalisel anyways).
cc @arsenm @aeubanks
|
non_process
|
amdgpu isel appears to be the name for selectiondagisel for some targets i m trying to write a test for some changes i m making to selectiondagbuilder the run lines of my filecheck tests on llvm ir have llc mtriple linux gnu tmp x ll stop after finalize isel o global isel fast isel to isolate my test to testing isel if i add start before finalize isel i no longer get any codegen from llc it still prints the input ir and yaml for mir but the mir output is gone if i replace start before finalize isel with print before all i see ir dump before instruction selection amdgpu isel huh why is amdgpu isel in there on a hunch testing start before amdgpu isel does what i d have expected from start before finalize isel expected llc mtriple linux gnu tmp x ll start before finalize isel stop after finalize isel o actual llc mtriple linux gnu tmp x ll start before amdgpu isel stop after finalize isel o though adding print before all to expected above shows what looks like amdgpu isel then local dynamic tls cleanup then finalize isel being run i see the same issue if i change the target triple to mtriple linux gnu as well so not specific to i think so not tagging this as related to globalisel since i don t want to test globalisel anyways cc arsenm aeubanks
| 0
|
10,409
| 13,204,505,355
|
IssuesEvent
|
2020-08-14 16:02:20
|
GetTerminus/terminus-oss
|
https://api.github.com/repos/GetTerminus/terminus-oss
|
opened
|
Tokens: plan out phase 2
|
Focus: tokens Goal: Process Improvement Needs: planning Type: chore
|
- outline exactly what phase 2 entails
- which packages are affected
- create issues for each affected package and attach to epic:
|
1.0
|
Tokens: plan out phase 2 - - outline exactly what phase 2 entails
- which packages are affected
- create issues for each affected package and attach to epic:
|
process
|
tokens plan out phase outline exactly what phase entails which packages are affected create issues for each affected package and attach to epic
| 1
|
1,555
| 4,156,188,142
|
IssuesEvent
|
2016-06-16 17:08:21
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Unhandled error [AssertionError] from child_process.js
|
child_process
|
Just got an unhandled error from child_process.js (Node v4.2.0)
AssertionError: UNHANDLED! false == true
File "internal/child_process.js", line 440, in Pipe.channel.onread
File "internal/child_process.js", line 686, in handleMessage
File "events.js", line 172, in ChildProcess.emit
File "events.js", line 92, in emitTwo
File "internal/child_process.js", line 464, in ChildProcess.<anonymous>
Perhaps related to this issue? : https://github.com/nodejs/node/issues/3341
|
1.0
|
Unhandled error [AssertionError] from child_process.js - Just got an unhandled error from child_process.js (Node v4.2.0)
AssertionError: UNHANDLED! false == true
File "internal/child_process.js", line 440, in Pipe.channel.onread
File "internal/child_process.js", line 686, in handleMessage
File "events.js", line 172, in ChildProcess.emit
File "events.js", line 92, in emitTwo
File "internal/child_process.js", line 464, in ChildProcess.<anonymous>
Perhaps related to this issue? : https://github.com/nodejs/node/issues/3341
|
process
|
unhandled error from child process js just got an unhandled error from child process js node assertionerror unhandled false true file internal child process js line in pipe channel onread file internal child process js line in handlemessage file events js line in childprocess emit file events js line in emittwo file internal child process js line in childprocess perhaps related to this issue
| 1
|
5,851
| 8,677,526,045
|
IssuesEvent
|
2018-11-30 17:01:23
|
googlegenomics/gcp-variant-transforms
|
https://api.github.com/repos/googlegenomics/gcp-variant-transforms
|
opened
|
Update GitHub docs to refer to the Cloud docs
|
P2 process
|
Variant Transforms docs are being updated on the cloud site:
https://cloud.google.com/genomics/docs/how-tos/variant-transforms
Once they're done, we should update all of our docs to refer to those pages rather than the github versions.
|
1.0
|
Update GitHub docs to refer to the Cloud docs - Variant Transforms docs are being updated on the cloud site:
https://cloud.google.com/genomics/docs/how-tos/variant-transforms
Once they're done, we should update all of our docs to refer to those pages rather than the github versions.
|
process
|
update github docs to refer to the cloud docs variant transforms docs are being updated on the cloud site once they re done we should update all of our docs to refer to those pages rather than the github versions
| 1
|
19,037
| 25,042,526,143
|
IssuesEvent
|
2022-11-04 22:54:07
|
USGS-WiM/StreamStats
|
https://api.github.com/repos/USGS-WiM/StreamStats
|
opened
|
BP: Add user information and submit
|
Batch Processor
|
Part #1455
- [ ] Add a text box that says "ID Field:". We may want to add a "?" icon next to this element, so that users can hover over the icon and get more information about what this means. Let us know if you'd like to see an example of this.
- [ ] Add a text box that says "Email Address:"
- [ ] Add a file uploader button that says "Upload Shapefile (.zip):"
- [ ] Add a button at the bottom of the form that says "Submit Batch"
- [ ] When the "Submit Batch" button is clicked, validate the form:
- [ ] Ensure a State/Region was selected
- [ ] If the "Compute Flow Statistics" checkbox was checked, ensure at least one Flow Statistic was checked
- [ ] If the "Compute Basin Characteristics" checkbox was checked, ensure at least one Basin Characteristic was checked
- [ ] Trim the text in the "ID Field" and ensure it is not empty
- [ ] Ensure the email address is a valid email address
- [ ] Ensure all shapefile is valid and that are files are present in the .zip
- [ ] If the form is invalid, provide a message to the user about the issue(s). This can be done in a number of ways-- toast message, highlight elements on the form, etc.
- [ ] If the form is valid, notify the user via a green/success toast message that the batch was submitted successfully and that they will be notified by email
|
1.0
|
BP: Add user information and submit - Part #1455
- [ ] Add a text box that says "ID Field:". We may want to add a "?" icon next to this element, so that users can hover over the icon and get more information about what this means. Let us know if you'd like to see an example of this.
- [ ] Add a text box that says "Email Address:"
- [ ] Add a file uploader button that says "Upload Shapefile (.zip):"
- [ ] Add a button at the bottom of the form that says "Submit Batch"
- [ ] When the "Submit Batch" button is clicked, validate the form:
- [ ] Ensure a State/Region was selected
- [ ] If the "Compute Flow Statistics" checkbox was checked, ensure at least one Flow Statistic was checked
- [ ] If the "Compute Basin Characteristics" checkbox was checked, ensure at least one Basin Characteristic was checked
- [ ] Trim the text in the "ID Field" and ensure it is not empty
- [ ] Ensure the email address is a valid email address
- [ ] Ensure all shapefile is valid and that are files are present in the .zip
- [ ] If the form is invalid, provide a message to the user about the issue(s). This can be done in a number of ways-- toast message, highlight elements on the form, etc.
- [ ] If the form is valid, notify the user via a green/success toast message that the batch was submitted successfully and that they will be notified by email
|
process
|
bp add user information and submit part add a text box that says id field we may want to add a icon next to this element so that users can hover over the icon and get more information about what this means let us know if you d like to see an example of this add a text box that says email address add a file uploader button that says upload shapefile zip add a button at the bottom of the form that says submit batch when the submit batch button is clicked validate the form ensure a state region was selected if the compute flow statistics checkbox was checked ensure at least one flow statistic was checked if the compute basin characteristics checkbox was checked ensure at least one basin characteristic was checked trim the text in the id field and ensure it is not empty ensure the email address is a valid email address ensure all shapefile is valid and that are files are present in the zip if the form is invalid provide a message to the user about the issue s this can be done in a number of ways toast message highlight elements on the form etc if the form is valid notify the user via a green success toast message that the batch was submitted successfully and that they will be notified by email
| 1
|
11,182
| 13,957,695,582
|
IssuesEvent
|
2020-10-24 08:11:35
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
HR: new harvesting
|
Geoportal HR - Croatia Harvesting process
|
Dear Angelo,
Can I have new harvesting of national catalouge, please?
Thank you,
Tanja
|
1.0
|
HR: new harvesting - Dear Angelo,
Can I have new harvesting of national catalouge, please?
Thank you,
Tanja
|
process
|
hr new harvesting dear angelo can i have new harvesting of national catalouge please thank you tanja
| 1
|
196,960
| 22,572,000,652
|
IssuesEvent
|
2022-06-28 01:45:14
|
xlizaluizax/datocms-Dating-WIKI---React-Blog
|
https://api.github.com/repos/xlizaluizax/datocms-Dating-WIKI---React-Blog
|
opened
|
CVE-2021-42740 (High) detected in shell-quote-1.7.2.tgz
|
security vulnerability
|
## CVE-2021-42740 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shell-quote-1.7.2.tgz</b></p></summary>
<p>quote and parse shell commands</p>
<p>Library home page: <a href="https://registry.npmjs.org/shell-quote/-/shell-quote-1.7.2.tgz">https://registry.npmjs.org/shell-quote/-/shell-quote-1.7.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/shell-quote/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- react-dev-utils-9.1.0.tgz
- :x: **shell-quote-1.7.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/xlizaluizax/datocms-Dating-WIKI---React-Blog/commit/f28f06b6cbc52fcdd0ce3f5eab40a431a262e08e">f28f06b6cbc52fcdd0ce3f5eab40a431a262e08e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The shell-quote package before 1.7.3 for Node.js allows command injection. An attacker can inject unescaped shell metacharacters through a regex designed to support Windows drive letters. If the output of this package is passed to a real shell as a quoted argument to a command with exec(), an attacker can inject arbitrary commands. This is because the Windows drive letter regex character class is {A-z] instead of the correct {A-Za-z]. Several shell metacharacters exist in the space between capital letter Z and lower case letter a, such as the backtick character.
<p>Publish Date: 2021-10-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42740>CVE-2021-42740</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42740">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42740</a></p>
<p>Release Date: 2021-10-21</p>
<p>Fix Resolution: shell-quote - 1.7.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-42740 (High) detected in shell-quote-1.7.2.tgz - ## CVE-2021-42740 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shell-quote-1.7.2.tgz</b></p></summary>
<p>quote and parse shell commands</p>
<p>Library home page: <a href="https://registry.npmjs.org/shell-quote/-/shell-quote-1.7.2.tgz">https://registry.npmjs.org/shell-quote/-/shell-quote-1.7.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/shell-quote/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- react-dev-utils-9.1.0.tgz
- :x: **shell-quote-1.7.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/xlizaluizax/datocms-Dating-WIKI---React-Blog/commit/f28f06b6cbc52fcdd0ce3f5eab40a431a262e08e">f28f06b6cbc52fcdd0ce3f5eab40a431a262e08e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The shell-quote package before 1.7.3 for Node.js allows command injection. An attacker can inject unescaped shell metacharacters through a regex designed to support Windows drive letters. If the output of this package is passed to a real shell as a quoted argument to a command with exec(), an attacker can inject arbitrary commands. This is because the Windows drive letter regex character class is {A-z] instead of the correct {A-Za-z]. Several shell metacharacters exist in the space between capital letter Z and lower case letter a, such as the backtick character.
<p>Publish Date: 2021-10-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42740>CVE-2021-42740</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42740">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42740</a></p>
<p>Release Date: 2021-10-21</p>
<p>Fix Resolution: shell-quote - 1.7.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in shell quote tgz cve high severity vulnerability vulnerable library shell quote tgz quote and parse shell commands library home page a href path to dependency file package json path to vulnerable library node modules shell quote package json dependency hierarchy react scripts tgz root library react dev utils tgz x shell quote tgz vulnerable library found in head commit a href vulnerability details the shell quote package before for node js allows command injection an attacker can inject unescaped shell metacharacters through a regex designed to support windows drive letters if the output of this package is passed to a real shell as a quoted argument to a command with exec an attacker can inject arbitrary commands this is because the windows drive letter regex character class is a z instead of the correct a za z several shell metacharacters exist in the space between capital letter z and lower case letter a such as the backtick character publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution shell quote step up your open source security game with mend
| 0
|
17,460
| 23,283,085,445
|
IssuesEvent
|
2022-08-05 13:56:35
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Validation fails to detect invalid `SetNull` referential action referencing non-optional fields
|
bug/2-confirmed kind/bug process/candidate team/schema topic: referential actions topic: relation-validation
|
This missed validation triggers a migration error when using MySQL, SQL Server, SQLite, and CockroachDB, but not on Postgres.
Example with MySQL:
```prisma
// schema.prisma
generator client {
provider = "prisma-client-js"
previewFeatures = ["referentialIntegrity"]
}
datasource db {
provider = "mysql"
url = env("DATABASE_URI_MYSQL")
}
model User {
id String @id
profile Profile?
enabled Boolean?
}
model Profile {
id String @id
user User @relation(fields: [userId], references: [id], onUpdate: SetNull, onDelete: SetNull)
// notice that this field should become optional in order to support `SetNull`
userId String @unique
enabled Boolean?
}
```
We can see that the schema is wrongfully considered valid:
```text
❯ prisma validate
Prisma schema loaded from prisma/schema.prisma
The schema at /.../reprod/prisma/schema.prisma is valid 🚀
```
If we attempt a push, we get the following error:
```text
❯ prisma db push --skip-generate
Prisma schema loaded from prisma/schema.prisma
Datasource "db": MySQL database "PRISMA_DB_NAME" at "localhost:3306"
MySQL database PRISMA_DB_NAME created at localhost:3306
Error: Column 'userId' cannot be NOT NULL: needed in a foreign key constraint 'Profile_userId_fkey' SET NULL
0: sql_migration_connector::apply_migration::migration_step
with step=AddForeignKey { foreign_key_id: ForeignKeyId(0) }
at migration-engine/connectors/sql-migration-connector/src/apply_migration.rs:21
1: sql_migration_connector::apply_migration::apply_migration
at migration-engine/connectors/sql-migration-connector/src/apply_migration.rs:10
2: migration_core::state::SchemaPush
at migration-engine/core/src/state.rs:349
```
If we try to create/update the `Profile` model via the Prisma client, we get the following migration error:
```ts
await prisma.$transaction([
prisma.user.create({
data: {
id: '1'.
profile: {
create: { id }
}
}
})
])
```
```text
Column 'userId' cannot be NOT NULL: needed in a foreign key constraint 'Profile_userId_fkey' SET NULL
0: sql_migration_connector::apply_migration::apply_migration
at migration-engine/connectors/sql-migration-connector/src/apply_migration.rs:9
1: migration_core::state::SchemaPush
at migration-engine/core/src/state.rs:349
```
On Postgres, oddly, `prisma db push` it doesn't fail, so migration error is thrown:
```text
❯ prisma db push --skip-generate
Prisma schema loaded from prisma/schema.prisma
Datasource "db": PostgreSQL database "PRISMA_DB_NAME", schema "public" at "localhost:5432"
PostgreSQL database PRISMA_DB_NAME created at localhost:5432
🚀 Your database is now in sync with your Prisma schema. Done in 64ms
```
|
1.0
|
Validation fails to detect invalid `SetNull` referential action referencing non-optional fields - This missed validation triggers a migration error when using MySQL, SQL Server, SQLite, and CockroachDB, but not on Postgres.
Example with MySQL:
```prisma
// schema.prisma
generator client {
provider = "prisma-client-js"
previewFeatures = ["referentialIntegrity"]
}
datasource db {
provider = "mysql"
url = env("DATABASE_URI_MYSQL")
}
model User {
id String @id
profile Profile?
enabled Boolean?
}
model Profile {
id String @id
user User @relation(fields: [userId], references: [id], onUpdate: SetNull, onDelete: SetNull)
// notice that this field should become optional in order to support `SetNull`
userId String @unique
enabled Boolean?
}
```
We can see that the schema is wrongfully considered valid:
```text
❯ prisma validate
Prisma schema loaded from prisma/schema.prisma
The schema at /.../reprod/prisma/schema.prisma is valid 🚀
```
If we attempt a push, we get the following error:
```text
❯ prisma db push --skip-generate
Prisma schema loaded from prisma/schema.prisma
Datasource "db": MySQL database "PRISMA_DB_NAME" at "localhost:3306"
MySQL database PRISMA_DB_NAME created at localhost:3306
Error: Column 'userId' cannot be NOT NULL: needed in a foreign key constraint 'Profile_userId_fkey' SET NULL
0: sql_migration_connector::apply_migration::migration_step
with step=AddForeignKey { foreign_key_id: ForeignKeyId(0) }
at migration-engine/connectors/sql-migration-connector/src/apply_migration.rs:21
1: sql_migration_connector::apply_migration::apply_migration
at migration-engine/connectors/sql-migration-connector/src/apply_migration.rs:10
2: migration_core::state::SchemaPush
at migration-engine/core/src/state.rs:349
```
If we try to create/update the `Profile` model via the Prisma client, we get the following migration error:
```ts
await prisma.$transaction([
prisma.user.create({
data: {
id: '1'.
profile: {
create: { id }
}
}
})
])
```
```text
Column 'userId' cannot be NOT NULL: needed in a foreign key constraint 'Profile_userId_fkey' SET NULL
0: sql_migration_connector::apply_migration::apply_migration
at migration-engine/connectors/sql-migration-connector/src/apply_migration.rs:9
1: migration_core::state::SchemaPush
at migration-engine/core/src/state.rs:349
```
On Postgres, oddly, `prisma db push` it doesn't fail, so migration error is thrown:
```text
❯ prisma db push --skip-generate
Prisma schema loaded from prisma/schema.prisma
Datasource "db": PostgreSQL database "PRISMA_DB_NAME", schema "public" at "localhost:5432"
PostgreSQL database PRISMA_DB_NAME created at localhost:5432
🚀 Your database is now in sync with your Prisma schema. Done in 64ms
```
|
process
|
validation fails to detect invalid setnull referential action referencing non optional fields this missed validation triggers a migration error when using mysql sql server sqlite and cockroachdb but not on postgres example with mysql prisma schema prisma generator client provider prisma client js previewfeatures datasource db provider mysql url env database uri mysql model user id string id profile profile enabled boolean model profile id string id user user relation fields references onupdate setnull ondelete setnull notice that this field should become optional in order to support setnull userid string unique enabled boolean we can see that the schema is wrongfully considered valid text ❯ prisma validate prisma schema loaded from prisma schema prisma the schema at reprod prisma schema prisma is valid 🚀 if we attempt a push we get the following error text ❯ prisma db push skip generate prisma schema loaded from prisma schema prisma datasource db mysql database prisma db name at localhost mysql database prisma db name created at localhost error column userid cannot be not null needed in a foreign key constraint profile userid fkey set null sql migration connector apply migration migration step with step addforeignkey foreign key id foreignkeyid at migration engine connectors sql migration connector src apply migration rs sql migration connector apply migration apply migration at migration engine connectors sql migration connector src apply migration rs migration core state schemapush at migration engine core src state rs if we try to create update the profile model via the prisma client we get the following migration error ts await prisma transaction prisma user create data id profile create id text column userid cannot be not null needed in a foreign key constraint profile userid fkey set null sql migration connector apply migration apply migration at migration engine connectors sql migration connector src apply migration rs migration core state schemapush at migration engine core src state rs on postgres oddly prisma db push it doesn t fail so migration error is thrown text ❯ prisma db push skip generate prisma schema loaded from prisma schema prisma datasource db postgresql database prisma db name schema public at localhost postgresql database prisma db name created at localhost 🚀 your database is now in sync with your prisma schema done in
| 1
|
14,077
| 3,374,111,484
|
IssuesEvent
|
2015-11-24 11:19:55
|
UHPeople/UHPeople
|
https://api.github.com/repos/UHPeople/UHPeople
|
closed
|
photogallery modal js navbar z-index fix does not work on chrome
|
bug coffeescript css jquery test needed
|
Makes bug where top of modal is not visible
|
1.0
|
photogallery modal js navbar z-index fix does not work on chrome - Makes bug where top of modal is not visible
|
non_process
|
photogallery modal js navbar z index fix does not work on chrome makes bug where top of modal is not visible
| 0
|
423,855
| 12,303,063,354
|
IssuesEvent
|
2020-05-11 18:02:14
|
adumortier/dev.to
|
https://api.github.com/repos/adumortier/dev.to
|
closed
|
Wireframe for Reader Stats
|
high priority
|
Add `section` or `div` to _sidebar.html.erb to hold the reader stats information.
|
1.0
|
Wireframe for Reader Stats - Add `section` or `div` to _sidebar.html.erb to hold the reader stats information.
|
non_process
|
wireframe for reader stats add section or div to sidebar html erb to hold the reader stats information
| 0
|
21,820
| 30,316,678,438
|
IssuesEvent
|
2023-07-10 16:01:55
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
New Term - "caste" for eusocial organisms
|
Term - add Class - Occurrence normative Process - complete
|
## New term
* Submitter: @wcornwell
* Efficacy Justification (why is this term necessary?): Eusocial organisms not well captured by existing terms: lifestage and sex
* Demand Justification (name at least two organizations that independently need this term): All museums with bee, ant, termite, or mole rat collections
* Stability Justification (what concerns are there that this might affect existing implementations?): Should not affect exiting implementations
* Implications for dwciri: namespace (does this change affect a dwciri term version)?:
Proposed attributes of the new term:
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): caste
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): occurrence
* Definition of the term (normative): Categorisation of individuals for eusocial species (including some mammals and arthropods)
* Usage comments (recommendations regarding content, etc., not normative): values could include worker, soldier, and reproductive
* Examples (not normative):
* Refines (identifier of the broader term this term refines; normative): this refines "sex" and "lifestage" for a particular set of organisms
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative):
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative):
|
1.0
|
New Term - "caste" for eusocial organisms - ## New term
* Submitter: @wcornwell
* Efficacy Justification (why is this term necessary?): Eusocial organisms not well captured by existing terms: lifestage and sex
* Demand Justification (name at least two organizations that independently need this term): All museums with bee, ant, termite, or mole rat collections
* Stability Justification (what concerns are there that this might affect existing implementations?): Should not affect exiting implementations
* Implications for dwciri: namespace (does this change affect a dwciri term version)?:
Proposed attributes of the new term:
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): caste
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): occurrence
* Definition of the term (normative): Categorisation of individuals for eusocial species (including some mammals and arthropods)
* Usage comments (recommendations regarding content, etc., not normative): values could include worker, soldier, and reproductive
* Examples (not normative):
* Refines (identifier of the broader term this term refines; normative): this refines "sex" and "lifestage" for a particular set of organisms
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative):
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative):
|
process
|
new term caste for eusocial organisms new term submitter wcornwell efficacy justification why is this term necessary eusocial organisms not well captured by existing terms lifestage and sex demand justification name at least two organizations that independently need this term all museums with bee ant termite or mole rat collections stability justification what concerns are there that this might affect existing implementations should not affect exiting implementations implications for dwciri namespace does this change affect a dwciri term version proposed attributes of the new term term name in lowercamelcase for properties uppercamelcase for classes caste organized in class e g occurrence event location taxon occurrence definition of the term normative categorisation of individuals for eusocial species including some mammals and arthropods usage comments recommendations regarding content etc not normative values could include worker soldier and reproductive examples not normative refines identifier of the broader term this term refines normative this refines sex and lifestage for a particular set of organisms replaces identifier of the existing term that would be deprecated and replaced by this term normative abcd xpath of the equivalent term in abcd or efg not normative
| 1
|
9,042
| 12,130,107,993
|
IssuesEvent
|
2020-04-23 00:30:40
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
remove gcp-devrel-py-tools from appengine/standard/sendgrid/requirements-test.txt
|
priority: p2 remove-gcp-devrel-py-tools type: process
|
remove gcp-devrel-py-tools from appengine/standard/sendgrid/requirements-test.txt
|
1.0
|
remove gcp-devrel-py-tools from appengine/standard/sendgrid/requirements-test.txt - remove gcp-devrel-py-tools from appengine/standard/sendgrid/requirements-test.txt
|
process
|
remove gcp devrel py tools from appengine standard sendgrid requirements test txt remove gcp devrel py tools from appengine standard sendgrid requirements test txt
| 1
|
268,926
| 8,415,819,029
|
IssuesEvent
|
2018-10-13 18:31:04
|
robot-lab/judyst-text-analysis
|
https://api.github.com/repos/robot-lab/judyst-text-analysis
|
opened
|
Сборка атрибутов решений.
|
idea priority 1
|
# Idea proposal
## В чём заключается идея
Судебное решение - сложная структура данных, поэтому необходимо научится выделять у данной структуры определенный набор атрибутов (например суд, регион, дата).
## Возможное решение
Выделить основные атрибуты которые нам необходимы на основе регулярных выражений и добавить эту информацию в таблицу БД.
## Дополнительный контекст или ссылки на связанные с данной задачей issues
Связано с исследовательской задачей: https://github.com/robot-lab/judyst-research/issues/2.
|
1.0
|
Сборка атрибутов решений. - # Idea proposal
## В чём заключается идея
Судебное решение - сложная структура данных, поэтому необходимо научится выделять у данной структуры определенный набор атрибутов (например суд, регион, дата).
## Возможное решение
Выделить основные атрибуты которые нам необходимы на основе регулярных выражений и добавить эту информацию в таблицу БД.
## Дополнительный контекст или ссылки на связанные с данной задачей issues
Связано с исследовательской задачей: https://github.com/robot-lab/judyst-research/issues/2.
|
non_process
|
сборка атрибутов решений idea proposal в чём заключается идея судебное решение сложная структура данных поэтому необходимо научится выделять у данной структуры определенный набор атрибутов например суд регион дата возможное решение выделить основные атрибуты которые нам необходимы на основе регулярных выражений и добавить эту информацию в таблицу бд дополнительный контекст или ссылки на связанные с данной задачей issues связано с исследовательской задачей
| 0
|
194,435
| 6,894,855,754
|
IssuesEvent
|
2017-11-23 11:36:00
|
spring-projects/spring-boot
|
https://api.github.com/repos/spring-projects/spring-boot
|
closed
|
JOOQ DSLContext remains non-transactional after auto-configuration
|
priority: normal type: bug
|
Noticed in Spring Boot **2.0.0.M6**.
With no additional configuration, the auto-wired JOOQ `DSLContext` still has `transactional` flag set to `false` and no transaction provider linked in after the application start.
It will become transactional if an explicit `PlatformTransactionManager` bean is provided:
```
@Configuration
@EnableTransactionManagement
public class AppConfig {
@Autowired
DataSource dataSource;
@Bean
public PlatformTransactionManager transactionManager() {
return new DataSourceTransactionManager(dataSource);
}
}
```
However, I have noticed that even without the explicit manager bean configuration above, an auto-wired `PlatformTransactionManager` within the same service will contain a reference. Apparently another auto-configuration creates it after all. Since transaction provider setup in the `JOOQAutoConfiguration` is declared as dependent on the `PlatformTransactionManager` bean existence, it seems like JOOQ auto-configuration runs too early to pick the manager's reference.
|
1.0
|
JOOQ DSLContext remains non-transactional after auto-configuration - Noticed in Spring Boot **2.0.0.M6**.
With no additional configuration, the auto-wired JOOQ `DSLContext` still has `transactional` flag set to `false` and no transaction provider linked in after the application start.
It will become transactional if an explicit `PlatformTransactionManager` bean is provided:
```
@Configuration
@EnableTransactionManagement
public class AppConfig {
@Autowired
DataSource dataSource;
@Bean
public PlatformTransactionManager transactionManager() {
return new DataSourceTransactionManager(dataSource);
}
}
```
However, I have noticed that even without the explicit manager bean configuration above, an auto-wired `PlatformTransactionManager` within the same service will contain a reference. Apparently another auto-configuration creates it after all. Since transaction provider setup in the `JOOQAutoConfiguration` is declared as dependent on the `PlatformTransactionManager` bean existence, it seems like JOOQ auto-configuration runs too early to pick the manager's reference.
|
non_process
|
jooq dslcontext remains non transactional after auto configuration noticed in spring boot with no additional configuration the auto wired jooq dslcontext still has transactional flag set to false and no transaction provider linked in after the application start it will become transactional if an explicit platformtransactionmanager bean is provided configuration enabletransactionmanagement public class appconfig autowired datasource datasource bean public platformtransactionmanager transactionmanager return new datasourcetransactionmanager datasource however i have noticed that even without the explicit manager bean configuration above an auto wired platformtransactionmanager within the same service will contain a reference apparently another auto configuration creates it after all since transaction provider setup in the jooqautoconfiguration is declared as dependent on the platformtransactionmanager bean existence it seems like jooq auto configuration runs too early to pick the manager s reference
| 0
|
9,804
| 12,815,256,629
|
IssuesEvent
|
2020-07-05 00:56:13
|
Okurdostu/aspnetcore-okurdostu.com
|
https://api.github.com/repos/Okurdostu/aspnetcore-okurdostu.com
|
closed
|
Hierarchical comments visual bug
|
bug enhancement in the process
|
I'm trying to do a hierarchical comment and reply system. I designed a database and visual for it but visual design isn't working right.
The database table design has two bases. When RelatedCommentId is null is the main comment and When it has data in RelatedCommentId column it's a reply.
one more thing, comment or reply can take endless reply. when a comment takes a lot of replies it doesn't seem right with my visual design.
Please, look at the images, you will understand the problem.
image from desktop

image from large mobile

Do you have any advice or solution?
I used the bootstrap grid system to do that and I couldn't solve this problem. I think it needs a special design or logical something and someone who has experience in design and a nice person can do that 😹 **thanks a lot in advance**.
|
1.0
|
Hierarchical comments visual bug - I'm trying to do a hierarchical comment and reply system. I designed a database and visual for it but visual design isn't working right.
The database table design has two bases. When RelatedCommentId is null is the main comment and When it has data in RelatedCommentId column it's a reply.
one more thing, comment or reply can take endless reply. when a comment takes a lot of replies it doesn't seem right with my visual design.
Please, look at the images, you will understand the problem.
image from desktop

image from large mobile

Do you have any advice or solution?
I used the bootstrap grid system to do that and I couldn't solve this problem. I think it needs a special design or logical something and someone who has experience in design and a nice person can do that 😹 **thanks a lot in advance**.
|
process
|
hierarchical comments visual bug i m trying to do a hierarchical comment and reply system i designed a database and visual for it but visual design isn t working right the database table design has two bases when relatedcommentid is null is the main comment and when it has data in relatedcommentid column it s a reply one more thing comment or reply can take endless reply when a comment takes a lot of replies it doesn t seem right with my visual design please look at the images you will understand the problem image from desktop image from large mobile do you have any advice or solution i used the bootstrap grid system to do that and i couldn t solve this problem i think it needs a special design or logical something and someone who has experience in design and a nice person can do that 😹 thanks a lot in advance
| 1
|
55,628
| 6,911,170,243
|
IssuesEvent
|
2017-11-28 07:00:20
|
AlexRoijals/ProjectWormhole
|
https://api.github.com/repos/AlexRoijals/ProjectWormhole
|
opened
|
Create a provisional playable terrain of the isle in 3D
|
3D ART CONCEPT ART GAME DESIGN
|
For block testing and in situ level design in Unity.
|
1.0
|
Create a provisional playable terrain of the isle in 3D - For block testing and in situ level design in Unity.
|
non_process
|
create a provisional playable terrain of the isle in for block testing and in situ level design in unity
| 0
|
2,955
| 5,955,170,490
|
IssuesEvent
|
2017-05-28 02:19:49
|
StoneCypher/jssm
|
https://api.github.com/repos/StoneCypher/jssm
|
closed
|
Apparently badges factor into npm quality, so, ... get badges
|
process
|
from shields.io, maybe others (coveralls mebbe?)
|
1.0
|
Apparently badges factor into npm quality, so, ... get badges - from shields.io, maybe others (coveralls mebbe?)
|
process
|
apparently badges factor into npm quality so get badges from shields io maybe others coveralls mebbe
| 1
|
7,386
| 10,515,362,557
|
IssuesEvent
|
2019-09-28 09:11:32
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
closed
|
Make multi-statements configurable
|
ADMIN AUTHENTICATION CONNECTION POOL QUERY PROCESSOR
|
ProxySQL supports multi-statements since 1.2.0f: see #546 and #547.
As a security enhancement, an interesting feature is to disable multi-statements, no matter what the client's connection wants.
# What
* [ ] add new global variable
* [ ] variable needs to be evaluated at connection phase, even if client may ignore it
* [ ] code that handles `COM_SET_OPTION` should either succeed or return an error (see #1122)
* [ ] tokenizer should return an error if a query contains semi-column and multi-statements is disabled
No milestone
|
1.0
|
Make multi-statements configurable - ProxySQL supports multi-statements since 1.2.0f: see #546 and #547.
As a security enhancement, an interesting feature is to disable multi-statements, no matter what the client's connection wants.
# What
* [ ] add new global variable
* [ ] variable needs to be evaluated at connection phase, even if client may ignore it
* [ ] code that handles `COM_SET_OPTION` should either succeed or return an error (see #1122)
* [ ] tokenizer should return an error if a query contains semi-column and multi-statements is disabled
No milestone
|
process
|
make multi statements configurable proxysql supports multi statements since see and as a security enhancement an interesting feature is to disable multi statements no matter what the client s connection wants what add new global variable variable needs to be evaluated at connection phase even if client may ignore it code that handles com set option should either succeed or return an error see tokenizer should return an error if a query contains semi column and multi statements is disabled no milestone
| 1
|
8,684
| 11,813,306,184
|
IssuesEvent
|
2020-03-19 22:03:47
|
googleapis/google-cloud-cpp-common
|
https://api.github.com/repos/googleapis/google-cloud-cpp-common
|
opened
|
Document what we consider an "API breakage"
|
type: process
|
I think we should create an ADR documenting what is an API breakage, or at least the things we do not consider part of the API. I am thinking of:
- We make no attempts to preserve the A**B**I. We change object sizes, add fields, change private members functions and fields without notice. We also may add default parameters to a function without notice.
- For the purposes of this document "API" means the "library API", the guarantees about the REST and/or gRPC APIs offered by GCP are discussed elsewhere.
- We try to preserve the API, but we will break the API if we think it is important, we will document -- to the best of our abilities -- any API breakage.
- Header files are part of the API, but you should not rely on indirect `#includes`.
- We consider type aliases and inline functions part of the API
- We do **not** consider types, functions or objects in the `internal` (or `foo_internal`) namespace to be part of the API, we will change those without notice.
- We do **not** consider types, functions or objects used in our tests part of the API, this includes anything in the `testing` or `foo_testing` namespace.
- We do **not** consider header files in the `internal/` or `testing/` or `testing_utils/` directories to be part of the API.
- This list is not exhaustive, as we discover other things that should not be considered part of the API we will update this list, but still will document any changes.
|
1.0
|
Document what we consider an "API breakage" - I think we should create an ADR documenting what is an API breakage, or at least the things we do not consider part of the API. I am thinking of:
- We make no attempts to preserve the A**B**I. We change object sizes, add fields, change private members functions and fields without notice. We also may add default parameters to a function without notice.
- For the purposes of this document "API" means the "library API", the guarantees about the REST and/or gRPC APIs offered by GCP are discussed elsewhere.
- We try to preserve the API, but we will break the API if we think it is important, we will document -- to the best of our abilities -- any API breakage.
- Header files are part of the API, but you should not rely on indirect `#includes`.
- We consider type aliases and inline functions part of the API
- We do **not** consider types, functions or objects in the `internal` (or `foo_internal`) namespace to be part of the API, we will change those without notice.
- We do **not** consider types, functions or objects used in our tests part of the API, this includes anything in the `testing` or `foo_testing` namespace.
- We do **not** consider header files in the `internal/` or `testing/` or `testing_utils/` directories to be part of the API.
- This list is not exhaustive, as we discover other things that should not be considered part of the API we will update this list, but still will document any changes.
|
process
|
document what we consider an api breakage i think we should create an adr documenting what is an api breakage or at least the things we do not consider part of the api i am thinking of we make no attempts to preserve the a b i we change object sizes add fields change private members functions and fields without notice we also may add default parameters to a function without notice for the purposes of this document api means the library api the guarantees about the rest and or grpc apis offered by gcp are discussed elsewhere we try to preserve the api but we will break the api if we think it is important we will document to the best of our abilities any api breakage header files are part of the api but you should not rely on indirect includes we consider type aliases and inline functions part of the api we do not consider types functions or objects in the internal or foo internal namespace to be part of the api we will change those without notice we do not consider types functions or objects used in our tests part of the api this includes anything in the testing or foo testing namespace we do not consider header files in the internal or testing or testing utils directories to be part of the api this list is not exhaustive as we discover other things that should not be considered part of the api we will update this list but still will document any changes
| 1
|
88,119
| 15,800,733,445
|
IssuesEvent
|
2021-04-03 01:02:37
|
ioana-nicolae/keycloak
|
https://api.github.com/repos/ioana-nicolae/keycloak
|
opened
|
CVE-2021-28165 (High) detected in multiple libraries
|
security vulnerability
|
## CVE-2021-28165 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jetty-io-9.4.5.v20170502.jar</b>, <b>jetty-io-9.4.29.v20200521.jar</b>, <b>jetty-io-9.4.6.v20170531.jar</b>, <b>jetty-io-9.4.3.v20170317.jar</b></p></summary>
<p>
<details><summary><b>jetty-io-9.4.5.v20170502.jar</b></p></summary>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.5.v20170502/jetty-io-9.4.5.v20170502.jar</p>
<p>
Dependency Hierarchy:
- graphene-webdriver-2.3.2.pom (Root Library)
- arquillian-drone-webdriver-depchain-2.5.2.pom
- htmlunit-driver-2.27.jar
- htmlunit-2.27.jar
- websocket-client-9.4.5.v20170502.jar
- :x: **jetty-io-9.4.5.v20170502.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-io-9.4.29.v20200521.jar</b></p></summary>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: keycloak/adapters/saml/jetty/jetty9.4/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.29.v20200521/jetty-io-9.4.29.v20200521.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.29.v20200521/jetty-io-9.4.29.v20200521.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.29.v20200521/jetty-io-9.4.29.v20200521.jar</p>
<p>
Dependency Hierarchy:
- jetty-server-9.4.29.v20200521.jar (Root Library)
- :x: **jetty-io-9.4.29.v20200521.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-io-9.4.6.v20170531.jar</b></p></summary>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: keycloak/testsuite/integration-arquillian/test-apps/fuse/camel/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.6.v20170531/jetty-io-9.4.6.v20170531.jar</p>
<p>
Dependency Hierarchy:
- camel-jetty9-2.21.2.jar (Root Library)
- jetty-server-9.4.6.v20170531.jar
- :x: **jetty-io-9.4.6.v20170531.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-io-9.4.3.v20170317.jar</b></p></summary>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: keycloak/testsuite/model/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.3.v20170317/jetty-io-9.4.3.v20170317.jar</p>
<p>
Dependency Hierarchy:
- integration-arquillian-tests-base-13.0.0-SNAPSHOT.jar (Root Library)
- graphene-webdriver-2.3.2.pom
- arquillian-drone-webdriver-depchain-2.4.3.pom
- htmlunit-driver-2.26.jar
- htmlunit-2.26.jar
- websocket-client-9.4.3.v20170317.jar
- :x: **jetty-io-9.4.3.v20170317.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Eclipse Jetty 7.2.2 to 9.4.38, 10.0.0.alpha0 to 10.0.1, and 11.0.0.alpha0 to 11.0.1, CPU usage can reach 100% upon receiving a large invalid TLS frame.
<p>Publish Date: 2021-04-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28165>CVE-2021-28165</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-26vr-8j45-3r4w">https://github.com/eclipse/jetty.project/security/advisories/GHSA-26vr-8j45-3r4w</a></p>
<p>Release Date: 2021-04-01</p>
<p>Fix Resolution: org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-io","packageVersion":"9.4.5.v20170502","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.jboss.arquillian.graphene:graphene-webdriver:2.3.2;org.jboss.arquillian.extension:arquillian-drone-webdriver-depchain:2.5.2;org.seleniumhq.selenium:htmlunit-driver:2.27;net.sourceforge.htmlunit:htmlunit:2.27;org.eclipse.jetty.websocket:websocket-client:9.4.5.v20170502;org.eclipse.jetty:jetty-io:9.4.5.v20170502","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2"},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-io","packageVersion":"9.4.29.v20200521","packageFilePaths":["/adapters/saml/jetty/jetty9.4/pom.xml","/adapters/oidc/jetty/jetty9.4/pom.xml","/testsuite/integration-arquillian/test-apps/fuse/cxf-jaxrs/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.eclipse.jetty:jetty-server:9.4.29.v20200521;org.eclipse.jetty:jetty-io:9.4.29.v20200521","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2"},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-io","packageVersion":"9.4.6.v20170531","packageFilePaths":["/testsuite/integration-arquillian/test-apps/fuse/camel/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.camel:camel-jetty9:2.21.2;org.eclipse.jetty:jetty-server:9.4.6.v20170531;org.eclipse.jetty:jetty-io:9.4.6.v20170531","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2"},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-io","packageVersion":"9.4.3.v20170317","packageFilePaths":["/testsuite/model/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.keycloak.testsuite:integration-arquillian-tests-base:13.0.0-SNAPSHOT;org.jboss.arquillian.graphene:graphene-webdriver:2.3.2;org.jboss.arquillian.extension:arquillian-drone-webdriver-depchain:2.4.3;org.seleniumhq.selenium:htmlunit-driver:2.26;net.sourceforge.htmlunit:htmlunit:2.26;org.eclipse.jetty.websocket:websocket-client:9.4.3.v20170317;org.eclipse.jetty:jetty-io:9.4.3.v20170317","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-28165","vulnerabilityDetails":"In Eclipse Jetty 7.2.2 to 9.4.38, 10.0.0.alpha0 to 10.0.1, and 11.0.0.alpha0 to 11.0.1, CPU usage can reach 100% upon receiving a large invalid TLS frame.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28165","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-28165 (High) detected in multiple libraries - ## CVE-2021-28165 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jetty-io-9.4.5.v20170502.jar</b>, <b>jetty-io-9.4.29.v20200521.jar</b>, <b>jetty-io-9.4.6.v20170531.jar</b>, <b>jetty-io-9.4.3.v20170317.jar</b></p></summary>
<p>
<details><summary><b>jetty-io-9.4.5.v20170502.jar</b></p></summary>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.5.v20170502/jetty-io-9.4.5.v20170502.jar</p>
<p>
Dependency Hierarchy:
- graphene-webdriver-2.3.2.pom (Root Library)
- arquillian-drone-webdriver-depchain-2.5.2.pom
- htmlunit-driver-2.27.jar
- htmlunit-2.27.jar
- websocket-client-9.4.5.v20170502.jar
- :x: **jetty-io-9.4.5.v20170502.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-io-9.4.29.v20200521.jar</b></p></summary>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: keycloak/adapters/saml/jetty/jetty9.4/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.29.v20200521/jetty-io-9.4.29.v20200521.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.29.v20200521/jetty-io-9.4.29.v20200521.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.29.v20200521/jetty-io-9.4.29.v20200521.jar</p>
<p>
Dependency Hierarchy:
- jetty-server-9.4.29.v20200521.jar (Root Library)
- :x: **jetty-io-9.4.29.v20200521.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-io-9.4.6.v20170531.jar</b></p></summary>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: keycloak/testsuite/integration-arquillian/test-apps/fuse/camel/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.6.v20170531/jetty-io-9.4.6.v20170531.jar</p>
<p>
Dependency Hierarchy:
- camel-jetty9-2.21.2.jar (Root Library)
- jetty-server-9.4.6.v20170531.jar
- :x: **jetty-io-9.4.6.v20170531.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-io-9.4.3.v20170317.jar</b></p></summary>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: keycloak/testsuite/model/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-io/9.4.3.v20170317/jetty-io-9.4.3.v20170317.jar</p>
<p>
Dependency Hierarchy:
- integration-arquillian-tests-base-13.0.0-SNAPSHOT.jar (Root Library)
- graphene-webdriver-2.3.2.pom
- arquillian-drone-webdriver-depchain-2.4.3.pom
- htmlunit-driver-2.26.jar
- htmlunit-2.26.jar
- websocket-client-9.4.3.v20170317.jar
- :x: **jetty-io-9.4.3.v20170317.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Eclipse Jetty 7.2.2 to 9.4.38, 10.0.0.alpha0 to 10.0.1, and 11.0.0.alpha0 to 11.0.1, CPU usage can reach 100% upon receiving a large invalid TLS frame.
<p>Publish Date: 2021-04-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28165>CVE-2021-28165</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-26vr-8j45-3r4w">https://github.com/eclipse/jetty.project/security/advisories/GHSA-26vr-8j45-3r4w</a></p>
<p>Release Date: 2021-04-01</p>
<p>Fix Resolution: org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-io","packageVersion":"9.4.5.v20170502","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.jboss.arquillian.graphene:graphene-webdriver:2.3.2;org.jboss.arquillian.extension:arquillian-drone-webdriver-depchain:2.5.2;org.seleniumhq.selenium:htmlunit-driver:2.27;net.sourceforge.htmlunit:htmlunit:2.27;org.eclipse.jetty.websocket:websocket-client:9.4.5.v20170502;org.eclipse.jetty:jetty-io:9.4.5.v20170502","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2"},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-io","packageVersion":"9.4.29.v20200521","packageFilePaths":["/adapters/saml/jetty/jetty9.4/pom.xml","/adapters/oidc/jetty/jetty9.4/pom.xml","/testsuite/integration-arquillian/test-apps/fuse/cxf-jaxrs/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.eclipse.jetty:jetty-server:9.4.29.v20200521;org.eclipse.jetty:jetty-io:9.4.29.v20200521","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2"},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-io","packageVersion":"9.4.6.v20170531","packageFilePaths":["/testsuite/integration-arquillian/test-apps/fuse/camel/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.camel:camel-jetty9:2.21.2;org.eclipse.jetty:jetty-server:9.4.6.v20170531;org.eclipse.jetty:jetty-io:9.4.6.v20170531","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2"},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-io","packageVersion":"9.4.3.v20170317","packageFilePaths":["/testsuite/model/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.keycloak.testsuite:integration-arquillian-tests-base:13.0.0-SNAPSHOT;org.jboss.arquillian.graphene:graphene-webdriver:2.3.2;org.jboss.arquillian.extension:arquillian-drone-webdriver-depchain:2.4.3;org.seleniumhq.selenium:htmlunit-driver:2.26;net.sourceforge.htmlunit:htmlunit:2.26;org.eclipse.jetty.websocket:websocket-client:9.4.3.v20170317;org.eclipse.jetty:jetty-io:9.4.3.v20170317","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-io:9.4.39, org.eclipse.jetty:jetty-io:10.0.2, org.eclipse.jetty:jetty-io:11.0.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-28165","vulnerabilityDetails":"In Eclipse Jetty 7.2.2 to 9.4.38, 10.0.0.alpha0 to 10.0.1, and 11.0.0.alpha0 to 11.0.1, CPU usage can reach 100% upon receiving a large invalid TLS frame.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28165","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jetty io jar jetty io jar jetty io jar jetty io jar jetty io jar the eclipse jetty project library home page a href path to vulnerable library home wss scanner repository org eclipse jetty jetty io jetty io jar dependency hierarchy graphene webdriver pom root library arquillian drone webdriver depchain pom htmlunit driver jar htmlunit jar websocket client jar x jetty io jar vulnerable library jetty io jar the eclipse jetty project library home page a href path to dependency file keycloak adapters saml jetty pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty io jetty io jar home wss scanner repository org eclipse jetty jetty io jetty io jar home wss scanner repository org eclipse jetty jetty io jetty io jar dependency hierarchy jetty server jar root library x jetty io jar vulnerable library jetty io jar the eclipse jetty project library home page a href path to dependency file keycloak testsuite integration arquillian test apps fuse camel pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty io jetty io jar dependency hierarchy camel jar root library jetty server jar x jetty io jar vulnerable library jetty io jar the eclipse jetty project library home page a href path to dependency file keycloak testsuite model pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty io jetty io jar dependency hierarchy integration arquillian tests base snapshot jar root library graphene webdriver pom arquillian drone webdriver depchain pom htmlunit driver jar htmlunit jar websocket client jar x jetty io jar vulnerable library found in base branch master vulnerability details in eclipse jetty to to and to cpu usage can reach upon receiving a large invalid tls frame publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty io org eclipse jetty jetty io org eclipse jetty jetty io isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org jboss arquillian graphene graphene webdriver org jboss arquillian extension arquillian drone webdriver depchain org seleniumhq selenium htmlunit driver net sourceforge htmlunit htmlunit org eclipse jetty websocket websocket client org eclipse jetty jetty io isminimumfixversionavailable true minimumfixversion org eclipse jetty jetty io org eclipse jetty jetty io org eclipse jetty jetty io packagetype java groupid org eclipse jetty packagename jetty io packageversion packagefilepaths istransitivedependency true dependencytree org eclipse jetty jetty server org eclipse jetty jetty io isminimumfixversionavailable true minimumfixversion org eclipse jetty jetty io org eclipse jetty jetty io org eclipse jetty jetty io packagetype java groupid org eclipse jetty packagename jetty io packageversion packagefilepaths istransitivedependency true dependencytree org apache camel camel org eclipse jetty jetty server org eclipse jetty jetty io isminimumfixversionavailable true minimumfixversion org eclipse jetty jetty io org eclipse jetty jetty io org eclipse jetty jetty io packagetype java groupid org eclipse jetty packagename jetty io packageversion packagefilepaths istransitivedependency true dependencytree org keycloak testsuite integration arquillian tests base snapshot org jboss arquillian graphene graphene webdriver org jboss arquillian extension arquillian drone webdriver depchain org seleniumhq selenium htmlunit driver net sourceforge htmlunit htmlunit org eclipse jetty websocket websocket client org eclipse jetty jetty io isminimumfixversionavailable true minimumfixversion org eclipse jetty jetty io org eclipse jetty jetty io org eclipse jetty jetty io basebranches vulnerabilityidentifier cve vulnerabilitydetails in eclipse jetty to to and to cpu usage can reach upon receiving a large invalid tls frame vulnerabilityurl
| 0
|
288,137
| 24,882,768,767
|
IssuesEvent
|
2022-10-28 03:47:11
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Orçamento - Execução - Presidente Olegário
|
generalization test development template - Memory (66) tag - Orçamento subtag - Execução
|
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Execução para o Município de Presidente Olegário.
|
1.0
|
Teste de generalizacao para a tag Orçamento - Execução - Presidente Olegário - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Execução para o Município de Presidente Olegário.
|
non_process
|
teste de generalizacao para a tag orçamento execução presidente olegário dod realizar o teste de generalização do validador da tag orçamento execução para o município de presidente olegário
| 0
|
242,209
| 20,205,678,534
|
IssuesEvent
|
2022-02-11 20:03:20
|
ValveSoftware/steam-for-linux
|
https://api.github.com/repos/ValveSoftware/steam-for-linux
|
closed
|
Login window flickering and no text is displayed
|
Intel drivers Steam client Need Retest Distro Family: openSUSE
|
#### Your system information
* Steam client version (build number or date): Updated on 14 September 2019
* Distribution (e.g. Ubuntu): openSUSE Tumbleweed
* Opted into Steam client beta?: [Yes/No] Apparently, yes
* Have you checked for system updates?: [Yes/No] Yes
#### Please describe your issue in as much detail as possible:
Steam login window should be displayed okay.

#### Steps for reproducing this issue:
1. Run steam from the command line
[SteamConsoleOutput.txt](https://github.com/ValveSoftware/steam-for-linux/files/3612407/SteamConsoleOutput.txt)
|
1.0
|
Login window flickering and no text is displayed - #### Your system information
* Steam client version (build number or date): Updated on 14 September 2019
* Distribution (e.g. Ubuntu): openSUSE Tumbleweed
* Opted into Steam client beta?: [Yes/No] Apparently, yes
* Have you checked for system updates?: [Yes/No] Yes
#### Please describe your issue in as much detail as possible:
Steam login window should be displayed okay.

#### Steps for reproducing this issue:
1. Run steam from the command line
[SteamConsoleOutput.txt](https://github.com/ValveSoftware/steam-for-linux/files/3612407/SteamConsoleOutput.txt)
|
non_process
|
login window flickering and no text is displayed your system information steam client version build number or date updated on september distribution e g ubuntu opensuse tumbleweed opted into steam client beta apparently yes have you checked for system updates yes please describe your issue in as much detail as possible steam login window should be displayed okay steps for reproducing this issue run steam from the command line
| 0
|
255,627
| 27,485,027,125
|
IssuesEvent
|
2023-03-04 01:49:01
|
panasalap/linux-4.1.15
|
https://api.github.com/repos/panasalap/linux-4.1.15
|
closed
|
CVE-2017-5550 (Medium) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246 - autoclosed
|
security vulnerability
|
## CVE-2017-5550 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b></p></summary>
<p>
<p>Linux kernel stable tree mirror</p>
<p>Library home page: <a href=https://github.com/gregkh/linux.git>https://github.com/gregkh/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/aae4c2fa46027fd4c477372871df090c6b94f3f1">aae4c2fa46027fd4c477372871df090c6b94f3f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Off-by-one error in the pipe_advance function in lib/iov_iter.c in the Linux kernel before 4.9.5 allows local users to obtain sensitive information from uninitialized heap-memory locations in opportunistic circumstances by reading from a pipe after an incorrect buffer-release decision.
<p>Publish Date: 2017-02-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-5550>CVE-2017-5550</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-5550">https://nvd.nist.gov/vuln/detail/CVE-2017-5550</a></p>
<p>Release Date: 2017-02-06</p>
<p>Fix Resolution: 4.9.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-5550 (Medium) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246 - autoclosed - ## CVE-2017-5550 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b></p></summary>
<p>
<p>Linux kernel stable tree mirror</p>
<p>Library home page: <a href=https://github.com/gregkh/linux.git>https://github.com/gregkh/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/aae4c2fa46027fd4c477372871df090c6b94f3f1">aae4c2fa46027fd4c477372871df090c6b94f3f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Off-by-one error in the pipe_advance function in lib/iov_iter.c in the Linux kernel before 4.9.5 allows local users to obtain sensitive information from uninitialized heap-memory locations in opportunistic circumstances by reading from a pipe after an incorrect buffer-release decision.
<p>Publish Date: 2017-02-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-5550>CVE-2017-5550</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-5550">https://nvd.nist.gov/vuln/detail/CVE-2017-5550</a></p>
<p>Release Date: 2017-02-06</p>
<p>Fix Resolution: 4.9.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in autoclosed cve medium severity vulnerability vulnerable library linux kernel stable tree mirror library home page a href found in head commit a href found in base branch master vulnerable source files lib iov iter c vulnerability details off by one error in the pipe advance function in lib iov iter c in the linux kernel before allows local users to obtain sensitive information from uninitialized heap memory locations in opportunistic circumstances by reading from a pipe after an incorrect buffer release decision publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
22,247
| 30,801,588,707
|
IssuesEvent
|
2023-08-01 02:11:27
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
roblox-pyc 2.26.113 has 3 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "2.26.113",
"result": {
"issues": 3,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-2.26.113/robloxpyc/installationmanager.py:19",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-2.26.113/robloxpyc/installationmanager.py:26",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-2.26.113/robloxpyc/installationmanager.py:79",
"code": " subprocess.call([\"npm\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpg6jdrhun/roblox-pyc"
}
}```
|
1.0
|
roblox-pyc 2.26.113 has 3 GuardDog issues - https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "2.26.113",
"result": {
"issues": 3,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-2.26.113/robloxpyc/installationmanager.py:19",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-2.26.113/robloxpyc/installationmanager.py:26",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-2.26.113/robloxpyc/installationmanager.py:79",
"code": " subprocess.call([\"npm\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpg6jdrhun/roblox-pyc"
}
}```
|
process
|
roblox pyc has guarddog issues dependency roblox pyc version result issues errors results silent process execution location roblox pyc robloxpyc installationmanager py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc installationmanager py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc robloxpyc installationmanager py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp roblox pyc
| 1
|
19,117
| 25,170,049,230
|
IssuesEvent
|
2022-11-11 01:50:18
|
googleapis/nodejs-dialogflow-cx
|
https://api.github.com/repos/googleapis/nodejs-dialogflow-cx
|
closed
|
Your .repo-metadata.json file has a problem 🤒
|
type: process api: dialogflow repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'dialogflow-cx' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'dialogflow-cx' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname dialogflow cx invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
16,760
| 21,927,824,515
|
IssuesEvent
|
2022-05-23 06:57:33
|
q191201771/lal
|
https://api.github.com/repos/q191201771/lal
|
closed
|
Unble push stream to Wowza RTMP Server with authencation
|
#Feature *In process * help wanted
|
Hi all,
I am trying push rtmp to wowza with user and password but i cann't work.
without password, it still works normally
My command format: pullrtsp2pushrtmp.go -i rtsp://root:@xxx.196:8555/live1s2.sdp -o rtmp://xxx:1935/hung/mystream -t 0

Plz help me
Tks u
|
1.0
|
Unble push stream to Wowza RTMP Server with authencation - Hi all,
I am trying push rtmp to wowza with user and password but i cann't work.
without password, it still works normally
My command format: pullrtsp2pushrtmp.go -i rtsp://root:@xxx.196:8555/live1s2.sdp -o rtmp://xxx:1935/hung/mystream -t 0

Plz help me
Tks u
|
process
|
unble push stream to wowza rtmp server with authencation hi all i am trying push rtmp to wowza with user and password but i cann t work without password it still works normally my command format go i rtsp root xxx sdp o rtmp xxx hung mystream t plz help me tks u
| 1
|
41,135
| 5,341,357,457
|
IssuesEvent
|
2017-02-17 02:30:36
|
Gallopsled/pwntools
|
https://api.github.com/repos/Gallopsled/pwntools
|
closed
|
Private functions and methods (underscore-prefixed) are not tested
|
bug testing
|
Apparently the Sphinx `doctest` module does not actually run all of the tests:
```py
def filter_members(self, members, want_all):
# type: (List[Tuple[unicode, Any]], bool) -> List[Tuple[unicode, Any, bool]]
"""Filter the given member list.
Members are skipped if
- they are private (except if given explicitly or the private-members
option is set)
- they are special methods (except if given explicitly or the
special-members option is set)
- they are undocumented (except if the undoc-members option is set)
The user can override the skipping decision by connecting to the
``autodoc-skip-member`` event.
"""
```
https://github.com/sphinx-doc/sphinx/blob/master/sphinx/ext/autodoc.py#L855-L861
|
1.0
|
Private functions and methods (underscore-prefixed) are not tested - Apparently the Sphinx `doctest` module does not actually run all of the tests:
```py
def filter_members(self, members, want_all):
# type: (List[Tuple[unicode, Any]], bool) -> List[Tuple[unicode, Any, bool]]
"""Filter the given member list.
Members are skipped if
- they are private (except if given explicitly or the private-members
option is set)
- they are special methods (except if given explicitly or the
special-members option is set)
- they are undocumented (except if the undoc-members option is set)
The user can override the skipping decision by connecting to the
``autodoc-skip-member`` event.
"""
```
https://github.com/sphinx-doc/sphinx/blob/master/sphinx/ext/autodoc.py#L855-L861
|
non_process
|
private functions and methods underscore prefixed are not tested apparently the sphinx doctest module does not actually run all of the tests py def filter members self members want all type list bool list filter the given member list members are skipped if they are private except if given explicitly or the private members option is set they are special methods except if given explicitly or the special members option is set they are undocumented except if the undoc members option is set the user can override the skipping decision by connecting to the autodoc skip member event
| 0
|
2,747
| 5,657,983,719
|
IssuesEvent
|
2017-04-10 08:51:49
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
closed
|
Duplicated process filters when using Tasks App
|
bug comp: activiti-processList
|
<!--
PLEASE FILL OUT THE FOLLOWING INFORMATION, THIS WILL HELP US TO RESOLVE YOUR PROBLEM FASTER.
REMEMBER FOR SUPPORT REQUESTS YOU CAN ALSO ASK ON OUR GITTER CHAT:
Please ask before on our gitter channel https://gitter.im/Alfresco/alfresco-ng2-components
-->
**Type of issue:** (check with "[x]")
```
- [ ] New feature request
- [x] Bug
- [ ] Support request
- [ ] Documentation
```
**Current behavior:**
If there are multiple apps deployed and having processes started or in progress, then filters get duplicated when accessing via `Procceses` tab with Tasks App

**Expected behavior:**
Filters should not be duplicated.
**Steps to reproduce the issue:**
<!-- Describe the steps to reproduce the issue. -->
**Component name and version:**
<!-- Example: ng2-alfresco-login. Check before if this issue is still present in the most recent version -->
**Browser and version:**
<!-- [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] -->
**Node version (for build issues):**
<!-- To check the version: node --version -->
**New feature request:**
<!-- Describe the feature, motivation and the concrete use case (only in case of new feature request) -->
|
1.0
|
Duplicated process filters when using Tasks App - <!--
PLEASE FILL OUT THE FOLLOWING INFORMATION, THIS WILL HELP US TO RESOLVE YOUR PROBLEM FASTER.
REMEMBER FOR SUPPORT REQUESTS YOU CAN ALSO ASK ON OUR GITTER CHAT:
Please ask before on our gitter channel https://gitter.im/Alfresco/alfresco-ng2-components
-->
**Type of issue:** (check with "[x]")
```
- [ ] New feature request
- [x] Bug
- [ ] Support request
- [ ] Documentation
```
**Current behavior:**
If there are multiple apps deployed and having processes started or in progress, then filters get duplicated when accessing via `Procceses` tab with Tasks App

**Expected behavior:**
Filters should not be duplicated.
**Steps to reproduce the issue:**
<!-- Describe the steps to reproduce the issue. -->
**Component name and version:**
<!-- Example: ng2-alfresco-login. Check before if this issue is still present in the most recent version -->
**Browser and version:**
<!-- [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] -->
**Node version (for build issues):**
<!-- To check the version: node --version -->
**New feature request:**
<!-- Describe the feature, motivation and the concrete use case (only in case of new feature request) -->
|
process
|
duplicated process filters when using tasks app please fill out the following information this will help us to resolve your problem faster remember for support requests you can also ask on our gitter chat please ask before on our gitter channel type of issue check with new feature request bug support request documentation current behavior if there are multiple apps deployed and having processes started or in progress then filters get duplicated when accessing via procceses tab with tasks app expected behavior filters should not be duplicated steps to reproduce the issue component name and version browser and version node version for build issues new feature request
| 1
|
7,140
| 10,282,217,186
|
IssuesEvent
|
2019-08-26 10:30:07
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Resource Group requirement for linking
|
Pri2 assigned-to-author automation/svc doc-enhancement process-automation/subsvc triaged
|
Can this article be updated to include the requirement that the log analytics workspace and automation account must belong to the same resource group? ([ref here](https://docs.microsoft.com/en-us/azure/azure-monitor/insights/solutions#log-analytics-workspace-and-automation-account)).
It can be frustrating to miss that requirement when you have created automation accounts and Log Analytics workspaces in a supported region mapping, but they belong to different resource groups.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f8f86bd0-7555-9be2-1015-76e3ab88062f
* Version Independent ID: 1d310402-fadf-d602-048a-b2e16bd86a7e
* Content: [Azure Automation and Log Analytics workspace mappings](https://docs.microsoft.com/en-us/azure/automation/how-to/region-mappings)
* Content Source: [articles/automation/how-to/region-mappings.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/how-to/region-mappings.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @bobbytreed
* Microsoft Alias: **robreed**
|
1.0
|
Resource Group requirement for linking - Can this article be updated to include the requirement that the log analytics workspace and automation account must belong to the same resource group? ([ref here](https://docs.microsoft.com/en-us/azure/azure-monitor/insights/solutions#log-analytics-workspace-and-automation-account)).
It can be frustrating to miss that requirement when you have created automation accounts and Log Analytics workspaces in a supported region mapping, but they belong to different resource groups.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f8f86bd0-7555-9be2-1015-76e3ab88062f
* Version Independent ID: 1d310402-fadf-d602-048a-b2e16bd86a7e
* Content: [Azure Automation and Log Analytics workspace mappings](https://docs.microsoft.com/en-us/azure/automation/how-to/region-mappings)
* Content Source: [articles/automation/how-to/region-mappings.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/how-to/region-mappings.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @bobbytreed
* Microsoft Alias: **robreed**
|
process
|
resource group requirement for linking can this article be updated to include the requirement that the log analytics workspace and automation account must belong to the same resource group it can be frustrating to miss that requirement when you have created automation accounts and log analytics workspaces in a supported region mapping but they belong to different resource groups document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id fadf content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
| 1
|
110,972
| 11,715,352,820
|
IssuesEvent
|
2020-03-09 14:00:29
|
TennersUnipd/sweDocs
|
https://api.github.com/repos/TennersUnipd/sweDocs
|
closed
|
NDP Correggere i rifermenti normativi e informativi
|
documentation enhancement
|
per suaprecisa definizione, lo standard ISO/IEC 12207 si istanzia e non si adotta in quanto tale, in tal modo confluendo nelle norme di progetto. Anche per ISO 8601, sarà più congruo usare un riferimento informativo, preferendo invece istanziare direttamente le (poche) norme di vostro interesse specifico
|
1.0
|
NDP Correggere i rifermenti normativi e informativi - per suaprecisa definizione, lo standard ISO/IEC 12207 si istanzia e non si adotta in quanto tale, in tal modo confluendo nelle norme di progetto. Anche per ISO 8601, sarà più congruo usare un riferimento informativo, preferendo invece istanziare direttamente le (poche) norme di vostro interesse specifico
|
non_process
|
ndp correggere i rifermenti normativi e informativi per suaprecisa definizione lo standard iso iec si istanzia e non si adotta in quanto tale in tal modo confluendo nelle norme di progetto anche per iso sarà più congruo usare un riferimento informativo preferendo invece istanziare direttamente le poche norme di vostro interesse specifico
| 0
|
6,760
| 9,885,304,454
|
IssuesEvent
|
2019-06-25 01:56:05
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Start and Stop Dates
|
automation/svc cxp process-automation/subsvc product-question triaged
|
Should give info in here about setting the dates as its very confusing when you set the date. Just sayin.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#modify-the-startup-and-shutdown-schedules)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @georgewallace
* Microsoft Alias: **gwallace**
|
1.0
|
Start and Stop Dates - Should give info in here about setting the dates as its very confusing when you set the date. Just sayin.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#modify-the-startup-and-shutdown-schedules)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @georgewallace
* Microsoft Alias: **gwallace**
|
process
|
start and stop dates should give info in here about setting the dates as its very confusing when you set the date just sayin document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login georgewallace microsoft alias gwallace
| 1
|
3,411
| 6,523,899,171
|
IssuesEvent
|
2017-08-29 10:28:05
|
w3c/w3process
|
https://api.github.com/repos/w3c/w3process
|
closed
|
Add reference to Code of Ethics and Professional Conduct (CEPC) in the Process Document
|
Process2018Candidate
|
Transferred from https://www.w3.org/community/w3process/track/issues/178
State: Raised
|
1.0
|
Add reference to Code of Ethics and Professional Conduct (CEPC) in the Process Document - Transferred from https://www.w3.org/community/w3process/track/issues/178
State: Raised
|
process
|
add reference to code of ethics and professional conduct cepc in the process document transferred from state raised
| 1
|
73,884
| 15,286,191,554
|
IssuesEvent
|
2021-02-23 14:26:04
|
idonthaveafifaaddiction/IoT-Smart-Gateway
|
https://api.github.com/repos/idonthaveafifaaddiction/IoT-Smart-Gateway
|
opened
|
CVE-2019-1010266 (Medium) detected in lodash-1.0.2.tgz
|
security vulnerability
|
## CVE-2019-1010266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-1.0.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, and extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p>
<p>Path to dependency file: IoT-Smart-Gateway/web/package.json</p>
<p>Path to vulnerable library: IoT-Smart-Gateway/web/node_modules/globule/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- gulp-3.9.1.tgz (Root Library)
- vinyl-fs-0.3.14.tgz
- glob-watcher-0.0.6.tgz
- gaze-0.5.2.tgz
- globule-0.1.0.tgz
- :x: **lodash-1.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/IoT-Smart-Gateway/commit/085c2936b8c1a6952ca6f0db9eaf75ff33e478a1">085c2936b8c1a6952ca6f0db9eaf75ff33e478a1</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p>
<p>Release Date: 2019-07-17</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.0.2","packageFilePaths":["/web/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp:3.9.1;vinyl-fs:0.3.14;glob-watcher:0.0.6;gaze:0.5.2;globule:0.1.0;lodash:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2019-1010266","vulnerabilityDetails":"lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-1010266 (Medium) detected in lodash-1.0.2.tgz - ## CVE-2019-1010266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-1.0.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, and extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p>
<p>Path to dependency file: IoT-Smart-Gateway/web/package.json</p>
<p>Path to vulnerable library: IoT-Smart-Gateway/web/node_modules/globule/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- gulp-3.9.1.tgz (Root Library)
- vinyl-fs-0.3.14.tgz
- glob-watcher-0.0.6.tgz
- gaze-0.5.2.tgz
- globule-0.1.0.tgz
- :x: **lodash-1.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/IoT-Smart-Gateway/commit/085c2936b8c1a6952ca6f0db9eaf75ff33e478a1">085c2936b8c1a6952ca6f0db9eaf75ff33e478a1</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p>
<p>Release Date: 2019-07-17</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.0.2","packageFilePaths":["/web/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp:3.9.1;vinyl-fs:0.3.14;glob-watcher:0.0.6;gaze:0.5.2;globule:0.1.0;lodash:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2019-1010266","vulnerabilityDetails":"lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file iot smart gateway web package json path to vulnerable library iot smart gateway web node modules globule node modules lodash package json dependency hierarchy gulp tgz root library vinyl fs tgz glob watcher tgz gaze tgz globule tgz x lodash tgz vulnerable library found in head commit a href found in base branch develop vulnerability details lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree gulp vinyl fs glob watcher gaze globule lodash isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is vulnerabilityurl
| 0
|
199,933
| 15,786,759,613
|
IssuesEvent
|
2021-04-01 18:12:56
|
NCAR/VAPOR
|
https://api.github.com/repos/NCAR/VAPOR
|
closed
|
Need a link to 500m BigBlueMarble.tms on website
|
Documentation Fixed
|
Our current website does not host our high-res BigBlueMarble tms image.
|
1.0
|
Need a link to 500m BigBlueMarble.tms on website - Our current website does not host our high-res BigBlueMarble tms image.
|
non_process
|
need a link to bigbluemarble tms on website our current website does not host our high res bigbluemarble tms image
| 0
|
161,434
| 6,130,165,756
|
IssuesEvent
|
2017-06-24 02:23:57
|
CAGoodman/CareWheelsCorp
|
https://api.github.com/repos/CAGoodman/CareWheelsCorp
|
opened
|
Need scrub of all the strings in the app
|
bug High Priority
|
Need a full review of all the user visible strings displayed by the app and make sure that they are informative and accurate. For example, the reminder text "Please check in with your CareWheels friendly customer support!" is not correct. Filed as separate issue.
|
1.0
|
Need scrub of all the strings in the app - Need a full review of all the user visible strings displayed by the app and make sure that they are informative and accurate. For example, the reminder text "Please check in with your CareWheels friendly customer support!" is not correct. Filed as separate issue.
|
non_process
|
need scrub of all the strings in the app need a full review of all the user visible strings displayed by the app and make sure that they are informative and accurate for example the reminder text please check in with your carewheels friendly customer support is not correct filed as separate issue
| 0
|
19,992
| 26,466,273,313
|
IssuesEvent
|
2023-01-17 00:11:19
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Use `directUrl` in `prisma studio`
|
process/candidate kind/improvement topic: cli topic: studio team/schema topic: data proxy
|
4.9.0 added `directUrl` for Migration and Introspection commands. We should also use `directUrl` for local Studio.
|
1.0
|
Use `directUrl` in `prisma studio` - 4.9.0 added `directUrl` for Migration and Introspection commands. We should also use `directUrl` for local Studio.
|
process
|
use directurl in prisma studio added directurl for migration and introspection commands we should also use directurl for local studio
| 1
|
273,628
| 20,800,025,984
|
IssuesEvent
|
2022-03-17 13:06:42
|
aws/karpenter
|
https://api.github.com/repos/aws/karpenter
|
opened
|
Add package docs so Karpenter displays better on https://pkg.go.dev/github.com/awslabs/karpenter
|
documentation feature
|
**Tell us about your request**
Improve package documentation so it's parsed and displayed at https://pkg.go.dev/github.com/awslabs/karpenter
**Tell us about the problem you're trying to solve. What are you trying to do, and why is it hard?**
A better user experience for people using https://pkg.go.dev/github.com/awslabs/karpenter to explore the Karpenter source code.
**Are you currently working around this issue?**
N/A
<!-- Please keep this note for the community -->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!-- Thank you for keeping this note for the community -->
|
1.0
|
Add package docs so Karpenter displays better on https://pkg.go.dev/github.com/awslabs/karpenter - **Tell us about your request**
Improve package documentation so it's parsed and displayed at https://pkg.go.dev/github.com/awslabs/karpenter
**Tell us about the problem you're trying to solve. What are you trying to do, and why is it hard?**
A better user experience for people using https://pkg.go.dev/github.com/awslabs/karpenter to explore the Karpenter source code.
**Are you currently working around this issue?**
N/A
<!-- Please keep this note for the community -->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!-- Thank you for keeping this note for the community -->
|
non_process
|
add package docs so karpenter displays better on tell us about your request improve package documentation so it s parsed and displayed at tell us about the problem you re trying to solve what are you trying to do and why is it hard a better user experience for people using to explore the karpenter source code are you currently working around this issue n a community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment
| 0
|
13,441
| 15,882,466,449
|
IssuesEvent
|
2021-04-09 16:03:33
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
"Error validating: The relation field ... uses the scalar fields .... At least one of those fields is required. Hence the relation field must be required as well." is too restrictive when using composite keys
|
bug/2-confirmed kind/bug process/candidate team/client team/migrations
|
I have to say first, that I really fell in love with Prisma. it is such an awesome tool, thanks! 🙏
## Bug description
When using composite keys the existence of relation fields is too strict. It can be optional if one of the key fields is optional - not all of them. I can better explain by example.
## How to reproduce
This schema is working:
```
model Object {
objectId Int @id
name String
parentId Int?
parent Object? @relation(fields: [parentId], references: [objectId])
}
```
`parentId` is optional and so is `parent`.
However my data model is more complex and there is no single unique column:
```
model ObjectCompositeKey {
clientId Int
objectId Int
name String
parentId Int?
parent ObjectCompositeKey? @relation(fields: [clientId, parentId], references: [clientId, objectId])
@@id([clientId, objectId])
}
```
This schema is not working and I receive the error:
```
Error validating: The relation field `parent` uses the scalar fields clientId, parentId. At least one of those fields is required. Hence the relation field must be required as well.
```
## Expected behavior
However, from my point of view the error is wrong. As soon as one of the scalar fields is optional, the relation field has to be optional as well.
It works if a make the `parent` relation field required, but that's not what I as `parentId` is optional and it would prevent me from inserting rows without `parent`.
## Environment & setup
```
prisma : 2.18.0
@prisma/client : 2.18.0
Current platform : windows
Query Engine : query-engine da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at ..\node_modules\@prisma\engines\query-engine-windows.exe)
Migration Engine : migration-engine-cli da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at ..\node_modules\@prisma\engines\migration-engine-windows.exe)
Introspection Engine : introspection-core da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at ..\node_modules\@prisma\engines\introspection-engine-windows.exe)
Format Binary : prisma-fmt da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at ..\node_modules\@prisma\engines\prisma-fmt-windows.exe)
Studio : 0.356.0
Preview Features : createMany, groupBy
```
|
1.0
|
"Error validating: The relation field ... uses the scalar fields .... At least one of those fields is required. Hence the relation field must be required as well." is too restrictive when using composite keys - I have to say first, that I really fell in love with Prisma. it is such an awesome tool, thanks! 🙏
## Bug description
When using composite keys the existence of relation fields is too strict. It can be optional if one of the key fields is optional - not all of them. I can better explain by example.
## How to reproduce
This schema is working:
```
model Object {
objectId Int @id
name String
parentId Int?
parent Object? @relation(fields: [parentId], references: [objectId])
}
```
`parentId` is optional and so is `parent`.
However my data model is more complex and there is no single unique column:
```
model ObjectCompositeKey {
clientId Int
objectId Int
name String
parentId Int?
parent ObjectCompositeKey? @relation(fields: [clientId, parentId], references: [clientId, objectId])
@@id([clientId, objectId])
}
```
This schema is not working and I receive the error:
```
Error validating: The relation field `parent` uses the scalar fields clientId, parentId. At least one of those fields is required. Hence the relation field must be required as well.
```
## Expected behavior
However, from my point of view the error is wrong. As soon as one of the scalar fields is optional, the relation field has to be optional as well.
It works if a make the `parent` relation field required, but that's not what I as `parentId` is optional and it would prevent me from inserting rows without `parent`.
## Environment & setup
```
prisma : 2.18.0
@prisma/client : 2.18.0
Current platform : windows
Query Engine : query-engine da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at ..\node_modules\@prisma\engines\query-engine-windows.exe)
Migration Engine : migration-engine-cli da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at ..\node_modules\@prisma\engines\migration-engine-windows.exe)
Introspection Engine : introspection-core da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at ..\node_modules\@prisma\engines\introspection-engine-windows.exe)
Format Binary : prisma-fmt da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at ..\node_modules\@prisma\engines\prisma-fmt-windows.exe)
Studio : 0.356.0
Preview Features : createMany, groupBy
```
|
process
|
error validating the relation field uses the scalar fields at least one of those fields is required hence the relation field must be required as well is too restrictive when using composite keys i have to say first that i really fell in love with prisma it is such an awesome tool thanks 🙏 bug description when using composite keys the existence of relation fields is too strict it can be optional if one of the key fields is optional not all of them i can better explain by example how to reproduce this schema is working model object objectid int id name string parentid int parent object relation fields references parentid is optional and so is parent however my data model is more complex and there is no single unique column model objectcompositekey clientid int objectid int name string parentid int parent objectcompositekey relation fields references id this schema is not working and i receive the error error validating the relation field parent uses the scalar fields clientid parentid at least one of those fields is required hence the relation field must be required as well expected behavior however from my point of view the error is wrong as soon as one of the scalar fields is optional the relation field has to be optional as well it works if a make the parent relation field required but that s not what i as parentid is optional and it would prevent me from inserting rows without parent environment setup prisma prisma client current platform windows query engine query engine at node modules prisma engines query engine windows exe migration engine migration engine cli at node modules prisma engines migration engine windows exe introspection engine introspection core at node modules prisma engines introspection engine windows exe format binary prisma fmt at node modules prisma engines prisma fmt windows exe studio preview features createmany groupby
| 1
|
63,904
| 18,052,289,797
|
IssuesEvent
|
2021-09-19 23:43:30
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
The sidebar is not able to collapse properly.
|
T-Defect
|
### Steps to reproduce
1. Drag the sidebar left-ward to collapse it so that its width shall span one column.
### What happened?
### What did you expect?
The side-bar should be displayed correctly.
### What happened?
http://user-images.githubusercontent.com/42837531/133940420-fbf9f0a3-21b7-487b-b637-8a27ee8876ba.PNG
### Operating system
http://silverblue.fedoraproject.org
### Application version
Element version: 1.8.4 Olm version: 3.2.3
### How did you install the app?
http://github.com/flathub/im.riot.Riot
### Homeserver
//MATRIX.ORG
### Have you submitted a rageshake?
No
|
1.0
|
The sidebar is not able to collapse properly. - ### Steps to reproduce
1. Drag the sidebar left-ward to collapse it so that its width shall span one column.
### What happened?
### What did you expect?
The side-bar should be displayed correctly.
### What happened?
http://user-images.githubusercontent.com/42837531/133940420-fbf9f0a3-21b7-487b-b637-8a27ee8876ba.PNG
### Operating system
http://silverblue.fedoraproject.org
### Application version
Element version: 1.8.4 Olm version: 3.2.3
### How did you install the app?
http://github.com/flathub/im.riot.Riot
### Homeserver
//MATRIX.ORG
### Have you submitted a rageshake?
No
|
non_process
|
the sidebar is not able to collapse properly steps to reproduce drag the sidebar left ward to collapse it so that its width shall span one column what happened what did you expect the side bar should be displayed correctly what happened operating system application version element version olm version how did you install the app homeserver matrix org have you submitted a rageshake no
| 0
|
4,423
| 7,302,064,730
|
IssuesEvent
|
2018-02-27 08:24:27
|
muflihun/residue
|
https://api.github.com/repos/muflihun/residue
|
closed
|
Separate log dispatchers and logging queue by clients
|
area: log-processing type: improvement
|
This is subject to [issue 613 on Easylogging++](https://github.com/muflihun/easyloggingpp/issues/613). That needs to be fixed before this should be worked on otherwise we may run in to race condition
This is an improvement to prevent one client blocking other client's logs in case of “rush hour”
We will have separate queue for unknown clients and one queue for each known client initialized at the startup
|
1.0
|
Separate log dispatchers and logging queue by clients - This is subject to [issue 613 on Easylogging++](https://github.com/muflihun/easyloggingpp/issues/613). That needs to be fixed before this should be worked on otherwise we may run in to race condition
This is an improvement to prevent one client blocking other client's logs in case of “rush hour”
We will have separate queue for unknown clients and one queue for each known client initialized at the startup
|
process
|
separate log dispatchers and logging queue by clients this is subject to that needs to be fixed before this should be worked on otherwise we may run in to race condition this is an improvement to prevent one client blocking other client s logs in case of “rush hour” we will have separate queue for unknown clients and one queue for each known client initialized at the startup
| 1
|
162,405
| 13,888,021,129
|
IssuesEvent
|
2020-10-19 05:25:32
|
NeilBaner/alpha
|
https://api.github.com/repos/NeilBaner/alpha
|
opened
|
Dududududu
|
severity.VeryLow type.DocumentationBug
|
# This is a test of Markdown syntax
## Let's see if headings work
**what about** *things* `like` ~~this~~?
```c
int main() {
printf("haha\n");
}
```
> do quotes work?
<!--session: 1603084447703-30b8f4fe-9a3f-489e-aeae-23241ae83ac7-->
|
1.0
|
Dududududu - # This is a test of Markdown syntax
## Let's see if headings work
**what about** *things* `like` ~~this~~?
```c
int main() {
printf("haha\n");
}
```
> do quotes work?
<!--session: 1603084447703-30b8f4fe-9a3f-489e-aeae-23241ae83ac7-->
|
non_process
|
dududududu this is a test of markdown syntax let s see if headings work what about things like this c int main printf haha n do quotes work
| 0
|
18,552
| 4,286,351,530
|
IssuesEvent
|
2016-07-16 02:36:03
|
ambisonictoolkit/atk-sc3
|
https://api.github.com/repos/ambisonictoolkit/atk-sc3
|
closed
|
Update soundfield transform figures to illustrate rE
|
documentation
|
Update the indication of perceived localization focus from |rV| to |rE|.
Mapping to |rE| offers a stronger sense of how localized an encoded planewave will appear after transform.
This change is parallel to [issue no. 24 for atk-reaper](https://github.com/ambisonictoolkit/atk-reaper/issues/24).
|
1.0
|
Update soundfield transform figures to illustrate rE - Update the indication of perceived localization focus from |rV| to |rE|.
Mapping to |rE| offers a stronger sense of how localized an encoded planewave will appear after transform.
This change is parallel to [issue no. 24 for atk-reaper](https://github.com/ambisonictoolkit/atk-reaper/issues/24).
|
non_process
|
update soundfield transform figures to illustrate re update the indication of perceived localization focus from rv to re mapping to re offers a stronger sense of how localized an encoded planewave will appear after transform this change is parallel to
| 0
|
9,412
| 12,406,938,484
|
IssuesEvent
|
2020-05-21 20:05:56
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Build.DefinitionName ?
|
Pri1 devops-cicd-process/tech devops/prod doc-bug
|
Both $(Build.DefinitionName) and $(BuildDefinitionName) are mentioned. Is the first one a typo ? If not, could you explain the difference ?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93
* Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7
* Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml#feedback)
* Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Build.DefinitionName ? - Both $(Build.DefinitionName) and $(BuildDefinitionName) are mentioned. Is the first one a typo ? If not, could you explain the difference ?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93
* Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7
* Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml#feedback)
* Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
build definitionname both build definitionname and builddefinitionname are mentioned is the first one a typo if not could you explain the difference document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
185,132
| 6,719,299,584
|
IssuesEvent
|
2017-10-15 22:35:41
|
adri/estimator
|
https://api.github.com/repos/adri/estimator
|
opened
|
Story: Add spectator role
|
Priority: Medium Status: To Do Type: Enhancement
|
_As a **moderator** I want to be able to **set a user as spectator** so he can follow the estimations without the necessity of voting_
- The spectator should't be able to see the votes before the round is done.
- The spectator can't vote
|
1.0
|
Story: Add spectator role - _As a **moderator** I want to be able to **set a user as spectator** so he can follow the estimations without the necessity of voting_
- The spectator should't be able to see the votes before the round is done.
- The spectator can't vote
|
non_process
|
story add spectator role as a moderator i want to be able to set a user as spectator so he can follow the estimations without the necessity of voting the spectator should t be able to see the votes before the round is done the spectator can t vote
| 0
|
13,000
| 15,360,329,351
|
IssuesEvent
|
2021-03-01 16:49:01
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Why don't we support condition property for template?
|
devops-cicd-process/tech devops/prod support-request
|
I have 4 tasks inside a template, is there any way to ignore a whole template rather than ignore every single task inside a template?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Why don't we support condition property for template? - I have 4 tasks inside a template, is there any way to ignore a whole template rather than ignore every single task inside a template?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
why don t we support condition property for template i have tasks inside a template is there any way to ignore a whole template rather than ignore every single task inside a template document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
353,976
| 10,561,387,664
|
IssuesEvent
|
2019-10-04 15:45:53
|
fgpv-vpgf/fgpv-vpgf
|
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
|
closed
|
Projection change warning label is not showing
|
bug-type: regression priority: low problem: bug type: corrective
|
The warning label on the basemap selector has not been showing for quite some time; the directive was broken in a refactor
|
1.0
|
Projection change warning label is not showing - The warning label on the basemap selector has not been showing for quite some time; the directive was broken in a refactor
|
non_process
|
projection change warning label is not showing the warning label on the basemap selector has not been showing for quite some time the directive was broken in a refactor
| 0
|
29,423
| 4,501,212,200
|
IssuesEvent
|
2016-09-01 08:37:12
|
mattbearman/lime
|
https://api.github.com/repos/mattbearman/lime
|
closed
|
BugMuncher Feedback Report
|
bug BugMuncher suggestion test
|
## Details ##
**Submitted:** March 11, 2016 22:03
**Category:** suggestion
**Sender Email:** matt.kabus@ampermusic.com
**Website:** BugMuncher App
**URL:** https://app.bugmuncher.com/user/dashboard
**Operating System:** Mac OS X Yosemite
**Browser:** Chrome 48.0.2564.116
**Browser Size:** 1442 x 837
**User Agent:**
Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/48.0.2564.116 Safari/537.36
**Description:**
TEST
## Custom Data ##
**user_id:** 1218
**account_id:** 803
**plan:** BugMuncher Corporate
## Browser Plugins ##
Widevine Content Decryption Module
Shockwave Flash
Chrome PDF Viewer
Native Client
## Events ##
**content:** Feedback Button Clicked
**timestamp:** Fri Mar 11 2016 17:03:38 GMT-0500 (EST)
**type:** bugmuncher
---
**type:** bugmuncher
**content:** Feedback Report Submitted
**timestamp:** Fri Mar 11 2016 17:03:51 GMT-0500 (EST)
---
|
1.0
|
BugMuncher Feedback Report - ## Details ##
**Submitted:** March 11, 2016 22:03
**Category:** suggestion
**Sender Email:** matt.kabus@ampermusic.com
**Website:** BugMuncher App
**URL:** https://app.bugmuncher.com/user/dashboard
**Operating System:** Mac OS X Yosemite
**Browser:** Chrome 48.0.2564.116
**Browser Size:** 1442 x 837
**User Agent:**
Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/48.0.2564.116 Safari/537.36
**Description:**
TEST
## Custom Data ##
**user_id:** 1218
**account_id:** 803
**plan:** BugMuncher Corporate
## Browser Plugins ##
Widevine Content Decryption Module
Shockwave Flash
Chrome PDF Viewer
Native Client
## Events ##
**content:** Feedback Button Clicked
**timestamp:** Fri Mar 11 2016 17:03:38 GMT-0500 (EST)
**type:** bugmuncher
---
**type:** bugmuncher
**content:** Feedback Report Submitted
**timestamp:** Fri Mar 11 2016 17:03:51 GMT-0500 (EST)
---
|
non_process
|
bugmuncher feedback report details submitted march category suggestion sender email matt kabus ampermusic com website bugmuncher app url operating system mac os x yosemite browser chrome browser size x user agent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari description test custom data user id account id plan bugmuncher corporate browser plugins widevine content decryption module shockwave flash chrome pdf viewer native client events content feedback button clicked timestamp fri mar gmt est type bugmuncher type bugmuncher content feedback report submitted timestamp fri mar gmt est
| 0
|
106,715
| 23,272,387,657
|
IssuesEvent
|
2022-08-05 01:35:16
|
fprime-community/fpp
|
https://api.github.com/repos/fprime-community/fpp
|
closed
|
Remove typedef enum from generated C++ code
|
code generation
|
It has been pointed out that in C++, instead of writing
```
typedef enum { ... } E;
```
we should write
```
enum E { ... };
```
|
1.0
|
Remove typedef enum from generated C++ code - It has been pointed out that in C++, instead of writing
```
typedef enum { ... } E;
```
we should write
```
enum E { ... };
```
|
non_process
|
remove typedef enum from generated c code it has been pointed out that in c instead of writing typedef enum e we should write enum e
| 0
|
3,627
| 6,663,545,879
|
IssuesEvent
|
2017-10-02 16:47:57
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Process.ProcessName is expensive on Windows
|
area-System.Diagnostics.Process enhancement tenet-performance
|
It causes an enumeration of all processes just to get the name of the current process
https://source.dot.net/#System.Diagnostics.Process/System/Diagnostics/Process.cs,900
https://source.dot.net/#System.Diagnostics.Process/System/Diagnostics/ProcessManager.Windows.cs,53
On Windows it could call just GetProcessName. There may be other places it could avoid examining all processes just to get information about the current process. (All this is arguably fallout of giving the Process class three responsibilities -- spawning processes, representing the current process, and accessing general process information)
Feedback from internal service that calls `Process.ProcessName` often for logging on Windows.
|
1.0
|
Process.ProcessName is expensive on Windows - It causes an enumeration of all processes just to get the name of the current process
https://source.dot.net/#System.Diagnostics.Process/System/Diagnostics/Process.cs,900
https://source.dot.net/#System.Diagnostics.Process/System/Diagnostics/ProcessManager.Windows.cs,53
On Windows it could call just GetProcessName. There may be other places it could avoid examining all processes just to get information about the current process. (All this is arguably fallout of giving the Process class three responsibilities -- spawning processes, representing the current process, and accessing general process information)
Feedback from internal service that calls `Process.ProcessName` often for logging on Windows.
|
process
|
process processname is expensive on windows it causes an enumeration of all processes just to get the name of the current process on windows it could call just getprocessname there may be other places it could avoid examining all processes just to get information about the current process all this is arguably fallout of giving the process class three responsibilities spawning processes representing the current process and accessing general process information feedback from internal service that calls process processname often for logging on windows
| 1
|
207,340
| 7,127,385,724
|
IssuesEvent
|
2018-01-20 21:05:01
|
ksu-svt/svt-event-web-app
|
https://api.github.com/repos/ksu-svt/svt-event-web-app
|
closed
|
Users can view detail pages of members/events.
|
Low Priority feature
|
Users should be able to navigate the member and event lists, click on one item, and be taken to a detail page that shows all appropriate information for that specific entry.
|
1.0
|
Users can view detail pages of members/events. - Users should be able to navigate the member and event lists, click on one item, and be taken to a detail page that shows all appropriate information for that specific entry.
|
non_process
|
users can view detail pages of members events users should be able to navigate the member and event lists click on one item and be taken to a detail page that shows all appropriate information for that specific entry
| 0
|
21,065
| 28,013,511,052
|
IssuesEvent
|
2023-03-27 20:31:00
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Routing based on the field value extracted from the logs is not working
|
bug processor/attributes
|
### Component(s)
processor/routing
### Describe the issue you're reporting
I have a requirement to send logs to different exporters through a collector gateway, for testing purpose i inserted some test messages into the logs which would capture a field value and based on that value, OTEL should send logs to different exporters. This is my processor information which obviously includes attributes processor which will add `environment` field to the logs and `routing` processor will use `environment` field values ( prod or non prod) to send events to different exporters.
```processors:
attributes/1:
actions:
- action: extract
key: message
pattern: \{\"message\"\:\s\"(?P<environment>\w+)
- key: environment
action: insert
from_attribute: message
routing:
from_attribute: environment
default_exporters: [file/no_rotation]
table:
- value : 'nonprod'
exporters: [ file/no_rotation ]
- value: 'prod'
exporters: [splunk_hec]
```
Here is the same log :
`{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "prod: some log data"}
{"message": "prod: some log data"}
"hec.log" 169L, 7223B
{"message": "prod: some log data"}
{"message": "prod: some log data"} `
As it can be seen from the `regex` pattern and sample logs, that only log lines which have `environment=prod` will go to `splunk_hec` and rest will go to `default` exporter. After several rounds of testing, i am still not able to send logs to non default exporters. Is there any guidlines around the type of logs which are suited for conditional routing ?
|
1.0
|
Routing based on the field value extracted from the logs is not working - ### Component(s)
processor/routing
### Describe the issue you're reporting
I have a requirement to send logs to different exporters through a collector gateway, for testing purpose i inserted some test messages into the logs which would capture a field value and based on that value, OTEL should send logs to different exporters. This is my processor information which obviously includes attributes processor which will add `environment` field to the logs and `routing` processor will use `environment` field values ( prod or non prod) to send events to different exporters.
```processors:
attributes/1:
actions:
- action: extract
key: message
pattern: \{\"message\"\:\s\"(?P<environment>\w+)
- key: environment
action: insert
from_attribute: message
routing:
from_attribute: environment
default_exporters: [file/no_rotation]
table:
- value : 'nonprod'
exporters: [ file/no_rotation ]
- value: 'prod'
exporters: [splunk_hec]
```
Here is the same log :
`{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "nonprod: some other log data"}
{"message": "prod: some log data"}
{"message": "prod: some log data"}
"hec.log" 169L, 7223B
{"message": "prod: some log data"}
{"message": "prod: some log data"} `
As it can be seen from the `regex` pattern and sample logs, that only log lines which have `environment=prod` will go to `splunk_hec` and rest will go to `default` exporter. After several rounds of testing, i am still not able to send logs to non default exporters. Is there any guidlines around the type of logs which are suited for conditional routing ?
|
process
|
routing based on the field value extracted from the logs is not working component s processor routing describe the issue you re reporting i have a requirement to send logs to different exporters through a collector gateway for testing purpose i inserted some test messages into the logs which would capture a field value and based on that value otel should send logs to different exporters this is my processor information which obviously includes attributes processor which will add environment field to the logs and routing processor will use environment field values prod or non prod to send events to different exporters processors attributes actions action extract key message pattern message s p w key environment action insert from attribute message routing from attribute environment default exporters table value nonprod exporters value prod exporters here is the same log message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message nonprod some other log data message prod some log data message prod some log data hec log message prod some log data message prod some log data as it can be seen from the regex pattern and sample logs that only log lines which have environment prod will go to splunk hec and rest will go to default exporter after several rounds of testing i am still not able to send logs to non default exporters is there any guidlines around the type of logs which are suited for conditional routing
| 1
|
23,887
| 3,864,007,424
|
IssuesEvent
|
2016-04-08 12:07:53
|
contao/core
|
https://api.github.com/repos/contao/core
|
closed
|
Widget::getPost() minor issue
|
defect
|
This statement in ```Widget::getPost()``` is always true and will be always executed:
https://github.com/contao/core/blob/master/system/modules/core/library/Contao/Widget.php#L828
Thus the last line of the method will never be executed:
https://github.com/contao/core/blob/master/system/modules/core/library/Contao/Widget.php#L845
|
1.0
|
Widget::getPost() minor issue - This statement in ```Widget::getPost()``` is always true and will be always executed:
https://github.com/contao/core/blob/master/system/modules/core/library/Contao/Widget.php#L828
Thus the last line of the method will never be executed:
https://github.com/contao/core/blob/master/system/modules/core/library/Contao/Widget.php#L845
|
non_process
|
widget getpost minor issue this statement in widget getpost is always true and will be always executed thus the last line of the method will never be executed
| 0
|
806,337
| 29,812,056,486
|
IssuesEvent
|
2023-06-16 15:48:30
|
tektoncd/pipeline
|
https://api.github.com/repos/tektoncd/pipeline
|
closed
|
Steps are getting volume mounts from other steps
|
kind/bug priority/critical-urgent
|
# Expected Behavior
I should get the volume mounts for the step that I specify.
# Actual Behavior
Steps are getting the wrong volume mounts which are the same as a previous step.
I think the root cause might be here: https://github.com/tektoncd/pipeline/blob/70a2d714dcbaefb952924edf6df0c217aff80548/pkg/pod/pod.go#L283 as I suspect that if you are in the case where `toAdd` is empty, the `append` may return the same slice from the iterator... though I think that could just be a mis-reading on my behalf
# Steps to Reproduce the Problem
Working on trying to get the minimal steps. I haven't finished cleaning this up and it may not be a functional test, but it's as close to the YAML that is reproducing the issue for me as I can edit by hand right now
I have a task that is something like this:
```yaml
apiVersion: tekton.dev/v1
kind: Task
metadata:
creationTimestamp: null
name: test-test-3c9a407df28bbcdbe55bd21af2f05c97b79fd29ccf1a8a20f7801
spec:
description: test
params:
- default: workflow_dispatch
name: eventName
type: string
- default: /home/runner
name: home
type: string
- name: needs_build_output_foo
type: string
- name: needs_build_result
type: string
results:
- name: status
type: string
stepTemplate:
computeResources: {}
volumeMounts:
- mountPath: $(params.home)/sbin
name: sbin
readOnly: true
- mountPath: $(params.home)
name: home
- mountPath: $(params.home)/var/event.json
name: event
readOnly: true
subPath: event.json
workingDir: $(params.home)/workspace
steps:
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: steps1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/state
name: state-s1
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-0/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionsfirstyaml
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionsfirstyaml-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionsfirstyaml
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-1/s1
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionssecondyaml
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionssecondyaml-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionssecondyamls1-7dbaa632b88e349a08a9d3b89637
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionssecondyaml
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-2/s1
name: state-fooactionssecondyamls1-7dbaa632b88e349a08a9d3b89637
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyamls2-42cfc2b2edd3d2cf335f46c323d0e
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionsfirstyaml-2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-3/s1
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
readOnly: true
- mountPath: $(params.home)/var/steps-3/s2
name: state-fooactionsfirstyamls2-42cfc2b2edd3d2cf335f46c323d0e
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionsfirstyaml-3
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-4/s1
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
readOnly: true
- mountPath: $(params.home)/var/steps-4/s2
name: state-fooactionsfirstyamls2-42cfc2b2edd3d2cf335f46c323d0e
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run-2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-5/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionssecondyaml2-337af8d140bc8aabc62dfadd24c3
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionssecondyaml-2-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml2-337af8d140bc8aabc62dfadd24c3
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionssecondyaml2s1-aea194106fc3b4402d0aacd8ed9
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionssecondyaml-2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-6/s1
name: state-fooactionssecondyaml2s1-aea194106fc3b4402d0aacd8ed9
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml2-337af8d140bc8aabc62dfadd24c3
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run-3
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-7/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionssecondyaml3-b743709819c6f07beac2c860aa3b
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionssecondyaml-3-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml3-b743709819c6f07beac2c860aa3b
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionssecondyaml3s1-77f8a545e3cf0829197392edba8
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionssecondyaml-3
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-8/s1
name: state-fooactionssecondyaml3s1-77f8a545e3cf0829197392edba8
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml3-b743709819c6f07beac2c860aa3b
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run-4
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-9/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionsfirstyaml-2-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionsfirstyaml-2-2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-10/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionssecondyaml-4-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionssecondyaml4s1-4eceec73b11f30970d1abc3c46d
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionssecondyaml-4
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-11/s1
name: state-fooactionssecondyaml4s1-4eceec73b11f30970d1abc3c46d
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyaml2s2-e94997f47e66fe37432a4ca30a7d
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionsfirstyaml-2-3
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-12/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: $(params.home)/var/steps-12/s2
name: state-fooactionsfirstyaml2s2-e94997f47e66fe37432a4ca30a7d
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionsfirstyaml-2-s2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-13/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: $(params.home)/var/state
name: state-s2
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: step--run-5
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-14/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/steps-14/s2
name: state-s2
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run-6
volumeMounts:
- mountPath: $(params.home)/var/steps-15/s2
name: state-s2
readOnly: true
- mountPath: $(params.home)/var/steps-15/s1
name: state-s1
readOnly: true
workingDir: $(params.home)/workspace
volumes:
- configMap:
name: test-event-35c9df848e0817e4c879b8cd5ba797c53a26877c9bf1899139b6
name: event
- emptyDir: {}
name: home
- emptyDir: {}
name: inputs-fooactionsfirstyaml
- emptyDir: {}
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
- emptyDir: {}
name: inputs-fooactionssecondyaml
- emptyDir: {}
name: inputs-fooactionssecondyaml2-337af8d140bc8aabc62dfadd24c3
- emptyDir: {}
name: inputs-fooactionssecondyaml3-b743709819c6f07beac2c860aa3b
- emptyDir: {}
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
- emptyDir: {}
name: sbin
- emptyDir: {}
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
- emptyDir: {}
name: state-fooactionsfirstyaml2s2-e94997f47e66fe37432a4ca30a7d
- emptyDir: {}
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
- emptyDir: {}
name: state-fooactionsfirstyamls2-42cfc2b2edd3d2cf335f46c323d0e
- emptyDir: {}
name: state-fooactionssecondyaml2s1-aea194106fc3b4402d0aacd8ed9
- emptyDir: {}
name: state-fooactionssecondyaml3s1-77f8a545e3cf0829197392edba8
- emptyDir: {}
name: state-fooactionssecondyaml4s1-4eceec73b11f30970d1abc3c46d
- emptyDir: {}
name: state-fooactionssecondyamls1-7dbaa632b88e349a08a9d3b89637
- emptyDir: {}
name: state-s1
- emptyDir: {}
name: state-s2
```
And then when I look at `stepfooactionsfirstyaml-2-s1` in Tekton Dashboard I will see on the details tab:
```yaml
name: stepfooactionsfirstyaml-2-s1
resources: {}
volumeMounts:
- mountPath: /home/runner/sbin
name: sbin
readOnly: true
- mountPath: /home/runner
name: home
- mountPath: /home/runner/var/event.json
name: event
readOnly: true
subPath: event.json
- mountPath: /home/runner/var/state
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
- mountPath: /home/runner/var/steps-13/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: /home/runner/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: /home/runner/var/steps-12/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: /home/runner/var/steps-12/s2
name: state-fooactionsfirstyaml2s2-e94997f47e66fe37432a4ca30a7d
readOnly: true
- mountPath: /home/runner/var/steps-11/s1
name: state-fooactionssecondyaml4s1-4eceec73b11f30970d1abc3c46d
readOnly: true
- mountPath: /home/runner/var/steps-10/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: /home/runner/var/with
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
workingDir: /home/runner/workspace
script: ls -lRA /home/runner
image: alpine:latest
onError: continue
```
Which seems like far far too many mounts
I am expecting to see
```yaml
- mountPath: /home/runner/sbin
name: sbin
readOnly: true
- mountPath: /home/runner
name: home
- mountPath: /home/runner/var/event.json
name: event
readOnly: true
subPath: event.json
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
```
# Additional Info
- Kubernetes version:
**Output of `kubectl version`:**
```
(Client Version: version.Info{Major:"1", Minor:"25", GitVersion:"v1.25.4", GitCommit:"872a965c6c6526caa949f0c6ac028ef7aff3fb78", GitTreeState:"clean", BuildDate:"2022-11-09T13:36:36Z", GoVersion:"go1.19.3", Compiler:"gc", Platform:"darwin/amd64"}
Kustomize Version: v4.5.7
Server Version: version.Info{Major:"1", Minor:"24", GitVersion:"v1.24.4+k3s1", GitCommit:"c3f830e9b9ed8a4d9d0e2aa663b4591b923a296e", GitTreeState:"clean", BuildDate:"2022-08-25T03:45:26Z", GoVersion:"go1.18.1", Compiler:"gc", Platform:"linux/amd64"}
``
- Tekton Pipeline version:
**Output of `tkn version` or `kubectl get pods -n tekton-pipelines -l app=tekton-pipelines-controller -o=jsonpath='{.items[0].metadata.labels.version}'`**
```
v0.45.0
```
<!-- Any other additional information -->
|
1.0
|
Steps are getting volume mounts from other steps - # Expected Behavior
I should get the volume mounts for the step that I specify.
# Actual Behavior
Steps are getting the wrong volume mounts which are the same as a previous step.
I think the root cause might be here: https://github.com/tektoncd/pipeline/blob/70a2d714dcbaefb952924edf6df0c217aff80548/pkg/pod/pod.go#L283 as I suspect that if you are in the case where `toAdd` is empty, the `append` may return the same slice from the iterator... though I think that could just be a mis-reading on my behalf
# Steps to Reproduce the Problem
Working on trying to get the minimal steps. I haven't finished cleaning this up and it may not be a functional test, but it's as close to the YAML that is reproducing the issue for me as I can edit by hand right now
I have a task that is something like this:
```yaml
apiVersion: tekton.dev/v1
kind: Task
metadata:
creationTimestamp: null
name: test-test-3c9a407df28bbcdbe55bd21af2f05c97b79fd29ccf1a8a20f7801
spec:
description: test
params:
- default: workflow_dispatch
name: eventName
type: string
- default: /home/runner
name: home
type: string
- name: needs_build_output_foo
type: string
- name: needs_build_result
type: string
results:
- name: status
type: string
stepTemplate:
computeResources: {}
volumeMounts:
- mountPath: $(params.home)/sbin
name: sbin
readOnly: true
- mountPath: $(params.home)
name: home
- mountPath: $(params.home)/var/event.json
name: event
readOnly: true
subPath: event.json
workingDir: $(params.home)/workspace
steps:
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: steps1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/state
name: state-s1
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-0/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionsfirstyaml
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionsfirstyaml-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionsfirstyaml
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-1/s1
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionssecondyaml
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionssecondyaml-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionssecondyamls1-7dbaa632b88e349a08a9d3b89637
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionssecondyaml
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-2/s1
name: state-fooactionssecondyamls1-7dbaa632b88e349a08a9d3b89637
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyamls2-42cfc2b2edd3d2cf335f46c323d0e
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionsfirstyaml-2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-3/s1
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
readOnly: true
- mountPath: $(params.home)/var/steps-3/s2
name: state-fooactionsfirstyamls2-42cfc2b2edd3d2cf335f46c323d0e
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionsfirstyaml-3
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-4/s1
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
readOnly: true
- mountPath: $(params.home)/var/steps-4/s2
name: state-fooactionsfirstyamls2-42cfc2b2edd3d2cf335f46c323d0e
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run-2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-5/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionssecondyaml2-337af8d140bc8aabc62dfadd24c3
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionssecondyaml-2-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml2-337af8d140bc8aabc62dfadd24c3
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionssecondyaml2s1-aea194106fc3b4402d0aacd8ed9
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionssecondyaml-2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-6/s1
name: state-fooactionssecondyaml2s1-aea194106fc3b4402d0aacd8ed9
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml2-337af8d140bc8aabc62dfadd24c3
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run-3
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-7/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionssecondyaml3-b743709819c6f07beac2c860aa3b
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionssecondyaml-3-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml3-b743709819c6f07beac2c860aa3b
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionssecondyaml3s1-77f8a545e3cf0829197392edba8
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionssecondyaml-3
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-8/s1
name: state-fooactionssecondyaml3s1-77f8a545e3cf0829197392edba8
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml3-b743709819c6f07beac2c860aa3b
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run-4
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-9/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionsfirstyaml-2-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionsfirstyaml-2-2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-10/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: $(params.home)/var/with
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionssecondyaml-4-s1
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionssecondyaml4s1-4eceec73b11f30970d1abc3c46d
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionssecondyaml-4
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-11/s1
name: state-fooactionssecondyaml4s1-4eceec73b11f30970d1abc3c46d
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyaml2s2-e94997f47e66fe37432a4ca30a7d
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: stepfooactionsfirstyaml-2-3
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-12/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: $(params.home)/var/steps-12/s2
name: state-fooactionsfirstyaml2s2-e94997f47e66fe37432a4ca30a7d
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: stepfooactionsfirstyaml-2-s2
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-13/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: $(params.home)/var/state
name: state-s2
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
name: step--run-5
onError: continue
volumeMounts:
- mountPath: $(params.home)/var/steps-14/s1
name: state-s1
readOnly: true
- mountPath: $(params.home)/var/steps-14/s2
name: state-s2
readOnly: true
workingDir: $(params.home)/workspace
- script: ls -lRa $(params.home)
computeResources: {}
image: alpine:latest
imagePullPolicy: Never
name: step--run-6
volumeMounts:
- mountPath: $(params.home)/var/steps-15/s2
name: state-s2
readOnly: true
- mountPath: $(params.home)/var/steps-15/s1
name: state-s1
readOnly: true
workingDir: $(params.home)/workspace
volumes:
- configMap:
name: test-event-35c9df848e0817e4c879b8cd5ba797c53a26877c9bf1899139b6
name: event
- emptyDir: {}
name: home
- emptyDir: {}
name: inputs-fooactionsfirstyaml
- emptyDir: {}
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
- emptyDir: {}
name: inputs-fooactionssecondyaml
- emptyDir: {}
name: inputs-fooactionssecondyaml2-337af8d140bc8aabc62dfadd24c3
- emptyDir: {}
name: inputs-fooactionssecondyaml3-b743709819c6f07beac2c860aa3b
- emptyDir: {}
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
- emptyDir: {}
name: sbin
- emptyDir: {}
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
- emptyDir: {}
name: state-fooactionsfirstyaml2s2-e94997f47e66fe37432a4ca30a7d
- emptyDir: {}
name: state-fooactionsfirstyamls1-e56457c408805b33bccee9465fabd
- emptyDir: {}
name: state-fooactionsfirstyamls2-42cfc2b2edd3d2cf335f46c323d0e
- emptyDir: {}
name: state-fooactionssecondyaml2s1-aea194106fc3b4402d0aacd8ed9
- emptyDir: {}
name: state-fooactionssecondyaml3s1-77f8a545e3cf0829197392edba8
- emptyDir: {}
name: state-fooactionssecondyaml4s1-4eceec73b11f30970d1abc3c46d
- emptyDir: {}
name: state-fooactionssecondyamls1-7dbaa632b88e349a08a9d3b89637
- emptyDir: {}
name: state-s1
- emptyDir: {}
name: state-s2
```
And then when I look at `stepfooactionsfirstyaml-2-s1` in Tekton Dashboard I will see on the details tab:
```yaml
name: stepfooactionsfirstyaml-2-s1
resources: {}
volumeMounts:
- mountPath: /home/runner/sbin
name: sbin
readOnly: true
- mountPath: /home/runner
name: home
- mountPath: /home/runner/var/event.json
name: event
readOnly: true
subPath: event.json
- mountPath: /home/runner/var/state
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
- mountPath: /home/runner/var/steps-13/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: /home/runner/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: /home/runner/var/steps-12/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: /home/runner/var/steps-12/s2
name: state-fooactionsfirstyaml2s2-e94997f47e66fe37432a4ca30a7d
readOnly: true
- mountPath: /home/runner/var/steps-11/s1
name: state-fooactionssecondyaml4s1-4eceec73b11f30970d1abc3c46d
readOnly: true
- mountPath: /home/runner/var/steps-10/s1
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
readOnly: true
- mountPath: /home/runner/var/with
name: inputs-fooactionssecondyaml4-f0436eac861647726e141a628e64
workingDir: /home/runner/workspace
script: ls -lRA /home/runner
image: alpine:latest
onError: continue
```
Which seems like far far too many mounts
I am expecting to see
```yaml
- mountPath: /home/runner/sbin
name: sbin
readOnly: true
- mountPath: /home/runner
name: home
- mountPath: /home/runner/var/event.json
name: event
readOnly: true
subPath: event.json
- mountPath: $(params.home)/var/inputs
name: inputs-fooactionsfirstyaml2-78427d02ce7e28440b543047a1d49
readOnly: true
- mountPath: $(params.home)/var/state
name: state-fooactionsfirstyaml2s1-0a86de14517d7378aa4d46d20caa
```
# Additional Info
- Kubernetes version:
**Output of `kubectl version`:**
```
(Client Version: version.Info{Major:"1", Minor:"25", GitVersion:"v1.25.4", GitCommit:"872a965c6c6526caa949f0c6ac028ef7aff3fb78", GitTreeState:"clean", BuildDate:"2022-11-09T13:36:36Z", GoVersion:"go1.19.3", Compiler:"gc", Platform:"darwin/amd64"}
Kustomize Version: v4.5.7
Server Version: version.Info{Major:"1", Minor:"24", GitVersion:"v1.24.4+k3s1", GitCommit:"c3f830e9b9ed8a4d9d0e2aa663b4591b923a296e", GitTreeState:"clean", BuildDate:"2022-08-25T03:45:26Z", GoVersion:"go1.18.1", Compiler:"gc", Platform:"linux/amd64"}
``
- Tekton Pipeline version:
**Output of `tkn version` or `kubectl get pods -n tekton-pipelines -l app=tekton-pipelines-controller -o=jsonpath='{.items[0].metadata.labels.version}'`**
```
v0.45.0
```
<!-- Any other additional information -->
|
non_process
|
steps are getting volume mounts from other steps expected behavior i should get the volume mounts for the step that i specify actual behavior steps are getting the wrong volume mounts which are the same as a previous step i think the root cause might be here as i suspect that if you are in the case where toadd is empty the append may return the same slice from the iterator though i think that could just be a mis reading on my behalf steps to reproduce the problem working on trying to get the minimal steps i haven t finished cleaning this up and it may not be a functional test but it s as close to the yaml that is reproducing the issue for me as i can edit by hand right now i have a task that is something like this yaml apiversion tekton dev kind task metadata creationtimestamp null name test test spec description test params default workflow dispatch name eventname type string default home runner name home type string name needs build output foo type string name needs build result type string results name status type string steptemplate computeresources volumemounts mountpath params home sbin name sbin readonly true mountpath params home name home mountpath params home var event json name event readonly true subpath event json workingdir params home workspace steps script ls lra params home computeresources image alpine latest name onerror continue volumemounts mountpath params home var state name state script ls lra params home computeresources image alpine latest imagepullpolicy never name step run onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var with name inputs fooactionsfirstyaml workingdir params home workspace script ls lra params home computeresources image alpine latest name stepfooactionsfirstyaml onerror continue volumemounts mountpath params home var inputs name inputs fooactionsfirstyaml readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name stepfooactionsfirstyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var inputs name inputs fooactionsfirstyaml readonly true mountpath params home var with name inputs fooactionssecondyaml workingdir params home workspace script ls lra params home computeresources image alpine latest name stepfooactionssecondyaml onerror continue volumemounts mountpath params home var inputs name inputs fooactionssecondyaml readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name stepfooactionssecondyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var inputs name inputs fooactionssecondyaml readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest name stepfooactionsfirstyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var steps name state readonly true mountpath params home var inputs name inputs fooactionsfirstyaml readonly true workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name stepfooactionsfirstyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var steps name state readonly true mountpath params home var inputs name inputs fooactionsfirstyaml readonly true workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name step run onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var with name inputs workingdir params home workspace script ls lra params home computeresources image alpine latest name stepfooactionssecondyaml onerror continue volumemounts mountpath params home var inputs name inputs readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name stepfooactionssecondyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var inputs name inputs readonly true workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name step run onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var with name inputs workingdir params home workspace script ls lra params home computeresources image alpine latest name stepfooactionssecondyaml onerror continue volumemounts mountpath params home var inputs name inputs readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name stepfooactionssecondyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var inputs name inputs readonly true workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name step run onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var with name inputs workingdir params home workspace script ls lra params home computeresources image alpine latest name stepfooactionsfirstyaml onerror continue volumemounts mountpath params home var inputs name inputs readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name stepfooactionsfirstyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var inputs name inputs readonly true mountpath params home var with name inputs workingdir params home workspace script ls lra params home computeresources image alpine latest name stepfooactionssecondyaml onerror continue volumemounts mountpath params home var inputs name inputs readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name stepfooactionssecondyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var inputs name inputs readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest name stepfooactionsfirstyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var steps name state readonly true mountpath params home var inputs name inputs readonly true workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name stepfooactionsfirstyaml onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var inputs name inputs readonly true mountpath params home var state name state workingdir params home workspace script ls lra params home computeresources image alpine latest name step run onerror continue volumemounts mountpath params home var steps name state readonly true mountpath params home var steps name state readonly true workingdir params home workspace script ls lra params home computeresources image alpine latest imagepullpolicy never name step run volumemounts mountpath params home var steps name state readonly true mountpath params home var steps name state readonly true workingdir params home workspace volumes configmap name test event name event emptydir name home emptydir name inputs fooactionsfirstyaml emptydir name inputs emptydir name inputs fooactionssecondyaml emptydir name inputs emptydir name inputs emptydir name inputs emptydir name sbin emptydir name state emptydir name state emptydir name state emptydir name state emptydir name state emptydir name state emptydir name state emptydir name state emptydir name state emptydir name state and then when i look at stepfooactionsfirstyaml in tekton dashboard i will see on the details tab yaml name stepfooactionsfirstyaml resources volumemounts mountpath home runner sbin name sbin readonly true mountpath home runner name home mountpath home runner var event json name event readonly true subpath event json mountpath home runner var state name state mountpath home runner var steps name state readonly true mountpath home runner var inputs name inputs readonly true mountpath home runner var steps name state readonly true mountpath home runner var steps name state readonly true mountpath home runner var steps name state readonly true mountpath home runner var steps name state readonly true mountpath home runner var with name inputs workingdir home runner workspace script ls lra home runner image alpine latest onerror continue which seems like far far too many mounts i am expecting to see yaml mountpath home runner sbin name sbin readonly true mountpath home runner name home mountpath home runner var event json name event readonly true subpath event json mountpath params home var inputs name inputs readonly true mountpath params home var state name state additional info kubernetes version output of kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform darwin kustomize version server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux tekton pipeline version output of tkn version or kubectl get pods n tekton pipelines l app tekton pipelines controller o jsonpath items metadata labels version
| 0
|
20,287
| 10,699,230,784
|
IssuesEvent
|
2019-10-23 20:28:42
|
0xProject/OpenZKP
|
https://api.github.com/repos/0xProject/OpenZKP
|
opened
|
Remove allocation of vectors
|
performance tracker
|
*On 2019-07-08 @pvienhage wrote in [`6d56638`](https://github.com/0xProject/OpenZKP/commit/6d566382f82046e2f1e87a8cdc93b0d83ee8285c) “Paul/reorg/proof chunks (#24)”:*
Remove allocation of vectors
```rust
self.write(hash)
}
}
}
// OPT - Remove allocation of vectors
impl Writable<&[FieldElement]> for ProverChannel {
fn write(&mut self, data: &[FieldElement]) {
let mut container = Vec::with_capacity(32 * data.len());
for element in data {
for byte in &element.as_montgomery().to_bytes_be() {
```
*From [`crypto/stark/src/channel.rs:230`](https://github.com/0xProject/OpenZKP/blob/de90fc95d1e1c5fda722c0e904c21cd1fd6b9c40/crypto/stark/src/channel.rs#L230)*
<!--{"commit-hash": "6d566382f82046e2f1e87a8cdc93b0d83ee8285c", "author": "Paul Vienhage", "author-mail": "<paul@0x.org>", "author-time": 1562611977, "author-tz": "-0400", "committer": "GitHub", "committer-mail": "<noreply@github.com>", "committer-time": 1562611977, "committer-tz": "-0400", "summary": "Paul/reorg/proof chunks (#24)", "previous": "8e2e175577121babb838127b42e0480a641ac749 src/channel.rs", "filename": "crypto/stark/src/channel.rs", "line": 229, "line_end": 230, "kind": "OPT", "issue": "Remove allocation of vectors", "head": "Remove allocation of vectors", "context": " self.write(hash)\n }\n }\n}\n\n// OPT - Remove allocation of vectors\nimpl Writable<&[FieldElement]> for ProverChannel {\n fn write(&mut self, data: &[FieldElement]) {\n let mut container = Vec::with_capacity(32 * data.len());\n for element in data {\n for byte in &element.as_montgomery().to_bytes_be() {\n", "repo": "0xProject/OpenZKP", "branch-hash": "de90fc95d1e1c5fda722c0e904c21cd1fd6b9c40"}-->
|
True
|
Remove allocation of vectors - *On 2019-07-08 @pvienhage wrote in [`6d56638`](https://github.com/0xProject/OpenZKP/commit/6d566382f82046e2f1e87a8cdc93b0d83ee8285c) “Paul/reorg/proof chunks (#24)”:*
Remove allocation of vectors
```rust
self.write(hash)
}
}
}
// OPT - Remove allocation of vectors
impl Writable<&[FieldElement]> for ProverChannel {
fn write(&mut self, data: &[FieldElement]) {
let mut container = Vec::with_capacity(32 * data.len());
for element in data {
for byte in &element.as_montgomery().to_bytes_be() {
```
*From [`crypto/stark/src/channel.rs:230`](https://github.com/0xProject/OpenZKP/blob/de90fc95d1e1c5fda722c0e904c21cd1fd6b9c40/crypto/stark/src/channel.rs#L230)*
<!--{"commit-hash": "6d566382f82046e2f1e87a8cdc93b0d83ee8285c", "author": "Paul Vienhage", "author-mail": "<paul@0x.org>", "author-time": 1562611977, "author-tz": "-0400", "committer": "GitHub", "committer-mail": "<noreply@github.com>", "committer-time": 1562611977, "committer-tz": "-0400", "summary": "Paul/reorg/proof chunks (#24)", "previous": "8e2e175577121babb838127b42e0480a641ac749 src/channel.rs", "filename": "crypto/stark/src/channel.rs", "line": 229, "line_end": 230, "kind": "OPT", "issue": "Remove allocation of vectors", "head": "Remove allocation of vectors", "context": " self.write(hash)\n }\n }\n}\n\n// OPT - Remove allocation of vectors\nimpl Writable<&[FieldElement]> for ProverChannel {\n fn write(&mut self, data: &[FieldElement]) {\n let mut container = Vec::with_capacity(32 * data.len());\n for element in data {\n for byte in &element.as_montgomery().to_bytes_be() {\n", "repo": "0xProject/OpenZKP", "branch-hash": "de90fc95d1e1c5fda722c0e904c21cd1fd6b9c40"}-->
|
non_process
|
remove allocation of vectors on pvienhage wrote in “paul reorg proof chunks ” remove allocation of vectors rust self write hash opt remove allocation of vectors impl writable for proverchannel fn write mut self data let mut container vec with capacity data len for element in data for byte in element as montgomery to bytes be from author time author tz committer github committer mail committer time committer tz summary paul reorg proof chunks previous src channel rs filename crypto stark src channel rs line line end kind opt issue remove allocation of vectors head remove allocation of vectors context self write hash n n n n n opt remove allocation of vectors nimpl writable for proverchannel n fn write mut self data n let mut container vec with capacity data len n for element in data n for byte in element as montgomery to bytes be n repo openzkp branch hash
| 0
|
827,205
| 31,759,121,408
|
IssuesEvent
|
2023-09-12 02:40:49
|
cloudflare/cloudflared
|
https://api.github.com/repos/cloudflare/cloudflared
|
closed
|
🐛Tunnel with mTLS - using Chrome (Mobile/ Ubuntu) doesn't work
|
Type: Bug Priority: Normal
|
**Describe the bug**
Using a cloudflared tunnel with requirement valid certificate (mTLS), since 2 weeks, using Chrome v.116.0.5845.187 (Ubuntu) suddenly stopped working with certain ISP DNS servers.
If I use the ISP's (T-Mobile) DNS servers, the mTLS feature doesn't work. If I use Adguard private DNS, it works without problems.
How the error looks:
Either the choose certificate popup dialogue doesn't open or the choose certificate popup opens, but Cloudflare refuses connection with error: forbidden, 403, default cloudflare page.
Same error result in private mode. Fresh installation. No browser extensions. But using Firefox (Ubuntu) it works, without problems - regardless of the DNS server used.
**To Reproduce**
Steps to reproduce the behavior:
1. Configure cloudflared tunnel with require valid mTLS certificate to connect
2. Use Chrome (Ubuntu) or Chrome (Android) to open website.
3. See error
If it's an issue with Cloudflare Tunnel:
4. Tunnel ID : tba
5. cloudflared config: tba
**Expected behavior**
Open choose certificate dialogue popup on Chrome and connect to website. Like it used to work about 2 weeks ago.
**Environment and versions**
- OS: Ubuntu Mate 22.04 LTS / Android 13
- Architecture: Server ARM64 (not soc raspberry)
- Version: latest
**Logs and errors**
no error in logs
**Additional context**
It used to work without any issues for months, 2 weeks ago. But suddly stopped working. Only affecting Chrome browser.
|
1.0
|
🐛Tunnel with mTLS - using Chrome (Mobile/ Ubuntu) doesn't work - **Describe the bug**
Using a cloudflared tunnel with requirement valid certificate (mTLS), since 2 weeks, using Chrome v.116.0.5845.187 (Ubuntu) suddenly stopped working with certain ISP DNS servers.
If I use the ISP's (T-Mobile) DNS servers, the mTLS feature doesn't work. If I use Adguard private DNS, it works without problems.
How the error looks:
Either the choose certificate popup dialogue doesn't open or the choose certificate popup opens, but Cloudflare refuses connection with error: forbidden, 403, default cloudflare page.
Same error result in private mode. Fresh installation. No browser extensions. But using Firefox (Ubuntu) it works, without problems - regardless of the DNS server used.
**To Reproduce**
Steps to reproduce the behavior:
1. Configure cloudflared tunnel with require valid mTLS certificate to connect
2. Use Chrome (Ubuntu) or Chrome (Android) to open website.
3. See error
If it's an issue with Cloudflare Tunnel:
4. Tunnel ID : tba
5. cloudflared config: tba
**Expected behavior**
Open choose certificate dialogue popup on Chrome and connect to website. Like it used to work about 2 weeks ago.
**Environment and versions**
- OS: Ubuntu Mate 22.04 LTS / Android 13
- Architecture: Server ARM64 (not soc raspberry)
- Version: latest
**Logs and errors**
no error in logs
**Additional context**
It used to work without any issues for months, 2 weeks ago. But suddly stopped working. Only affecting Chrome browser.
|
non_process
|
🐛tunnel with mtls using chrome mobile ubuntu doesn t work describe the bug using a cloudflared tunnel with requirement valid certificate mtls since weeks using chrome v ubuntu suddenly stopped working with certain isp dns servers if i use the isp s t mobile dns servers the mtls feature doesn t work if i use adguard private dns it works without problems how the error looks either the choose certificate popup dialogue doesn t open or the choose certificate popup opens but cloudflare refuses connection with error forbidden default cloudflare page same error result in private mode fresh installation no browser extensions but using firefox ubuntu it works without problems regardless of the dns server used to reproduce steps to reproduce the behavior configure cloudflared tunnel with require valid mtls certificate to connect use chrome ubuntu or chrome android to open website see error if it s an issue with cloudflare tunnel tunnel id tba cloudflared config tba expected behavior open choose certificate dialogue popup on chrome and connect to website like it used to work about weeks ago environment and versions os ubuntu mate lts android architecture server not soc raspberry version latest logs and errors no error in logs additional context it used to work without any issues for months weeks ago but suddly stopped working only affecting chrome browser
| 0
|
14,760
| 10,211,531,673
|
IssuesEvent
|
2019-08-14 17:10:06
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Add instructions to debug on mac
|
Pri2 assigned-to-author doc-enhancement service-fabric/svc triaged
|
Hi,
It would be great to add instructions about how to debug SF on Mac. Since the service is running though docker it is not quite sure how it is done.
Kind Regards,
/ José.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7ea176d4-3324-727d-667a-570c7f829d53
* Version Independent ID: b86fc40d-7a5c-9ece-53c7-db8d491dc34a
* Content: [Develop .NET Core Azure Service Fabric applications with Visual Studio Code](https://docs.microsoft.com/en-us/azure/service-fabric/service-fabric-develop-csharp-applications-with-vs-code#feedback)
* Content Source: [articles/service-fabric/service-fabric-develop-csharp-applications-with-vs-code.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-fabric/service-fabric-develop-csharp-applications-with-vs-code.md)
* Service: **service-fabric**
* GitHub Login: @peterpogorski
* Microsoft Alias: **pepogors**
|
1.0
|
Add instructions to debug on mac - Hi,
It would be great to add instructions about how to debug SF on Mac. Since the service is running though docker it is not quite sure how it is done.
Kind Regards,
/ José.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7ea176d4-3324-727d-667a-570c7f829d53
* Version Independent ID: b86fc40d-7a5c-9ece-53c7-db8d491dc34a
* Content: [Develop .NET Core Azure Service Fabric applications with Visual Studio Code](https://docs.microsoft.com/en-us/azure/service-fabric/service-fabric-develop-csharp-applications-with-vs-code#feedback)
* Content Source: [articles/service-fabric/service-fabric-develop-csharp-applications-with-vs-code.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-fabric/service-fabric-develop-csharp-applications-with-vs-code.md)
* Service: **service-fabric**
* GitHub Login: @peterpogorski
* Microsoft Alias: **pepogors**
|
non_process
|
add instructions to debug on mac hi it would be great to add instructions about how to debug sf on mac since the service is running though docker it is not quite sure how it is done kind regards josé document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service service fabric github login peterpogorski microsoft alias pepogors
| 0
|
37,559
| 18,523,306,902
|
IssuesEvent
|
2021-10-20 17:19:48
|
keras-team/keras
|
https://api.github.com/repos/keras-team/keras
|
closed
|
mixed_precision returns gradient zeros when the model input size is large
|
type:bug/performance keras-team-review-pending
|
**This question is moved from Tensorflow git as instructed. please see [https://github.com/tensorflow/tensorflow/issues/52236](url) from previous discussions**
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): Noe
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 20.04
- TensorFlow installed from (source or binary): binary
- TensorFlow version (use command below): v2.6.0-rc2-32-g919f693420e 2.6.0
- Python version: 3.9
- CUDA/cuDNN version: 11.1/8.1.1
- GPU model and memory: RTX 3090/24GB
**Describe the current behavior**
When using mixed_precision policy described in [https://www.tensorflow.org/guide/mixed_precision](url) with large model input size, for example `(256, 368, 368,)`, the returned gradient are constantly ZEROS. However, if remove the mixed_precision policy, the returned gradient is normal with non-zeros numbers.
Moreover, if we use small model input size, let's say `(16, 16, 16)`, the returned gradient is normal no matter the mixed_precision is allowed or not.
My model is a typical U-net like model.
**Describe the expected behavior**
With large model input size like `(256, 368, 368)` used above, the returned gradient should be at least non-zeros. Otherwise the model won't be trained.
**Standalone code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary to generate
the problem. If possible, please share a link to Colab/Jupyter/any notebook.
The code below set the input shape to `[1, 256, 368, 368, 1]` and allows `mixed_precision.Policy('mixed_float16')`. It will return zeros gradients in the end (hence, no training at all). Setting `tf16_flag=False` will returns normal gradient behavior.
Also, by change `shape = [1, 16, 16, 16, 1]`, the gradient behaves normally no matter allows `mixed_precision.Policy('mixed_float16')` or not
```
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import layers
from tensorflow.keras import mixed_precision
import numpy as np
from tqdm import tqdm
gpus = tf.config.experimental.list_physical_devices('GPU')
for gpu in gpus:
tf.config.experimental.set_memory_growth(gpu, True)
tf16_flag = True
if tf16_flag:
policy = mixed_precision.Policy('mixed_float16')
mixed_precision.set_global_policy(policy)
shape = [1, 256, 368, 368, 1]
# shape = [1, 16, 16, 16, 1]
def forward_conv(x, filters, kernels, name='forward', padding='same'):
i = 0
for flt, kernel in zip(filters, kernels):
x = layers.Conv3D(flt, kernel, activation='relu', padding=padding, dilation_rate=(1, 1, 1),
use_bias=False, name=str(i) + '_' + name)(x)
x = layers.BatchNormalization(name=str(i) + '_bn_' + name)(x)
i += 1
return x
def part_one(ipt):
l1 = forward_conv(ipt, (4, 4), (3, 3), name='enc1')
d2 = layers.MaxPool3D(pool_size=(2, 2, 2))(l1)
l2 = forward_conv(d2, (4, 4), (3, 3), name='enc2')
return l1, l2
def part_inner(ipt1, ipt2):
l1 = forward_conv(ipt1, (4, 4), (3, 3), name='enc1')
l2 = forward_conv(ipt2, (4, 4), (3, 3), name='enc2')
return l1, l2
def part_two(ipt1, ipt2):
l2 = forward_conv(ipt2, (4, 4), (3, 3), name='dec2')
u1 = layers.UpSampling3D(size=(2, 2, 2))(l2)
r1 = forward_conv(ipt1 + u1, (4, 4), (3, 3), name='dec1')
return r1
initial = tf.ones(shape, dtype=tf.float16) if tf16_flag \
else tf.ones(shape, dtype=tf.float32)
tf.random.set_seed(1)
with tf.GradientTape() as g:
g.watch(initial)
l1_, l2_ = part_one(initial)
for _ in range(2):
l1_, l2_ = part_inner(l1_, l2_)
opt_ = part_two(l1_, l2_)
loss = tf.reduce_mean(l1_) + tf.reduce_mean(opt_)
gd = g.gradient(loss, initial)
print('-' * 100)
print(f'loss is {loss} and grad is {np.sum(gd)} with input shape {shape}')
```
|
True
|
mixed_precision returns gradient zeros when the model input size is large - **This question is moved from Tensorflow git as instructed. please see [https://github.com/tensorflow/tensorflow/issues/52236](url) from previous discussions**
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): Noe
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 20.04
- TensorFlow installed from (source or binary): binary
- TensorFlow version (use command below): v2.6.0-rc2-32-g919f693420e 2.6.0
- Python version: 3.9
- CUDA/cuDNN version: 11.1/8.1.1
- GPU model and memory: RTX 3090/24GB
**Describe the current behavior**
When using mixed_precision policy described in [https://www.tensorflow.org/guide/mixed_precision](url) with large model input size, for example `(256, 368, 368,)`, the returned gradient are constantly ZEROS. However, if remove the mixed_precision policy, the returned gradient is normal with non-zeros numbers.
Moreover, if we use small model input size, let's say `(16, 16, 16)`, the returned gradient is normal no matter the mixed_precision is allowed or not.
My model is a typical U-net like model.
**Describe the expected behavior**
With large model input size like `(256, 368, 368)` used above, the returned gradient should be at least non-zeros. Otherwise the model won't be trained.
**Standalone code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary to generate
the problem. If possible, please share a link to Colab/Jupyter/any notebook.
The code below set the input shape to `[1, 256, 368, 368, 1]` and allows `mixed_precision.Policy('mixed_float16')`. It will return zeros gradients in the end (hence, no training at all). Setting `tf16_flag=False` will returns normal gradient behavior.
Also, by change `shape = [1, 16, 16, 16, 1]`, the gradient behaves normally no matter allows `mixed_precision.Policy('mixed_float16')` or not
```
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import layers
from tensorflow.keras import mixed_precision
import numpy as np
from tqdm import tqdm
gpus = tf.config.experimental.list_physical_devices('GPU')
for gpu in gpus:
tf.config.experimental.set_memory_growth(gpu, True)
tf16_flag = True
if tf16_flag:
policy = mixed_precision.Policy('mixed_float16')
mixed_precision.set_global_policy(policy)
shape = [1, 256, 368, 368, 1]
# shape = [1, 16, 16, 16, 1]
def forward_conv(x, filters, kernels, name='forward', padding='same'):
i = 0
for flt, kernel in zip(filters, kernels):
x = layers.Conv3D(flt, kernel, activation='relu', padding=padding, dilation_rate=(1, 1, 1),
use_bias=False, name=str(i) + '_' + name)(x)
x = layers.BatchNormalization(name=str(i) + '_bn_' + name)(x)
i += 1
return x
def part_one(ipt):
l1 = forward_conv(ipt, (4, 4), (3, 3), name='enc1')
d2 = layers.MaxPool3D(pool_size=(2, 2, 2))(l1)
l2 = forward_conv(d2, (4, 4), (3, 3), name='enc2')
return l1, l2
def part_inner(ipt1, ipt2):
l1 = forward_conv(ipt1, (4, 4), (3, 3), name='enc1')
l2 = forward_conv(ipt2, (4, 4), (3, 3), name='enc2')
return l1, l2
def part_two(ipt1, ipt2):
l2 = forward_conv(ipt2, (4, 4), (3, 3), name='dec2')
u1 = layers.UpSampling3D(size=(2, 2, 2))(l2)
r1 = forward_conv(ipt1 + u1, (4, 4), (3, 3), name='dec1')
return r1
initial = tf.ones(shape, dtype=tf.float16) if tf16_flag \
else tf.ones(shape, dtype=tf.float32)
tf.random.set_seed(1)
with tf.GradientTape() as g:
g.watch(initial)
l1_, l2_ = part_one(initial)
for _ in range(2):
l1_, l2_ = part_inner(l1_, l2_)
opt_ = part_two(l1_, l2_)
loss = tf.reduce_mean(l1_) + tf.reduce_mean(opt_)
gd = g.gradient(loss, initial)
print('-' * 100)
print(f'loss is {loss} and grad is {np.sum(gd)} with input shape {shape}')
```
|
non_process
|
mixed precision returns gradient zeros when the model input size is large this question is moved from tensorflow git as instructed please see url from previous discussions system information have i written custom code as opposed to using a stock example script provided in tensorflow noe os platform and distribution e g linux ubuntu ubuntu tensorflow installed from source or binary binary tensorflow version use command below python version cuda cudnn version gpu model and memory rtx describe the current behavior when using mixed precision policy described in url with large model input size for example the returned gradient are constantly zeros however if remove the mixed precision policy the returned gradient is normal with non zeros numbers moreover if we use small model input size let s say the returned gradient is normal no matter the mixed precision is allowed or not my model is a typical u net like model describe the expected behavior with large model input size like used above the returned gradient should be at least non zeros otherwise the model won t be trained standalone code to reproduce the issue provide a reproducible test case that is the bare minimum necessary to generate the problem if possible please share a link to colab jupyter any notebook the code below set the input shape to and allows mixed precision policy mixed it will return zeros gradients in the end hence no training at all setting flag false will returns normal gradient behavior also by change shape the gradient behaves normally no matter allows mixed precision policy mixed or not import tensorflow as tf from tensorflow import keras from tensorflow keras import layers from tensorflow keras import mixed precision import numpy as np from tqdm import tqdm gpus tf config experimental list physical devices gpu for gpu in gpus tf config experimental set memory growth gpu true flag true if flag policy mixed precision policy mixed mixed precision set global policy policy shape shape def forward conv x filters kernels name forward padding same i for flt kernel in zip filters kernels x layers flt kernel activation relu padding padding dilation rate use bias false name str i name x x layers batchnormalization name str i bn name x i return x def part one ipt forward conv ipt name layers pool size forward conv name return def part inner forward conv name forward conv name return def part two forward conv name layers size forward conv name return initial tf ones shape dtype tf if flag else tf ones shape dtype tf tf random set seed with tf gradienttape as g g watch initial part one initial for in range part inner opt part two loss tf reduce mean tf reduce mean opt gd g gradient loss initial print print f loss is loss and grad is np sum gd with input shape shape
| 0
|
15,685
| 19,847,916,173
|
IssuesEvent
|
2022-01-21 09:01:12
|
ooi-data/CE01ISSM-RID16-07-NUTNRB000-recovered_host-nutnr_b_dcl_conc_instrument_recovered
|
https://api.github.com/repos/ooi-data/CE01ISSM-RID16-07-NUTNRB000-recovered_host-nutnr_b_dcl_conc_instrument_recovered
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:01:12.275767.
## Details
Flow name: `CE01ISSM-RID16-07-NUTNRB000-recovered_host-nutnr_b_dcl_conc_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:01:12.275767.
## Details
Flow name: `CE01ISSM-RID16-07-NUTNRB000-recovered_host-nutnr_b_dcl_conc_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host nutnr b dcl conc instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
537,638
| 15,732,315,156
|
IssuesEvent
|
2021-03-29 18:08:40
|
Malikil/PYOP-Checker
|
https://api.github.com/repos/Malikil/PYOP-Checker
|
opened
|
Display more pending maps when mods are specified
|
enhancement low priority
|
Right now the bot will only display maps of a certain mod up until the size limit for one field in an embed. That's fine when multiple mods are trying to be displayed and we don't want to hit the absolute cap. But when a single mod is specified and we probably won't be hitting the cap through other mods then extra fields can be added for the selected mod so more maps can be checked at once.
|
1.0
|
Display more pending maps when mods are specified - Right now the bot will only display maps of a certain mod up until the size limit for one field in an embed. That's fine when multiple mods are trying to be displayed and we don't want to hit the absolute cap. But when a single mod is specified and we probably won't be hitting the cap through other mods then extra fields can be added for the selected mod so more maps can be checked at once.
|
non_process
|
display more pending maps when mods are specified right now the bot will only display maps of a certain mod up until the size limit for one field in an embed that s fine when multiple mods are trying to be displayed and we don t want to hit the absolute cap but when a single mod is specified and we probably won t be hitting the cap through other mods then extra fields can be added for the selected mod so more maps can be checked at once
| 0
|
4,465
| 7,331,979,026
|
IssuesEvent
|
2018-03-05 15:06:34
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
URL to LUIS website is not displayed
|
assigned-to-author cognitive-services doc-bug in-process triaged
|
The markdown-like tags [LUIS][luis-reference-regions.md] displayed instead.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7449f6c8-ce19-3f32-d5de-e0308df266cf
* Version Independent ID: 6e26fc4a-270f-3eba-ee79-6ee7e7538f5b
* [Content](https://docs.microsoft.com/en-us/azure/cognitive-services/luis/home)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/LUIS/Home.md)
* Service: cognitive-services
|
1.0
|
URL to LUIS website is not displayed - The markdown-like tags [LUIS][luis-reference-regions.md] displayed instead.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7449f6c8-ce19-3f32-d5de-e0308df266cf
* Version Independent ID: 6e26fc4a-270f-3eba-ee79-6ee7e7538f5b
* [Content](https://docs.microsoft.com/en-us/azure/cognitive-services/luis/home)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/LUIS/Home.md)
* Service: cognitive-services
|
process
|
url to luis website is not displayed the markdown like tags displayed instead document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id service cognitive services
| 1
|
49,805
| 13,187,274,877
|
IssuesEvent
|
2020-08-13 02:53:51
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
lilliput - test_minuit2_migrad.py fails with ROOT 6's Minuit2 (Trac #2114)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2114">https://code.icecube.wisc.edu/ticket/2114</a>, reported by nega and owned by kkrings</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:13",
"description": "`test_minuit2_migrad.py` fails with ROOT 6's Minuit2\n\nsee the SL6 SPTS bot: http://builds.icecube.wisc.edu/builders/Scientific%20Linux%206%20%28SPTS%29/builds/24/steps/test/logs/stdio\n\nconfirmed locally on Ubuntu 14 with CVMFS::py2-v3\n\nNote: The SPTS bot is SL6 w/ a **local** copy of CVMFS::py2-v3 + extras",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1550067313248429",
"component": "combo reconstruction",
"summary": "lilliput - test_minuit2_migrad.py fails with ROOT 6's Minuit2",
"priority": "normal",
"keywords": "lilliput minuit2",
"time": "2017-11-16T21:47:24",
"milestone": "",
"owner": "kkrings",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
lilliput - test_minuit2_migrad.py fails with ROOT 6's Minuit2 (Trac #2114) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2114">https://code.icecube.wisc.edu/ticket/2114</a>, reported by nega and owned by kkrings</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:13",
"description": "`test_minuit2_migrad.py` fails with ROOT 6's Minuit2\n\nsee the SL6 SPTS bot: http://builds.icecube.wisc.edu/builders/Scientific%20Linux%206%20%28SPTS%29/builds/24/steps/test/logs/stdio\n\nconfirmed locally on Ubuntu 14 with CVMFS::py2-v3\n\nNote: The SPTS bot is SL6 w/ a **local** copy of CVMFS::py2-v3 + extras",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1550067313248429",
"component": "combo reconstruction",
"summary": "lilliput - test_minuit2_migrad.py fails with ROOT 6's Minuit2",
"priority": "normal",
"keywords": "lilliput minuit2",
"time": "2017-11-16T21:47:24",
"milestone": "",
"owner": "kkrings",
"type": "defect"
}
```
</p>
</details>
|
non_process
|
lilliput test migrad py fails with root s trac migrated from json status closed changetime description test migrad py fails with root s n nsee the spts bot locally on ubuntu with cvmfs n nnote the spts bot is w a local copy of cvmfs extras reporter nega cc resolution fixed ts component combo reconstruction summary lilliput test migrad py fails with root s priority normal keywords lilliput time milestone owner kkrings type defect
| 0
|
3,914
| 6,827,731,550
|
IssuesEvent
|
2017-11-08 17:59:29
|
trilinos/Trilinos
|
https://api.github.com/repos/trilinos/Trilinos
|
closed
|
Issue and Pull Request Templates?
|
Framework in progress process improvement question
|
@trilinos/framework, has Trilinos ever considered adding `ISSUE_TEMPLATE.md` and `PULL_REQUEST_TEMPLATE.md` files to the repository to attempt to standardize the information we get in an issue or pull request? If you like, I can mock some up (I have for GitLab, [here](https://gitlab.sandia.gov/jmgate/testingTemplates)), but it occurred to me that this may have come up before and was decided against for some reason.
|
1.0
|
Issue and Pull Request Templates? - @trilinos/framework, has Trilinos ever considered adding `ISSUE_TEMPLATE.md` and `PULL_REQUEST_TEMPLATE.md` files to the repository to attempt to standardize the information we get in an issue or pull request? If you like, I can mock some up (I have for GitLab, [here](https://gitlab.sandia.gov/jmgate/testingTemplates)), but it occurred to me that this may have come up before and was decided against for some reason.
|
process
|
issue and pull request templates trilinos framework has trilinos ever considered adding issue template md and pull request template md files to the repository to attempt to standardize the information we get in an issue or pull request if you like i can mock some up i have for gitlab but it occurred to me that this may have come up before and was decided against for some reason
| 1
|
260,129
| 22,594,271,501
|
IssuesEvent
|
2022-06-28 23:57:17
|
Azure/azure-sdk-for-js
|
https://api.github.com/repos/Azure/azure-sdk-for-js
|
closed
|
[Digital Twins - Core] Deploy test resources failing in nightly runs
|
Client Digital Twins test-reliability
|
Digital Twins - Core nightly test runs are failing with:
> Error message:
>BadRequest
Headers:
Cache-Control : no-cache
Pragma : no-cache
x-ms-failure-cause : gateway
x-ms-ratelimit-remaining-subscription-writes: 1197
x-ms-request-id : f2aa134a-babe-48fd-8806-aa0c67f1fdcf
x-ms-correlation-request-id : f2aa134a-babe-48fd-8806-aa0c67f1fdcf
x-ms-routing-request-id : WESTUS:20220623T092439Z:f2aa134a-babe-48fd-8806-aa0c67f1fdcf
Strict-Transport-Security : max-age=31536000; includeSubDomains
X-Content-Type-Options : nosniff
Date : Thu, 23 Jun 2022 09:24:38 GMT
Body:
{
"error": {
"code": "LocationNotAvailableForResourceType",
"message": "The provided location 'westus' is not available for resource type 'Microsoft.DigitalTwins/digitalTwinsInstances'. List of available regions for the resource type is 'westcentralus,westus2,northeurope,australiaeast,westeurope,eastus,southcentralus,southeastasia,uksouth,eastus2'."
}
}
For more details check here:
- https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1662353&view=results
|
1.0
|
[Digital Twins - Core] Deploy test resources failing in nightly runs - Digital Twins - Core nightly test runs are failing with:
> Error message:
>BadRequest
Headers:
Cache-Control : no-cache
Pragma : no-cache
x-ms-failure-cause : gateway
x-ms-ratelimit-remaining-subscription-writes: 1197
x-ms-request-id : f2aa134a-babe-48fd-8806-aa0c67f1fdcf
x-ms-correlation-request-id : f2aa134a-babe-48fd-8806-aa0c67f1fdcf
x-ms-routing-request-id : WESTUS:20220623T092439Z:f2aa134a-babe-48fd-8806-aa0c67f1fdcf
Strict-Transport-Security : max-age=31536000; includeSubDomains
X-Content-Type-Options : nosniff
Date : Thu, 23 Jun 2022 09:24:38 GMT
Body:
{
"error": {
"code": "LocationNotAvailableForResourceType",
"message": "The provided location 'westus' is not available for resource type 'Microsoft.DigitalTwins/digitalTwinsInstances'. List of available regions for the resource type is 'westcentralus,westus2,northeurope,australiaeast,westeurope,eastus,southcentralus,southeastasia,uksouth,eastus2'."
}
}
For more details check here:
- https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1662353&view=results
|
non_process
|
deploy test resources failing in nightly runs digital twins core nightly test runs are failing with error message badrequest headers cache control no cache pragma no cache x ms failure cause gateway x ms ratelimit remaining subscription writes x ms request id babe x ms correlation request id babe x ms routing request id westus babe strict transport security max age includesubdomains x content type options nosniff date thu jun gmt body error code locationnotavailableforresourcetype message the provided location westus is not available for resource type microsoft digitaltwins digitaltwinsinstances list of available regions for the resource type is westcentralus northeurope australiaeast westeurope eastus southcentralus southeastasia uksouth for more details check here
| 0
|
20,545
| 27,195,021,110
|
IssuesEvent
|
2023-02-20 03:56:47
|
cse442-at-ub/project_s23-team-infinity
|
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
|
opened
|
Create an alert for conflicting/overlapping events
|
Processing Task UI Task
|
**Testing:**
*Test 1*
1. Create an event on a specific date.
2. Create another event at that overlaps the previous event by a couple of minutes.
3. Ensure there is an alert notifying the user of overlapping events
*Test 2*
1. Create an event on a specific date.
2. Create another event on the same date at the same time.
3. Ensure there is an alert notifying the user of conflicting events.
|
1.0
|
Create an alert for conflicting/overlapping events - **Testing:**
*Test 1*
1. Create an event on a specific date.
2. Create another event at that overlaps the previous event by a couple of minutes.
3. Ensure there is an alert notifying the user of overlapping events
*Test 2*
1. Create an event on a specific date.
2. Create another event on the same date at the same time.
3. Ensure there is an alert notifying the user of conflicting events.
|
process
|
create an alert for conflicting overlapping events testing test create an event on a specific date create another event at that overlaps the previous event by a couple of minutes ensure there is an alert notifying the user of overlapping events test create an event on a specific date create another event on the same date at the same time ensure there is an alert notifying the user of conflicting events
| 1
|
43,182
| 9,383,971,125
|
IssuesEvent
|
2019-04-05 06:22:59
|
ThronesDB/thronesdb
|
https://api.github.com/repos/ThronesDB/thronesdb
|
closed
|
get rid of broken bulk upload for decks
|
code quality
|
this is just busted. i don't see the value in fixing this right now, or ever. yank it.
get rid of the "download all" decks action, it's not accessible via the UI anywhere.
|
1.0
|
get rid of broken bulk upload for decks - this is just busted. i don't see the value in fixing this right now, or ever. yank it.
get rid of the "download all" decks action, it's not accessible via the UI anywhere.
|
non_process
|
get rid of broken bulk upload for decks this is just busted i don t see the value in fixing this right now or ever yank it get rid of the download all decks action it s not accessible via the ui anywhere
| 0
|
19,674
| 2,622,159,126
|
IssuesEvent
|
2015-03-04 00:09:18
|
byzhang/terrastore
|
https://api.github.com/repos/byzhang/terrastore
|
closed
|
Drop support for array-based json documents.
|
auto-migrated Milestone-0.8.0 Priority-High Project-Terrastore Type-Enhancement
|
```
See mailing list threads:
http://groups.google.com/group/terrastore-discussions/t/c648539ec785e59f
http://groups.google.com/group/terrastore-discussions/t/e8e3b91d6ccfc4b
```
Original issue reported on code.google.com by `sergio.b...@gmail.com` on 1 Nov 2010 at 11:34
* Blocking: #119
|
1.0
|
Drop support for array-based json documents. - ```
See mailing list threads:
http://groups.google.com/group/terrastore-discussions/t/c648539ec785e59f
http://groups.google.com/group/terrastore-discussions/t/e8e3b91d6ccfc4b
```
Original issue reported on code.google.com by `sergio.b...@gmail.com` on 1 Nov 2010 at 11:34
* Blocking: #119
|
non_process
|
drop support for array based json documents see mailing list threads original issue reported on code google com by sergio b gmail com on nov at blocking
| 0
|
7,179
| 10,320,871,241
|
IssuesEvent
|
2019-08-30 22:07:17
|
NuGet/Home
|
https://api.github.com/repos/NuGet/Home
|
opened
|
Publish release notes for 5.3
|
Area:Release Process
|
Likely should ship release notes for 5.3 when preview3 ships.
Finalize soon.
|
1.0
|
Publish release notes for 5.3 - Likely should ship release notes for 5.3 when preview3 ships.
Finalize soon.
|
process
|
publish release notes for likely should ship release notes for when ships finalize soon
| 1
|
628,877
| 20,016,851,776
|
IssuesEvent
|
2022-02-01 12:58:27
|
ibissource/frank-flow
|
https://api.github.com/repos/ibissource/frank-flow
|
closed
|
Redesign the Frank!Flow (styling only)
|
feature priority:medium work:medium
|
**Is your feature request related to a problem? Please describe.**
The Frank!Flow looks flat and dated. It should look like a modern and simple tool: what it is.
**Describe the solution you'd like**
Change up the squares and add some shadows.
**Additional context**
The shadows shouldn't be taken too far. Which I always find hard. Either nothing gets shadow, or everything.
|
1.0
|
Redesign the Frank!Flow (styling only) - **Is your feature request related to a problem? Please describe.**
The Frank!Flow looks flat and dated. It should look like a modern and simple tool: what it is.
**Describe the solution you'd like**
Change up the squares and add some shadows.
**Additional context**
The shadows shouldn't be taken too far. Which I always find hard. Either nothing gets shadow, or everything.
|
non_process
|
redesign the frank flow styling only is your feature request related to a problem please describe the frank flow looks flat and dated it should look like a modern and simple tool what it is describe the solution you d like change up the squares and add some shadows additional context the shadows shouldn t be taken too far which i always find hard either nothing gets shadow or everything
| 0
|
292,774
| 8,967,879,518
|
IssuesEvent
|
2019-01-29 05:45:18
|
actiontech/dble
|
https://api.github.com/repos/actiontech/dble
|
closed
|
query with 2 "in subquery" with global table may get error "ERROR 4004 (42S22): column autoalias_tt.autoalias_scalar not found"
|
Priority/High complex-query resolve
|
* **dble version:**
5.6.29-dble-9.9.9.9-f2ee68f267caf5c53371cb68de2068998b363b72-20190128080033
* **preconditions :**
table tt is a table in schema's default node
* **configs:**
**schema.xml**
```
<table dataNode="dn1,dn2,dn3,dn4" name="test" type="global" />
```
**rule.xml**
```
```
**server.xml**
```
```
* **steps:**
step1. execute queries with dble:
```
mysql> create table tt(id int);
Query OK, 0 rows affected (0.01 sec)
mysql> insert into tt values(1),(2),(3),(4);
Query OK, 4 rows affected (0.00 sec)
mysql> create table test(id int, k int);
Query OK, 0 rows affected (0.08 sec)
mysql> insert into test values(1,1),(2,2),(3,3),(null,null);
Query OK, 4 rows affected (0.02 sec)
```
step2.execute queries with dble:
```
mysql> select * from test where k=1 and k in (select id from tt) and id in (select id from tt);
ERROR 4004 (42S22): column autoalias_tt.autoalias_scalar not found
```
* **expect result:**
1.step2 query success
* **real result:**
1.step2 query got error
* **supplements:**
1.sharding table has no such problem
/label ~BUG
|
1.0
|
query with 2 "in subquery" with global table may get error "ERROR 4004 (42S22): column autoalias_tt.autoalias_scalar not found" - * **dble version:**
5.6.29-dble-9.9.9.9-f2ee68f267caf5c53371cb68de2068998b363b72-20190128080033
* **preconditions :**
table tt is a table in schema's default node
* **configs:**
**schema.xml**
```
<table dataNode="dn1,dn2,dn3,dn4" name="test" type="global" />
```
**rule.xml**
```
```
**server.xml**
```
```
* **steps:**
step1. execute queries with dble:
```
mysql> create table tt(id int);
Query OK, 0 rows affected (0.01 sec)
mysql> insert into tt values(1),(2),(3),(4);
Query OK, 4 rows affected (0.00 sec)
mysql> create table test(id int, k int);
Query OK, 0 rows affected (0.08 sec)
mysql> insert into test values(1,1),(2,2),(3,3),(null,null);
Query OK, 4 rows affected (0.02 sec)
```
step2.execute queries with dble:
```
mysql> select * from test where k=1 and k in (select id from tt) and id in (select id from tt);
ERROR 4004 (42S22): column autoalias_tt.autoalias_scalar not found
```
* **expect result:**
1.step2 query success
* **real result:**
1.step2 query got error
* **supplements:**
1.sharding table has no such problem
/label ~BUG
|
non_process
|
query with in subquery with global table may get error error column autoalias tt autoalias scalar not found dble version: dble preconditions : table tt is a table in schema s default node configs: schema xml rule xml server xml steps: execute queries with dble mysql create table tt id int query ok rows affected sec mysql insert into tt values query ok rows affected sec mysql create table test id int k int query ok rows affected sec mysql insert into test values null null query ok rows affected sec execute queries with dble mysql select from test where k and k in select id from tt and id in select id from tt error column autoalias tt autoalias scalar not found expect result: query success real result: query got error supplements: sharding table has no such problem label bug
| 0
|
506,607
| 14,668,711,877
|
IssuesEvent
|
2020-12-29 22:03:52
|
bcgov/entity
|
https://api.github.com/repos/bcgov/entity
|
closed
|
Clicking on Cancel button throws the Paybc error.
|
ENTITY Name Request Priority1 bug
|
**Describe the bug in current situation**
When a user clicks on Cancel and Refund from the Paybc screen then the error dialog box seems to appear showing paybc error
**Impact of this bug**
All users clicking Cancel and Refund
Steps to reproduce the behavior:
1. Go to 'https://test.bcregistry.ca/namerequest/'
2. Click on 'Cancel and Return button'
3. See error
**Screenshots/ Visual Reference/ Source**

|
1.0
|
Clicking on Cancel button throws the Paybc error. - **Describe the bug in current situation**
When a user clicks on Cancel and Refund from the Paybc screen then the error dialog box seems to appear showing paybc error
**Impact of this bug**
All users clicking Cancel and Refund
Steps to reproduce the behavior:
1. Go to 'https://test.bcregistry.ca/namerequest/'
2. Click on 'Cancel and Return button'
3. See error
**Screenshots/ Visual Reference/ Source**

|
non_process
|
clicking on cancel button throws the paybc error describe the bug in current situation when a user clicks on cancel and refund from the paybc screen then the error dialog box seems to appear showing paybc error impact of this bug all users clicking cancel and refund steps to reproduce the behavior go to click on cancel and return button see error screenshots visual reference source
| 0
|
101,090
| 16,490,889,953
|
IssuesEvent
|
2021-05-25 03:32:16
|
uniquelyparticular/shipengine-request
|
https://api.github.com/repos/uniquelyparticular/shipengine-request
|
opened
|
CVE-2021-23383 (High) detected in handlebars-4.1.2.tgz
|
security vulnerability
|
## CVE-2021-23383 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.2.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz</a></p>
<p>Path to dependency file: shipengine-request/package.json</p>
<p>Path to vulnerable library: shipengine-request/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- semantic-release-15.14.0.tgz (Root Library)
- release-notes-generator-7.1.4.tgz
- conventional-changelog-writer-4.0.3.tgz
- :x: **handlebars-4.1.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package handlebars before 4.7.7 are vulnerable to Prototype Pollution when selecting certain compiling options to compile templates coming from an untrusted source.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23383>CVE-2021-23383</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: handlebars - v4.7.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23383 (High) detected in handlebars-4.1.2.tgz - ## CVE-2021-23383 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.2.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz</a></p>
<p>Path to dependency file: shipengine-request/package.json</p>
<p>Path to vulnerable library: shipengine-request/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- semantic-release-15.14.0.tgz (Root Library)
- release-notes-generator-7.1.4.tgz
- conventional-changelog-writer-4.0.3.tgz
- :x: **handlebars-4.1.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package handlebars before 4.7.7 are vulnerable to Prototype Pollution when selecting certain compiling options to compile templates coming from an untrusted source.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23383>CVE-2021-23383</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: handlebars - v4.7.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in handlebars tgz cve high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file shipengine request package json path to vulnerable library shipengine request node modules handlebars package json dependency hierarchy semantic release tgz root library release notes generator tgz conventional changelog writer tgz x handlebars tgz vulnerable library found in base branch develop vulnerability details the package handlebars before are vulnerable to prototype pollution when selecting certain compiling options to compile templates coming from an untrusted source publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.