Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
242,994
26,277,893,409
IssuesEvent
2023-01-07 01:26:03
turkdevops/snyk
https://api.github.com/repos/turkdevops/snyk
closed
CVE-2011-4969 (Low) detected in yiisoft/yii-1.1.14 - autoclosed
security vulnerability
## CVE-2011-4969 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yiisoft/yii-1.1.14</b></p></summary> <p>Yii Web Programming Framework</p> <p>Library home page: <a href="https://api.github.com/repos/yiisoft/yii/zipball/f0fee98ee84f70f1f3652f65562c9670e919cb4e">https://api.github.com/repos/yiisoft/yii/zipball/f0fee98ee84f70f1f3652f65562c9670e919cb4e</a></p> <p> Dependency Hierarchy: - :x: **yiisoft/yii-1.1.14** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/snyk/commit/9505f4ca92405cc9273dc3726c2d274ce28a4407">9505f4ca92405cc9273dc3726c2d274ce28a4407</a></p> <p>Found in base branch: <b>ALL_HANDS/major-secrets</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag. <p>Publish Date: 2013-03-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-4969>CVE-2011-4969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p> <p>Release Date: 2013-03-08</p> <p>Fix Resolution: 1.6.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2011-4969 (Low) detected in yiisoft/yii-1.1.14 - autoclosed - ## CVE-2011-4969 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yiisoft/yii-1.1.14</b></p></summary> <p>Yii Web Programming Framework</p> <p>Library home page: <a href="https://api.github.com/repos/yiisoft/yii/zipball/f0fee98ee84f70f1f3652f65562c9670e919cb4e">https://api.github.com/repos/yiisoft/yii/zipball/f0fee98ee84f70f1f3652f65562c9670e919cb4e</a></p> <p> Dependency Hierarchy: - :x: **yiisoft/yii-1.1.14** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/snyk/commit/9505f4ca92405cc9273dc3726c2d274ce28a4407">9505f4ca92405cc9273dc3726c2d274ce28a4407</a></p> <p>Found in base branch: <b>ALL_HANDS/major-secrets</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag. <p>Publish Date: 2013-03-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-4969>CVE-2011-4969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p> <p>Release Date: 2013-03-08</p> <p>Fix Resolution: 1.6.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve low detected in yiisoft yii autoclosed cve low severity vulnerability vulnerable library yiisoft yii yii web programming framework library home page a href dependency hierarchy x yiisoft yii vulnerable library found in head commit a href found in base branch all hands major secrets vulnerability details cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
105,051
22,832,341,291
IssuesEvent
2022-07-12 13:55:48
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
[Accessibility]: `N/A` hard to understand when read aloud
team/code-intelligence accessibility team/frontend-platform wcag/2.1/fixing wcag/2.1
### Audit type Screen reader navigation ### User journey audit issue #33516, #33518 ### Problem description <img width="884" alt="n:a" src="https://user-images.githubusercontent.com/103087/167504500-6c228e31-688b-4262-ba26-14d18129ffd1.png"> I couldn't understand the text for `N/A` here. ### Expected behavior Rename text or add an aria-label that can say `not applicable` or some other long-form version that can be understood. ### Additional details _No response_ ### Assigning labels - [X] Please give this issue an estimate by applying a label like `estimate/Xd`, where X is the estimated number of days it will take to complete. - [X] If this issue is specific to a specific Sourcegraph product, please assign the appropriate team label to this issue. - [X] If this issue will require input from designers in order to complete, please assign the label `needs-design`. - [X] If you are confident that this issue should be fixed by GitStart, please assign the label `gitstart`. ### Owner This issue will be fixed by my team, I have assigned a relevant member to it, or I will do so in the near future.
1.0
[Accessibility]: `N/A` hard to understand when read aloud - ### Audit type Screen reader navigation ### User journey audit issue #33516, #33518 ### Problem description <img width="884" alt="n:a" src="https://user-images.githubusercontent.com/103087/167504500-6c228e31-688b-4262-ba26-14d18129ffd1.png"> I couldn't understand the text for `N/A` here. ### Expected behavior Rename text or add an aria-label that can say `not applicable` or some other long-form version that can be understood. ### Additional details _No response_ ### Assigning labels - [X] Please give this issue an estimate by applying a label like `estimate/Xd`, where X is the estimated number of days it will take to complete. - [X] If this issue is specific to a specific Sourcegraph product, please assign the appropriate team label to this issue. - [X] If this issue will require input from designers in order to complete, please assign the label `needs-design`. - [X] If you are confident that this issue should be fixed by GitStart, please assign the label `gitstart`. ### Owner This issue will be fixed by my team, I have assigned a relevant member to it, or I will do so in the near future.
non_process
n a hard to understand when read aloud audit type screen reader navigation user journey audit issue problem description img width alt n a src i couldn t understand the text for n a here expected behavior rename text or add an aria label that can say not applicable or some other long form version that can be understood additional details no response assigning labels please give this issue an estimate by applying a label like estimate xd where x is the estimated number of days it will take to complete if this issue is specific to a specific sourcegraph product please assign the appropriate team label to this issue if this issue will require input from designers in order to complete please assign the label needs design if you are confident that this issue should be fixed by gitstart please assign the label gitstart owner this issue will be fixed by my team i have assigned a relevant member to it or i will do so in the near future
0
6,790
9,921,948,721
IssuesEvent
2019-06-30 22:59:12
GroceriStar/fetch-constants
https://api.github.com/repos/GroceriStar/fetch-constants
closed
#### [Recipe Search][URLS][part1]
in-process
https://chickenkyiv.github.io/search-api-documentation/docs/db-schema https://chickenkyiv.github.io/search-api-documentation/docs/db-schema By using names on URLS from this [page](https://chickenkyiv.github.io/search-api-documentation/docs/database-tables-models/attribute/allergy) In order to make it better, we'll create set of constants, each for a different method. Example: *http://localhost:3000/api/attribute?filter[where][type]=allergy* will became `export const ATTRIBUTE_FILTER_TYPE_ALLERGY = "ATTRIBUTE_FILTER_TYPE_ALLERGY";`
1.0
#### [Recipe Search][URLS][part1] - https://chickenkyiv.github.io/search-api-documentation/docs/db-schema https://chickenkyiv.github.io/search-api-documentation/docs/db-schema By using names on URLS from this [page](https://chickenkyiv.github.io/search-api-documentation/docs/database-tables-models/attribute/allergy) In order to make it better, we'll create set of constants, each for a different method. Example: *http://localhost:3000/api/attribute?filter[where][type]=allergy* will became `export const ATTRIBUTE_FILTER_TYPE_ALLERGY = "ATTRIBUTE_FILTER_TYPE_ALLERGY";`
process
by using names on urls from this in order to make it better we ll create set of constants each for a different method example allergy will became export const attribute filter type allergy attribute filter type allergy
1
4,300
7,194,970,240
IssuesEvent
2018-02-04 12:13:39
w3c/html
https://api.github.com/repos/w3c/html
closed
Add attribution of WHATWG HTML, as required by the CC-BY license
editorial process
WHATWG HTML now has the following copyright notice: > Copyright © 2018 WHATWG (Apple, Google, Mozilla, Microsoft). This work is licensed under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/). Unlike the previous license, this license explicitly requires attribution. Although previous copying can be interpreted as under the previous license, newer copying must follow the CC-BY 4.0 license and therefore requires attribution. The following commits appears to copy text from WHATWG HTML without attribution: https://github.com/w3c/html/commit/25aaa34bc4b8fbbad11a125d57a441f21256206d https://github.com/w3c/html/commit/6d380f2936c0b08ca4a2dc64e334c029780f1e2f Copying this text is allowed by the copyright license. However, the license requires proper attribution. Please add the attribution required by the new copyright license. Here is an example of an appropriate attribution CC-BY: > Portions derived from [HTML Living Standard](https://html.spec.whatwg.org), Copyright © 2018 WHATWG (Apple, Google, Mozilla, Microsoft) under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/).
1.0
Add attribution of WHATWG HTML, as required by the CC-BY license - WHATWG HTML now has the following copyright notice: > Copyright © 2018 WHATWG (Apple, Google, Mozilla, Microsoft). This work is licensed under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/). Unlike the previous license, this license explicitly requires attribution. Although previous copying can be interpreted as under the previous license, newer copying must follow the CC-BY 4.0 license and therefore requires attribution. The following commits appears to copy text from WHATWG HTML without attribution: https://github.com/w3c/html/commit/25aaa34bc4b8fbbad11a125d57a441f21256206d https://github.com/w3c/html/commit/6d380f2936c0b08ca4a2dc64e334c029780f1e2f Copying this text is allowed by the copyright license. However, the license requires proper attribution. Please add the attribution required by the new copyright license. Here is an example of an appropriate attribution CC-BY: > Portions derived from [HTML Living Standard](https://html.spec.whatwg.org), Copyright © 2018 WHATWG (Apple, Google, Mozilla, Microsoft) under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/).
process
add attribution of whatwg html as required by the cc by license whatwg html now has the following copyright notice copyright © whatwg apple google mozilla microsoft this work is licensed under a unlike the previous license this license explicitly requires attribution although previous copying can be interpreted as under the previous license newer copying must follow the cc by license and therefore requires attribution the following commits appears to copy text from whatwg html without attribution copying this text is allowed by the copyright license however the license requires proper attribution please add the attribution required by the new copyright license here is an example of an appropriate attribution cc by portions derived from copyright © whatwg apple google mozilla microsoft under a
1
3,523
6,564,753,246
IssuesEvent
2017-09-08 03:58:05
zero-os/0-Disk
https://api.github.com/repos/zero-os/0-Disk
closed
zeroctl rebalance command
process_wontfix type_feature
Part of the self-healing as described in https://github.com/zero-os/0-orchestrator/blob/master/specs/selfhealing/storage.md. TODO: + describe parameters and how the command would look like; + describe what the command would do and to which storage types it would apply (if specific at all);
1.0
zeroctl rebalance command - Part of the self-healing as described in https://github.com/zero-os/0-orchestrator/blob/master/specs/selfhealing/storage.md. TODO: + describe parameters and how the command would look like; + describe what the command would do and to which storage types it would apply (if specific at all);
process
zeroctl rebalance command part of the self healing as described in todo describe parameters and how the command would look like describe what the command would do and to which storage types it would apply if specific at all
1
6,044
8,854,377,592
IssuesEvent
2019-01-09 01:04:16
NottingHack/hms2
https://api.github.com/repos/NottingHack/hms2
opened
member.ex audit
Process
After 6 years we should remove ex member records but to keep DB integrity we likely obfuscate the data need to look carefully at what tables are have already that need cleaning and as we add more this audit will constantly need updating going to be best to break this down in `Entity` base files in say `app/HMS/Obfuscate/` so when adding a new entity we add a new `EntityObfuscate.php` file and the command runs uses auto discovery to find them all and call handle/run methods
1.0
member.ex audit - After 6 years we should remove ex member records but to keep DB integrity we likely obfuscate the data need to look carefully at what tables are have already that need cleaning and as we add more this audit will constantly need updating going to be best to break this down in `Entity` base files in say `app/HMS/Obfuscate/` so when adding a new entity we add a new `EntityObfuscate.php` file and the command runs uses auto discovery to find them all and call handle/run methods
process
member ex audit after years we should remove ex member records but to keep db integrity we likely obfuscate the data need to look carefully at what tables are have already that need cleaning and as we add more this audit will constantly need updating going to be best to break this down in entity base files in say app hms obfuscate so when adding a new entity we add a new entityobfuscate php file and the command runs uses auto discovery to find them all and call handle run methods
1
247,773
26,728,863,316
IssuesEvent
2023-01-30 01:13:34
Yash-Handa/GitHub-Org-Geographics
https://api.github.com/repos/Yash-Handa/GitHub-Org-Geographics
opened
CVE-2022-48285 (Medium) detected in jszip-3.2.1.tgz
security vulnerability
## CVE-2022-48285 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jszip-3.2.1.tgz</b></p></summary> <p>Create, read and edit .zip files with JavaScript http://stuartk.com/jszip</p> <p>Library home page: <a href="https://registry.npmjs.org/jszip/-/jszip-3.2.1.tgz">https://registry.npmjs.org/jszip/-/jszip-3.2.1.tgz</a></p> <p>Path to dependency file: /GitHub-Org-Geographics/package.json</p> <p>Path to vulnerable library: /node_modules/jszip/package.json</p> <p> Dependency Hierarchy: - protractor-5.4.2.tgz (Root Library) - selenium-webdriver-3.6.0.tgz - :x: **jszip-3.2.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> loadAsync in JSZip before 3.8.0 allows Directory Traversal via a crafted ZIP archive. <p>Publish Date: 2023-01-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-48285>CVE-2022-48285</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-48285">https://www.cve.org/CVERecord?id=CVE-2022-48285</a></p> <p>Release Date: 2023-01-29</p> <p>Fix Resolution (jszip): 3.8.0</p> <p>Direct dependency fix Resolution (protractor): 5.4.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-48285 (Medium) detected in jszip-3.2.1.tgz - ## CVE-2022-48285 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jszip-3.2.1.tgz</b></p></summary> <p>Create, read and edit .zip files with JavaScript http://stuartk.com/jszip</p> <p>Library home page: <a href="https://registry.npmjs.org/jszip/-/jszip-3.2.1.tgz">https://registry.npmjs.org/jszip/-/jszip-3.2.1.tgz</a></p> <p>Path to dependency file: /GitHub-Org-Geographics/package.json</p> <p>Path to vulnerable library: /node_modules/jszip/package.json</p> <p> Dependency Hierarchy: - protractor-5.4.2.tgz (Root Library) - selenium-webdriver-3.6.0.tgz - :x: **jszip-3.2.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> loadAsync in JSZip before 3.8.0 allows Directory Traversal via a crafted ZIP archive. <p>Publish Date: 2023-01-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-48285>CVE-2022-48285</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-48285">https://www.cve.org/CVERecord?id=CVE-2022-48285</a></p> <p>Release Date: 2023-01-29</p> <p>Fix Resolution (jszip): 3.8.0</p> <p>Direct dependency fix Resolution (protractor): 5.4.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jszip tgz cve medium severity vulnerability vulnerable library jszip tgz create read and edit zip files with javascript library home page a href path to dependency file github org geographics package json path to vulnerable library node modules jszip package json dependency hierarchy protractor tgz root library selenium webdriver tgz x jszip tgz vulnerable library found in base branch master vulnerability details loadasync in jszip before allows directory traversal via a crafted zip archive publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jszip direct dependency fix resolution protractor step up your open source security game with mend
0
40,398
6,822,351,597
IssuesEvent
2017-11-07 19:46:46
flyve-mdm/flyve-mdm-glpi-demo
https://api.github.com/repos/flyve-mdm/flyve-mdm-glpi-demo
closed
Move Wiki articles to project site
documentation
Hi, @Naylin15 Could you move this article to the Wiki in the project site? https://github.com/flyve-mdm/flyve-mdm-glpi-demo/wiki/Self_create_an_user_account Thank you.
1.0
Move Wiki articles to project site - Hi, @Naylin15 Could you move this article to the Wiki in the project site? https://github.com/flyve-mdm/flyve-mdm-glpi-demo/wiki/Self_create_an_user_account Thank you.
non_process
move wiki articles to project site hi could you move this article to the wiki in the project site thank you
0
9,172
12,225,607,296
IssuesEvent
2020-05-03 06:25:24
labnote-ant/labnote
https://api.github.com/repos/labnote-ant/labnote
closed
Make checkbox for conditions in process
process-view
Make checkbox for True/False conditions (ex. gradually, heating and etc.) in process.
1.0
Make checkbox for conditions in process - Make checkbox for True/False conditions (ex. gradually, heating and etc.) in process.
process
make checkbox for conditions in process make checkbox for true false conditions ex gradually heating and etc in process
1
10,700
13,495,112,155
IssuesEvent
2020-09-11 22:59:34
googleapis/java-secretmanager
https://api.github.com/repos/googleapis/java-secretmanager
closed
Promote to GA
api: secretmanager type: process
Package name: **google-cloud-secretmanager** Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] 28 days elapsed since last beta release with new API surface - [ ] Server API is GA - [ ] Package API is stable, and we can commit to backward compatibility - [ ] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
Promote to GA - Package name: **google-cloud-secretmanager** Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] 28 days elapsed since last beta release with new API surface - [ ] Server API is GA - [ ] Package API is stable, and we can commit to backward compatibility - [ ] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
promote to ga package name google cloud secretmanager current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
15,118
18,851,809,719
IssuesEvent
2021-11-11 22:00:34
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
TablesClient: make predict no longer call `get_model` and remove related arguments [breaking change]
type: process
Currently, for every [`predict`](https://github.com/googleapis/google-cloud-python/blob/master/automl/google/cloud/automl_v1beta1/tables/tables_client.py#L2593-L2603) call, two API requests are made: 1. [`get_model`](https://github.com/googleapis/google-cloud-python/blob/master/automl/google/cloud/automl_v1beta1/tables/tables_client.py#L221-L232) 2. `predict` `get_model` is costly, often 0.5s, but sometimes 10s in my tests (when the project contains huge number of models). Planning to never do `get_model` call in `predict`, and the user has to do a manual `get_model` call before making (repeated) predictions. So the cost is explicit and the user will never be caught off guard. This will be a breaking change, we will need to bump the major version. CC: @busunkim96 @sirtorry
1.0
TablesClient: make predict no longer call `get_model` and remove related arguments [breaking change] - Currently, for every [`predict`](https://github.com/googleapis/google-cloud-python/blob/master/automl/google/cloud/automl_v1beta1/tables/tables_client.py#L2593-L2603) call, two API requests are made: 1. [`get_model`](https://github.com/googleapis/google-cloud-python/blob/master/automl/google/cloud/automl_v1beta1/tables/tables_client.py#L221-L232) 2. `predict` `get_model` is costly, often 0.5s, but sometimes 10s in my tests (when the project contains huge number of models). Planning to never do `get_model` call in `predict`, and the user has to do a manual `get_model` call before making (repeated) predictions. So the cost is explicit and the user will never be caught off guard. This will be a breaking change, we will need to bump the major version. CC: @busunkim96 @sirtorry
process
tablesclient make predict no longer call get model and remove related arguments currently for every call two api requests are made predict get model is costly often but sometimes in my tests when the project contains huge number of models planning to never do get model call in predict and the user has to do a manual get model call before making repeated predictions so the cost is explicit and the user will never be caught off guard this will be a breaking change we will need to bump the major version cc sirtorry
1
21,308
28,502,253,305
IssuesEvent
2023-04-18 18:18:03
daviddrysdale/python-phonenumbers
https://api.github.com/repos/daviddrysdale/python-phonenumbers
closed
United States area code 557 is not working in python-phonenumbers
process
This is a python specific issue. The parent repository shows this as a valid phone number: Python Version: v3.10.8 Library Version: v8.13.9 Phone number area code not working: 557 Parent Repo Validation: ![image](https://user-images.githubusercontent.com/131063967/232620085-db566cb0-7cfa-497e-ac67-cd99cf1ea007.png)
1.0
United States area code 557 is not working in python-phonenumbers - This is a python specific issue. The parent repository shows this as a valid phone number: Python Version: v3.10.8 Library Version: v8.13.9 Phone number area code not working: 557 Parent Repo Validation: ![image](https://user-images.githubusercontent.com/131063967/232620085-db566cb0-7cfa-497e-ac67-cd99cf1ea007.png)
process
united states area code is not working in python phonenumbers this is a python specific issue the parent repository shows this as a valid phone number python version library version phone number area code not working parent repo validation
1
5,924
8,744,130,340
IssuesEvent
2018-12-12 21:16:43
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
ESS modifes built-in variables in ess-noweb-mode
process:eval
ESS modifes built-in variables in: https://github.com/emacs-ess/ESS/blob/d947e772639fb8b5957ea3afbbebff583e9a46cf/lisp/ess-noweb-mode.el#L411-L426 That can break other modes, see https://github.com/abo-abo/swiper/issues/1660
1.0
ESS modifes built-in variables in ess-noweb-mode - ESS modifes built-in variables in: https://github.com/emacs-ess/ESS/blob/d947e772639fb8b5957ea3afbbebff583e9a46cf/lisp/ess-noweb-mode.el#L411-L426 That can break other modes, see https://github.com/abo-abo/swiper/issues/1660
process
ess modifes built in variables in ess noweb mode ess modifes built in variables in that can break other modes see
1
435,193
30,491,596,754
IssuesEvent
2023-07-18 08:05:20
LiskHQ/app-registry
https://api.github.com/repos/LiskHQ/app-registry
closed
Add contributions information to the README
documentation type: improvement
### Description Currently, the README doesn't provide instructions for the users to submit a PR with the off-chain information for the sidechain applications. ### Acceptance Criteria - The contributions section in the README is updated with proper instructions for users to submit a PR ### Additional Information Ref: https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request-from-a-fork
1.0
Add contributions information to the README - ### Description Currently, the README doesn't provide instructions for the users to submit a PR with the off-chain information for the sidechain applications. ### Acceptance Criteria - The contributions section in the README is updated with proper instructions for users to submit a PR ### Additional Information Ref: https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request-from-a-fork
non_process
add contributions information to the readme description currently the readme doesn t provide instructions for the users to submit a pr with the off chain information for the sidechain applications acceptance criteria the contributions section in the readme is updated with proper instructions for users to submit a pr additional information ref
0
21,800
11,387,477,327
IssuesEvent
2020-01-29 15:05:21
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
CustomPaint does not respect isComplex/willChange if painters are null
customer: dream (g3) perf: speed severe: performance
# Background and reproduction One of our customer wants to use `CustomPaint` to trigger `RasterCache` by setting `isComplex = true, willChange = false` as below: ``` import 'package:flutter/material.dart'; class CostlyToRasterize extends StatelessWidget { @override Widget build(BuildContext context) => Container(color: Colors.blue); } void main() => runApp( MaterialApp( title: 'RootView', checkerboardRasterCacheImages: true, home: PageView.builder( itemBuilder: (BuildContext _, int index) { return CustomPaint( isComplex: true, willChange: false, child: CostlyToRasterize(), ); }, ), ), ); ``` (A more realistic app is available in [this gist](https://gist.github.com/liyuqian/8699bf9300ed2c749b4e69710ee43ef4).) However, the `isComplex` isn't sent to engine and the raster cache isn't triggered as expected. # Cause and solution The root cause is the following snippet of code (https://github.com/flutter/flutter/blob/master/packages/flutter/lib/src/rendering/custom_paint.dart#L571) ``` if (_painter != null) { _paintWithPainter(context.canvas, offset, _painter); _setRasterCacheHints(context); } ``` which ignores the raster cache hints if painter is null. To fix this, we could do either of the following: 1. Update our API doc to specify that raster cache hints are ignored if painters aren't provided. Hence the user needs to provide at least a dummy painter to trigger the raster cache hints. 2. Change our code to always `_setRasterCacheHints` no matter whether painter exists. CC @Hixie and @goderbauer for suggestions about which solution should be adopted. @ignatz: for the time being, you should be able to trigger the raster cache by providing a dummy painter like the following ``` import 'package:flutter/material.dart'; class CostlyToRasterize extends StatelessWidget { @override Widget build(BuildContext context) => Container(color: Colors.blue); } void main() => runApp( MaterialApp( title: 'RootView', checkerboardRasterCacheImages: true, home: PageView.builder( itemBuilder: (BuildContext _, int index) { return CustomPaint( painter: EmptyPainter(), isComplex: true, willChange: false, child: CostlyToRasterize(), ); }, ), ), ); class EmptyPainter extends CustomPainter { @override void paint(Canvas canvas, Size size) {} @override bool shouldRepaint(CustomPainter oldDelegate) => false; } ```
True
CustomPaint does not respect isComplex/willChange if painters are null - # Background and reproduction One of our customer wants to use `CustomPaint` to trigger `RasterCache` by setting `isComplex = true, willChange = false` as below: ``` import 'package:flutter/material.dart'; class CostlyToRasterize extends StatelessWidget { @override Widget build(BuildContext context) => Container(color: Colors.blue); } void main() => runApp( MaterialApp( title: 'RootView', checkerboardRasterCacheImages: true, home: PageView.builder( itemBuilder: (BuildContext _, int index) { return CustomPaint( isComplex: true, willChange: false, child: CostlyToRasterize(), ); }, ), ), ); ``` (A more realistic app is available in [this gist](https://gist.github.com/liyuqian/8699bf9300ed2c749b4e69710ee43ef4).) However, the `isComplex` isn't sent to engine and the raster cache isn't triggered as expected. # Cause and solution The root cause is the following snippet of code (https://github.com/flutter/flutter/blob/master/packages/flutter/lib/src/rendering/custom_paint.dart#L571) ``` if (_painter != null) { _paintWithPainter(context.canvas, offset, _painter); _setRasterCacheHints(context); } ``` which ignores the raster cache hints if painter is null. To fix this, we could do either of the following: 1. Update our API doc to specify that raster cache hints are ignored if painters aren't provided. Hence the user needs to provide at least a dummy painter to trigger the raster cache hints. 2. Change our code to always `_setRasterCacheHints` no matter whether painter exists. CC @Hixie and @goderbauer for suggestions about which solution should be adopted. @ignatz: for the time being, you should be able to trigger the raster cache by providing a dummy painter like the following ``` import 'package:flutter/material.dart'; class CostlyToRasterize extends StatelessWidget { @override Widget build(BuildContext context) => Container(color: Colors.blue); } void main() => runApp( MaterialApp( title: 'RootView', checkerboardRasterCacheImages: true, home: PageView.builder( itemBuilder: (BuildContext _, int index) { return CustomPaint( painter: EmptyPainter(), isComplex: true, willChange: false, child: CostlyToRasterize(), ); }, ), ), ); class EmptyPainter extends CustomPainter { @override void paint(Canvas canvas, Size size) {} @override bool shouldRepaint(CustomPainter oldDelegate) => false; } ```
non_process
custompaint does not respect iscomplex willchange if painters are null background and reproduction one of our customer wants to use custompaint to trigger rastercache by setting iscomplex true willchange false as below import package flutter material dart class costlytorasterize extends statelesswidget override widget build buildcontext context container color colors blue void main runapp materialapp title rootview checkerboardrastercacheimages true home pageview builder itembuilder buildcontext int index return custompaint iscomplex true willchange false child costlytorasterize a more realistic app is available in however the iscomplex isn t sent to engine and the raster cache isn t triggered as expected cause and solution the root cause is the following snippet of code if painter null paintwithpainter context canvas offset painter setrastercachehints context which ignores the raster cache hints if painter is null to fix this we could do either of the following update our api doc to specify that raster cache hints are ignored if painters aren t provided hence the user needs to provide at least a dummy painter to trigger the raster cache hints change our code to always setrastercachehints no matter whether painter exists cc hixie and goderbauer for suggestions about which solution should be adopted ignatz for the time being you should be able to trigger the raster cache by providing a dummy painter like the following import package flutter material dart class costlytorasterize extends statelesswidget override widget build buildcontext context container color colors blue void main runapp materialapp title rootview checkerboardrastercacheimages true home pageview builder itembuilder buildcontext int index return custompaint painter emptypainter iscomplex true willchange false child costlytorasterize class emptypainter extends custompainter override void paint canvas canvas size size override bool shouldrepaint custompainter olddelegate false
0
12,298
14,854,473,458
IssuesEvent
2021-01-18 11:21:51
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Error messages should be updated as per document
Bug P1 Participant manager Process: Fixed Process: Tested dev
Error messages should be updated as per following document https://docs.google.com/spreadsheets/d/1CEcgbG2Et3FCc5qbavewlNQapxNHG5kTSNMFYzdz4nc/edit?usp=sharing
2.0
Error messages should be updated as per document - Error messages should be updated as per following document https://docs.google.com/spreadsheets/d/1CEcgbG2Et3FCc5qbavewlNQapxNHG5kTSNMFYzdz4nc/edit?usp=sharing
process
error messages should be updated as per document error messages should be updated as per following document
1
27,062
4,867,181,025
IssuesEvent
2016-11-15 03:03:29
TNGSB/eWallet
https://api.github.com/repos/TNGSB/eWallet
closed
e-Wallet_WebAdmin 07112016
Defect - High (Sev-2)
As for today testing, there are 3 defects found in web admin module. -Card Management -Customer Service Kindly refer the attached defects document for your perusal. [Sev_2.zip](https://github.com/TNGSB/eWallet/files/574946/Sev_2.zip)
1.0
e-Wallet_WebAdmin 07112016 - As for today testing, there are 3 defects found in web admin module. -Card Management -Customer Service Kindly refer the attached defects document for your perusal. [Sev_2.zip](https://github.com/TNGSB/eWallet/files/574946/Sev_2.zip)
non_process
e wallet webadmin as for today testing there are defects found in web admin module card management customer service kindly refer the attached defects document for your perusal
0
21,947
30,448,636,913
IssuesEvent
2023-07-16 02:00:07
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Fri, 14 Jul 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Temporal Label-Refinement for Weakly-Supervised Audio-Visual Event Localization - **Authors:** Kalyan Ramakrishnan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2307.06385 - **Pdf link:** https://arxiv.org/pdf/2307.06385 - **Abstract** Audio-Visual Event Localization (AVEL) is the task of temporally localizing and classifying \emph{audio-visual events}, i.e., events simultaneously visible and audible in a video. In this paper, we solve AVEL in a weakly-supervised setting, where only video-level event labels (their presence/absence, but not their locations in time) are available as supervision for training. Our idea is to use a base model to estimate labels on the training data at a finer temporal resolution than at the video level and re-train the model with these labels. I.e., we determine the subset of labels for each \emph{slice} of frames in a training video by (i) replacing the frames outside the slice with those from a second video having no overlap in video-level labels, and (ii) feeding this synthetic video into the base model to extract labels for just the slice in question. To handle the out-of-distribution nature of our synthetic videos, we propose an auxiliary objective for the base model that induces more reliable predictions of the localized event labels as desired. Our three-stage pipeline outperforms several existing AVEL methods with no architectural changes and improves performance on a related weakly-supervised task as well. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### ConvNeXt-ChARM: ConvNeXt-based Transform for Efficient Neural Image Compression - **Authors:** Ahmed Ghorbel, Wassim Hamidouche, Luce Morin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.06342 - **Pdf link:** https://arxiv.org/pdf/2307.06342 - **Abstract** Over the last few years, neural image compression has gained wide attention from research and industry, yielding promising end-to-end deep neural codecs outperforming their conventional counterparts in rate-distortion performance. Despite significant advancement, current methods, including attention-based transform coding, still need to be improved in reducing the coding rate while preserving the reconstruction fidelity, especially in non-homogeneous textured image areas. Those models also require more parameters and a higher decoding time. To tackle the above challenges, we propose ConvNeXt-ChARM, an efficient ConvNeXt-based transform coding framework, paired with a compute-efficient channel-wise auto-regressive prior to capturing both global and local contexts from the hyper and quantized latent representations. The proposed architecture can be optimized end-to-end to fully exploit the context information and extract compact latent representation while reconstructing higher-quality images. Experimental results on four widely-used datasets showed that ConvNeXt-ChARM brings consistent and significant BD-rate (PSNR) reductions estimated on average to 5.24% and 1.22% over the versatile video coding (VVC) reference encoder (VTM-18.0) and the state-of-the-art learned image compression method SwinT-ChARM, respectively. Moreover, we provide model scaling studies to verify the computational efficiency of our approach and conduct several objective and subjective analyses to bring to the fore the performance gap between the next generation ConvNet, namely ConvNeXt, and Swin Transformer. ## Keyword: RAW ### Image Transformation Sequence Retrieval with General Reinforcement Learning - **Authors:** Enrique Mas-Candela, Antonio Ríos-Vila, Jorge Calvo-Zaragoza - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2307.06630 - **Pdf link:** https://arxiv.org/pdf/2307.06630 - **Abstract** In this work, the novel Image Transformation Sequence Retrieval (ITSR) task is presented, in which a model must retrieve the sequence of transformations between two given images that act as source and target, respectively. Given certain characteristics of the challenge such as the multiplicity of a correct sequence or the correlation between consecutive steps of the process, we propose a solution to ITSR using a general model-based Reinforcement Learning such as Monte Carlo Tree Search (MCTS), which is combined with a deep neural network. Our experiments provide a benchmark in both synthetic and real domains, where the proposed approach is compared with supervised training. The results report that a model trained with MCTS is able to outperform its supervised counterpart in both the simplest and the most complex cases. Our work draws interesting conclusions about the nature of ITSR and its associated challenges. ## Keyword: raw image There is no result
2.0
New submissions for Fri, 14 Jul 23 - ## Keyword: events ### Temporal Label-Refinement for Weakly-Supervised Audio-Visual Event Localization - **Authors:** Kalyan Ramakrishnan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2307.06385 - **Pdf link:** https://arxiv.org/pdf/2307.06385 - **Abstract** Audio-Visual Event Localization (AVEL) is the task of temporally localizing and classifying \emph{audio-visual events}, i.e., events simultaneously visible and audible in a video. In this paper, we solve AVEL in a weakly-supervised setting, where only video-level event labels (their presence/absence, but not their locations in time) are available as supervision for training. Our idea is to use a base model to estimate labels on the training data at a finer temporal resolution than at the video level and re-train the model with these labels. I.e., we determine the subset of labels for each \emph{slice} of frames in a training video by (i) replacing the frames outside the slice with those from a second video having no overlap in video-level labels, and (ii) feeding this synthetic video into the base model to extract labels for just the slice in question. To handle the out-of-distribution nature of our synthetic videos, we propose an auxiliary objective for the base model that induces more reliable predictions of the localized event labels as desired. Our three-stage pipeline outperforms several existing AVEL methods with no architectural changes and improves performance on a related weakly-supervised task as well. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### ConvNeXt-ChARM: ConvNeXt-based Transform for Efficient Neural Image Compression - **Authors:** Ahmed Ghorbel, Wassim Hamidouche, Luce Morin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.06342 - **Pdf link:** https://arxiv.org/pdf/2307.06342 - **Abstract** Over the last few years, neural image compression has gained wide attention from research and industry, yielding promising end-to-end deep neural codecs outperforming their conventional counterparts in rate-distortion performance. Despite significant advancement, current methods, including attention-based transform coding, still need to be improved in reducing the coding rate while preserving the reconstruction fidelity, especially in non-homogeneous textured image areas. Those models also require more parameters and a higher decoding time. To tackle the above challenges, we propose ConvNeXt-ChARM, an efficient ConvNeXt-based transform coding framework, paired with a compute-efficient channel-wise auto-regressive prior to capturing both global and local contexts from the hyper and quantized latent representations. The proposed architecture can be optimized end-to-end to fully exploit the context information and extract compact latent representation while reconstructing higher-quality images. Experimental results on four widely-used datasets showed that ConvNeXt-ChARM brings consistent and significant BD-rate (PSNR) reductions estimated on average to 5.24% and 1.22% over the versatile video coding (VVC) reference encoder (VTM-18.0) and the state-of-the-art learned image compression method SwinT-ChARM, respectively. Moreover, we provide model scaling studies to verify the computational efficiency of our approach and conduct several objective and subjective analyses to bring to the fore the performance gap between the next generation ConvNet, namely ConvNeXt, and Swin Transformer. ## Keyword: RAW ### Image Transformation Sequence Retrieval with General Reinforcement Learning - **Authors:** Enrique Mas-Candela, Antonio Ríos-Vila, Jorge Calvo-Zaragoza - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2307.06630 - **Pdf link:** https://arxiv.org/pdf/2307.06630 - **Abstract** In this work, the novel Image Transformation Sequence Retrieval (ITSR) task is presented, in which a model must retrieve the sequence of transformations between two given images that act as source and target, respectively. Given certain characteristics of the challenge such as the multiplicity of a correct sequence or the correlation between consecutive steps of the process, we propose a solution to ITSR using a general model-based Reinforcement Learning such as Monte Carlo Tree Search (MCTS), which is combined with a deep neural network. Our experiments provide a benchmark in both synthetic and real domains, where the proposed approach is compared with supervised training. The results report that a model trained with MCTS is able to outperform its supervised counterpart in both the simplest and the most complex cases. Our work draws interesting conclusions about the nature of ITSR and its associated challenges. ## Keyword: raw image There is no result
process
new submissions for fri jul keyword events temporal label refinement for weakly supervised audio visual event localization authors kalyan ramakrishnan subjects computer vision and pattern recognition cs cv machine learning cs lg sound cs sd audio and speech processing eess as arxiv link pdf link abstract audio visual event localization avel is the task of temporally localizing and classifying emph audio visual events i e events simultaneously visible and audible in a video in this paper we solve avel in a weakly supervised setting where only video level event labels their presence absence but not their locations in time are available as supervision for training our idea is to use a base model to estimate labels on the training data at a finer temporal resolution than at the video level and re train the model with these labels i e we determine the subset of labels for each emph slice of frames in a training video by i replacing the frames outside the slice with those from a second video having no overlap in video level labels and ii feeding this synthetic video into the base model to extract labels for just the slice in question to handle the out of distribution nature of our synthetic videos we propose an auxiliary objective for the base model that induces more reliable predictions of the localized event labels as desired our three stage pipeline outperforms several existing avel methods with no architectural changes and improves performance on a related weakly supervised task as well keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression convnext charm convnext based transform for efficient neural image compression authors ahmed ghorbel wassim hamidouche luce morin subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract over the last few years neural image compression has gained wide attention from research and industry yielding promising end to end deep neural codecs outperforming their conventional counterparts in rate distortion performance despite significant advancement current methods including attention based transform coding still need to be improved in reducing the coding rate while preserving the reconstruction fidelity especially in non homogeneous textured image areas those models also require more parameters and a higher decoding time to tackle the above challenges we propose convnext charm an efficient convnext based transform coding framework paired with a compute efficient channel wise auto regressive prior to capturing both global and local contexts from the hyper and quantized latent representations the proposed architecture can be optimized end to end to fully exploit the context information and extract compact latent representation while reconstructing higher quality images experimental results on four widely used datasets showed that convnext charm brings consistent and significant bd rate psnr reductions estimated on average to and over the versatile video coding vvc reference encoder vtm and the state of the art learned image compression method swint charm respectively moreover we provide model scaling studies to verify the computational efficiency of our approach and conduct several objective and subjective analyses to bring to the fore the performance gap between the next generation convnet namely convnext and swin transformer keyword raw image transformation sequence retrieval with general reinforcement learning authors enrique mas candela antonio ríos vila jorge calvo zaragoza subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract in this work the novel image transformation sequence retrieval itsr task is presented in which a model must retrieve the sequence of transformations between two given images that act as source and target respectively given certain characteristics of the challenge such as the multiplicity of a correct sequence or the correlation between consecutive steps of the process we propose a solution to itsr using a general model based reinforcement learning such as monte carlo tree search mcts which is combined with a deep neural network our experiments provide a benchmark in both synthetic and real domains where the proposed approach is compared with supervised training the results report that a model trained with mcts is able to outperform its supervised counterpart in both the simplest and the most complex cases our work draws interesting conclusions about the nature of itsr and its associated challenges keyword raw image there is no result
1
8,892
11,986,967,840
IssuesEvent
2020-04-07 20:18:32
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
internal/gapicgen: Go-based binary dependencies are using master instead of latest release
type: process
We've noticed that the `internal/gapicgen` system is pulling `master` of Go-based binary dependencies i.e. `protoc-gen-go_gapic` and `goimports`, rather than the version specified in `internal/gapicgen/go.mod`. We need to ensure that the pinned versions are respected by the generation tooling so that we don't introduce unintended changes.
1.0
internal/gapicgen: Go-based binary dependencies are using master instead of latest release - We've noticed that the `internal/gapicgen` system is pulling `master` of Go-based binary dependencies i.e. `protoc-gen-go_gapic` and `goimports`, rather than the version specified in `internal/gapicgen/go.mod`. We need to ensure that the pinned versions are respected by the generation tooling so that we don't introduce unintended changes.
process
internal gapicgen go based binary dependencies are using master instead of latest release we ve noticed that the internal gapicgen system is pulling master of go based binary dependencies i e protoc gen go gapic and goimports rather than the version specified in internal gapicgen go mod we need to ensure that the pinned versions are respected by the generation tooling so that we don t introduce unintended changes
1
40,120
6,800,157,238
IssuesEvent
2017-11-02 13:04:50
onury/geolocator
https://api.github.com/repos/onury/geolocator
closed
Demo pages not using HTTPS
documentation
Hello, As stated in your own README: > Make sure you're calling Geolocation APIs (such as geolocator.locate() and geolocator.watch()) from a secure origin (i.e. an HTTPS page). In Chrome 50, Geolocation API is removed from unsecured origins. Other browsers are expected to follow. Since the demo is not using HTTPS, the demo doesn't ask for location in Google Chrome. Please put the demo pages under HTTPS, its free using Let's Encrypt now!
1.0
Demo pages not using HTTPS - Hello, As stated in your own README: > Make sure you're calling Geolocation APIs (such as geolocator.locate() and geolocator.watch()) from a secure origin (i.e. an HTTPS page). In Chrome 50, Geolocation API is removed from unsecured origins. Other browsers are expected to follow. Since the demo is not using HTTPS, the demo doesn't ask for location in Google Chrome. Please put the demo pages under HTTPS, its free using Let's Encrypt now!
non_process
demo pages not using https hello as stated in your own readme make sure you re calling geolocation apis such as geolocator locate and geolocator watch from a secure origin i e an https page in chrome geolocation api is removed from unsecured origins other browsers are expected to follow since the demo is not using https the demo doesn t ask for location in google chrome please put the demo pages under https its free using let s encrypt now
0
12,482
14,950,090,914
IssuesEvent
2021-01-26 12:33:46
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
opened
Expand AWS ALB log parser to include Desync mitigation mode fields
data inputs story team:data processing
### Description [AWS ALB access logs](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/load-balancer-access-logs.html#access-log-file-format) include 4 additional fields that the current [log parser](https://github.com/panther-labs/panther/blob/2a27d25374cde5ab88d67142d3ebc42568d2dd94/internal/log_analysis/log_processor/parsers/awslogs/alb.go#L37) seems to ignore: - `"target:port_list"` - `"target_status_code_list"` - `"classification"` - `"classification_reason"` The last two fields seem particularly important in the security context since they are specified by the [Desync mitigation mode](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/application-load-balancers.html#desync-mitigation-mode) protection. > Desync mitigation mode protects your application from issues due to HTTP Desync. The load balancer classifies each request based on its threat level, allows safe requests, and then mitigates risk as specified by the mitigation mode that you specify. The desync mitigation modes are monitor, defensive, and strictest. ### Additional References - https://github.com/aws/http-desync-guardian - https://aws.amazon.com/about-aws/whats-new/2020/08/application-and-classic-load-balancers-adding-defense-in-depth-with-introduction-of-desync-mitigation-mode/ ### Related Services - Log Processing / ALB Log Parser - AWS Glue metadata for ALB access log table ### Acceptance Criteria - Parse AWS ALB log records with the 4 additional fields. - Output the additional fields accordingly and make them available for rules.
1.0
Expand AWS ALB log parser to include Desync mitigation mode fields - ### Description [AWS ALB access logs](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/load-balancer-access-logs.html#access-log-file-format) include 4 additional fields that the current [log parser](https://github.com/panther-labs/panther/blob/2a27d25374cde5ab88d67142d3ebc42568d2dd94/internal/log_analysis/log_processor/parsers/awslogs/alb.go#L37) seems to ignore: - `"target:port_list"` - `"target_status_code_list"` - `"classification"` - `"classification_reason"` The last two fields seem particularly important in the security context since they are specified by the [Desync mitigation mode](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/application-load-balancers.html#desync-mitigation-mode) protection. > Desync mitigation mode protects your application from issues due to HTTP Desync. The load balancer classifies each request based on its threat level, allows safe requests, and then mitigates risk as specified by the mitigation mode that you specify. The desync mitigation modes are monitor, defensive, and strictest. ### Additional References - https://github.com/aws/http-desync-guardian - https://aws.amazon.com/about-aws/whats-new/2020/08/application-and-classic-load-balancers-adding-defense-in-depth-with-introduction-of-desync-mitigation-mode/ ### Related Services - Log Processing / ALB Log Parser - AWS Glue metadata for ALB access log table ### Acceptance Criteria - Parse AWS ALB log records with the 4 additional fields. - Output the additional fields accordingly and make them available for rules.
process
expand aws alb log parser to include desync mitigation mode fields description include additional fields that the current seems to ignore target port list target status code list classification classification reason the last two fields seem particularly important in the security context since they are specified by the protection desync mitigation mode protects your application from issues due to http desync the load balancer classifies each request based on its threat level allows safe requests and then mitigates risk as specified by the mitigation mode that you specify the desync mitigation modes are monitor defensive and strictest additional references related services log processing alb log parser aws glue metadata for alb access log table acceptance criteria parse aws alb log records with the additional fields output the additional fields accordingly and make them available for rules
1
1,193
3,690,399,345
IssuesEvent
2016-02-25 19:51:34
hoodiehq/editorial
https://api.github.com/repos/hoodiehq/editorial
closed
Define a formal method in which contributors are added to the Editorial team
process
I updated the `team-roles.md` file with descriptions in https://github.com/hoodiehq/editorial/pull/23. In addition to adding the descriptions, I think we need to define a formal process in which contributors are added to the Editorials team. Having this early on makes the transition from one (or few) contributor to many slide through with much less friction than when it is developed later. Tasks for getting this done: - [x] Separate file or put in `team-roles.md`? - [x] Different path to contributor for each of the defined roles in `team-roles.md` or a unified path? - [x] ~~Define what `active` and `contribution` mean.~~ - [ ] Link to CoC, `CONTRIBUTING.md`, and anything else? - [ ] {{other points?}}
1.0
Define a formal method in which contributors are added to the Editorial team - I updated the `team-roles.md` file with descriptions in https://github.com/hoodiehq/editorial/pull/23. In addition to adding the descriptions, I think we need to define a formal process in which contributors are added to the Editorials team. Having this early on makes the transition from one (or few) contributor to many slide through with much less friction than when it is developed later. Tasks for getting this done: - [x] Separate file or put in `team-roles.md`? - [x] Different path to contributor for each of the defined roles in `team-roles.md` or a unified path? - [x] ~~Define what `active` and `contribution` mean.~~ - [ ] Link to CoC, `CONTRIBUTING.md`, and anything else? - [ ] {{other points?}}
process
define a formal method in which contributors are added to the editorial team i updated the team roles md file with descriptions in in addition to adding the descriptions i think we need to define a formal process in which contributors are added to the editorials team having this early on makes the transition from one or few contributor to many slide through with much less friction than when it is developed later tasks for getting this done separate file or put in team roles md different path to contributor for each of the defined roles in team roles md or a unified path define what active and contribution mean link to coc contributing md and anything else other points
1
3,559
3,955,259,057
IssuesEvent
2016-04-29 20:12:21
commercialhaskell/stack
https://api.github.com/repos/commercialhaskell/stack
closed
Avoid unnecessarily loading the hackage index
awaiting pr help wanted type: performance
Loading the hackage is expensive, because the encoded index cache is 18Mb. On my fairly fast computer, it takes 0.3 seconds to decode it. Search `getPackageCaches` to find the spots it's used. There are two cases where loading the hackage currently happens unnecessarily: 1) In `loadSourceMap`, because the user might have specified a package target on the commandline which is not in the resolver. In this case, it uses the latest from hackage. 2) In `constructPlan`, because plan construction errors get annotated with "latest applicable" versions. For both of these, loading the hackage index is only necessary under specific and abormal circumstances. However, currently this expensive operation is being done regardless of need. It's also used in `resolveBuildPlan`, but that only seems to be used by the tests.
True
Avoid unnecessarily loading the hackage index - Loading the hackage is expensive, because the encoded index cache is 18Mb. On my fairly fast computer, it takes 0.3 seconds to decode it. Search `getPackageCaches` to find the spots it's used. There are two cases where loading the hackage currently happens unnecessarily: 1) In `loadSourceMap`, because the user might have specified a package target on the commandline which is not in the resolver. In this case, it uses the latest from hackage. 2) In `constructPlan`, because plan construction errors get annotated with "latest applicable" versions. For both of these, loading the hackage index is only necessary under specific and abormal circumstances. However, currently this expensive operation is being done regardless of need. It's also used in `resolveBuildPlan`, but that only seems to be used by the tests.
non_process
avoid unnecessarily loading the hackage index loading the hackage is expensive because the encoded index cache is on my fairly fast computer it takes seconds to decode it search getpackagecaches to find the spots it s used there are two cases where loading the hackage currently happens unnecessarily in loadsourcemap because the user might have specified a package target on the commandline which is not in the resolver in this case it uses the latest from hackage in constructplan because plan construction errors get annotated with latest applicable versions for both of these loading the hackage index is only necessary under specific and abormal circumstances however currently this expensive operation is being done regardless of need it s also used in resolvebuildplan but that only seems to be used by the tests
0
248,990
26,870,749,324
IssuesEvent
2023-02-04 12:35:24
turkdevops/electron-api-demos
https://api.github.com/repos/turkdevops/electron-api-demos
closed
CVE-2022-25881 (Medium) detected in http-cache-semantics-4.1.0.tgz - autoclosed
security vulnerability
## CVE-2022-25881 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-cache-semantics-4.1.0.tgz</b></p></summary> <p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p> <p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz</a></p> <p> Dependency Hierarchy: <p>Found in HEAD commit: <a href="https://github.com/turkdevops/electron-api-demos/commit/8b3c67fde2016f47e681b745f49afdea23a50ed4">8b3c67fde2016f47e681b745f49afdea23a50ed4</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library. <p>Publish Date: 2023-01-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p> <p>Release Date: 2023-01-31</p> <p>Fix Resolution: http-cache-semantics - 4.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-25881 (Medium) detected in http-cache-semantics-4.1.0.tgz - autoclosed - ## CVE-2022-25881 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-cache-semantics-4.1.0.tgz</b></p></summary> <p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p> <p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz</a></p> <p> Dependency Hierarchy: <p>Found in HEAD commit: <a href="https://github.com/turkdevops/electron-api-demos/commit/8b3c67fde2016f47e681b745f49afdea23a50ed4">8b3c67fde2016f47e681b745f49afdea23a50ed4</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library. <p>Publish Date: 2023-01-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p> <p>Release Date: 2023-01-31</p> <p>Fix Resolution: http-cache-semantics - 4.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in http cache semantics tgz autoclosed cve medium severity vulnerability vulnerable library http cache semantics tgz parses cache control and other headers helps building correct http caches and proxies library home page a href dependency hierarchy found in head commit a href found in base branch master vulnerability details this affects versions of the package http cache semantics before the issue can be exploited via malicious request header values sent to a server when that server reads the cache policy from the request using this library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution http cache semantics step up your open source security game with mend
0
16,413
21,191,507,959
IssuesEvent
2022-04-08 17:59:19
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Webkit: Run internal tests across Webkit browser
process: tests type: chore browser: webkit stage: icebox
In addition to Chrome and Firefox, run our internal tests across Webkit browser. Skip any tests in the Webkit browser that do not currently pass.
1.0
Webkit: Run internal tests across Webkit browser - In addition to Chrome and Firefox, run our internal tests across Webkit browser. Skip any tests in the Webkit browser that do not currently pass.
process
webkit run internal tests across webkit browser in addition to chrome and firefox run our internal tests across webkit browser skip any tests in the webkit browser that do not currently pass
1
299,408
25,901,504,710
IssuesEvent
2022-12-15 06:18:48
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: liquibase failed
C-test-failure O-robot O-roachtest T-sql-sessions branch-release-22.2.0
roachtest.liquibase [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6784087?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6784087?buildTab=artifacts#/liquibase) on release-22.2.0 @ [9bb8a7faf624dcd23ce60e2a8a805ef863b10f72](https://github.com/cockroachdb/cockroach/commits/9bb8a7faf624dcd23ce60e2a8a805ef863b10f72): ``` test artifacts and logs in: /artifacts/liquibase/run_1 orm_helpers.go:191,orm_helpers.go:117,java_helpers.go:220,liquibase.go:123,liquibase.go:140,test_runner.go:930: Tests run on Cockroach v22.2.0-beta.2-106-g9bb8a7faf6 Tests run against liquibase 1790ddef2d0339c5c96839ac60ac424c130dadd8 47 Total Tests Run 38 tests passed 9 tests failed 17 tests skipped 0 tests ignored 0 tests passed unexpectedly 6 tests failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- FAIL: liquibase.harness.change.ChangeObjectTests.apply createSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply addDefaultValueSequenceNext against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply alterSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply addAutoIncrement against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply dropSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply renameSequence against cockroachdb 20.2 - unknown (unexpected) For a full summary look at the liquibase artifacts An updated blocklist (liquibaseBlocklist22_2) is available in the artifacts' liquibase log ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #87586 roachtest: liquibase failed [C-test-failure O-roachtest O-robot branch-release-22.2] - #83313 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.2] - #83050 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-master] - #83048 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1] </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*liquibase.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-20236
2.0
roachtest: liquibase failed - roachtest.liquibase [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6784087?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6784087?buildTab=artifacts#/liquibase) on release-22.2.0 @ [9bb8a7faf624dcd23ce60e2a8a805ef863b10f72](https://github.com/cockroachdb/cockroach/commits/9bb8a7faf624dcd23ce60e2a8a805ef863b10f72): ``` test artifacts and logs in: /artifacts/liquibase/run_1 orm_helpers.go:191,orm_helpers.go:117,java_helpers.go:220,liquibase.go:123,liquibase.go:140,test_runner.go:930: Tests run on Cockroach v22.2.0-beta.2-106-g9bb8a7faf6 Tests run against liquibase 1790ddef2d0339c5c96839ac60ac424c130dadd8 47 Total Tests Run 38 tests passed 9 tests failed 17 tests skipped 0 tests ignored 0 tests passed unexpectedly 6 tests failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- FAIL: liquibase.harness.change.ChangeObjectTests.apply createSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply addDefaultValueSequenceNext against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply alterSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply addAutoIncrement against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply dropSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply renameSequence against cockroachdb 20.2 - unknown (unexpected) For a full summary look at the liquibase artifacts An updated blocklist (liquibaseBlocklist22_2) is available in the artifacts' liquibase log ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #87586 roachtest: liquibase failed [C-test-failure O-roachtest O-robot branch-release-22.2] - #83313 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.2] - #83050 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-master] - #83048 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1] </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*liquibase.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-20236
non_process
roachtest liquibase failed roachtest liquibase with on release test artifacts and logs in artifacts liquibase run orm helpers go orm helpers go java helpers go liquibase go liquibase go test runner go tests run on cockroach beta tests run against liquibase total tests run tests passed tests failed tests skipped tests ignored tests passed unexpectedly tests failed unexpectedly tests expected failed but skipped tests expected failed but not run fail liquibase harness change changeobjecttests apply createsequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply adddefaultvaluesequencenext against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply altersequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply addautoincrement against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply dropsequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply renamesequence against cockroachdb unknown unexpected for a full summary look at the liquibase artifacts an updated blocklist is available in the artifacts liquibase log parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest liquibase failed roachtest liquibase failed roachtest liquibase failed roachtest liquibase failed cc cockroachdb sql experience jira issue crdb
0
378,623
11,205,407,359
IssuesEvent
2020-01-05 14:04:23
kubernetes-sigs/kubefed
https://api.github.com/repos/kubernetes-sigs/kubefed
closed
federation controller disaster recovery Support
kind/feature lifecycle/rotten priority/backlog
<!-- Please only use this template for submitting enhancement requests --> **What would you like to be added**: Currently, we can only have one host cluster, but if the host cluster goes down, then the federation controller will also goes down, and then the federation will not work. **Why is this needed**: It is better to enable the federation controller can support disaster recovery mode, such as we can introduce active-passive mode for federation controller, if the primary controller goes down, then another controller will take over to manage all of the clusters. <!-- DO NOT EDIT BELOW THIS LINE --> /kind feature /cc @marun @kubernetes-sigs/federation-wg
1.0
federation controller disaster recovery Support - <!-- Please only use this template for submitting enhancement requests --> **What would you like to be added**: Currently, we can only have one host cluster, but if the host cluster goes down, then the federation controller will also goes down, and then the federation will not work. **Why is this needed**: It is better to enable the federation controller can support disaster recovery mode, such as we can introduce active-passive mode for federation controller, if the primary controller goes down, then another controller will take over to manage all of the clusters. <!-- DO NOT EDIT BELOW THIS LINE --> /kind feature /cc @marun @kubernetes-sigs/federation-wg
non_process
federation controller disaster recovery support what would you like to be added currently we can only have one host cluster but if the host cluster goes down then the federation controller will also goes down and then the federation will not work why is this needed it is better to enable the federation controller can support disaster recovery mode such as we can introduce active passive mode for federation controller if the primary controller goes down then another controller will take over to manage all of the clusters kind feature cc marun kubernetes sigs federation wg
0
700
3,197,184,575
IssuesEvent
2015-10-01 02:00:35
18F/CMS.gov-developer
https://api.github.com/repos/18F/CMS.gov-developer
closed
Plan training dates
process
* [ ] CMS to consider what kind of trainings and with you. * [ ] CMS to pick tentative dates and times. * [ ] Add these to the [scheduled trainings document](https://github.com/18F/CMS.gov-developer/blob/master/deliverables/draft/training_schedule.md).
1.0
Plan training dates - * [ ] CMS to consider what kind of trainings and with you. * [ ] CMS to pick tentative dates and times. * [ ] Add these to the [scheduled trainings document](https://github.com/18F/CMS.gov-developer/blob/master/deliverables/draft/training_schedule.md).
process
plan training dates cms to consider what kind of trainings and with you cms to pick tentative dates and times add these to the
1
174,844
21,300,487,115
IssuesEvent
2022-04-15 01:59:18
YJSoft/nedb
https://api.github.com/repos/YJSoft/nedb
closed
WS-2017-0247 (Low) detected in ms-0.3.0.tgz - autoclosed
security vulnerability
## WS-2017-0247 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ms-0.3.0.tgz</b></p></summary> <p>Tiny ms conversion utility</p> <p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.3.0.tgz">https://registry.npmjs.org/ms/-/ms-0.3.0.tgz</a></p> <p>Path to dependency file: nedb/package.json</p> <p>Path to vulnerable library: nedb/node_modules/ms</p> <p> Dependency Hierarchy: - mocha-1.4.3.tgz (Root Library) - :x: **ms-0.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/YJSoft/nedb/commit/c3b8d81a383e4da24efcf18e552e70fb79fe742d">c3b8d81a383e4da24efcf18e552e70fb79fe742d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS). <p>Publish Date: 2017-04-12 <p>URL: <a href=https://github.com/zeit/ms/commit/305f2ddcd4eff7cc7c518aca6bb2b2d2daad8fef>WS-2017-0247</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>3.4</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/vercel/ms/pull/89">https://github.com/vercel/ms/pull/89</a></p> <p>Release Date: 2017-04-12</p> <p>Fix Resolution: 2.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2017-0247 (Low) detected in ms-0.3.0.tgz - autoclosed - ## WS-2017-0247 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ms-0.3.0.tgz</b></p></summary> <p>Tiny ms conversion utility</p> <p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.3.0.tgz">https://registry.npmjs.org/ms/-/ms-0.3.0.tgz</a></p> <p>Path to dependency file: nedb/package.json</p> <p>Path to vulnerable library: nedb/node_modules/ms</p> <p> Dependency Hierarchy: - mocha-1.4.3.tgz (Root Library) - :x: **ms-0.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/YJSoft/nedb/commit/c3b8d81a383e4da24efcf18e552e70fb79fe742d">c3b8d81a383e4da24efcf18e552e70fb79fe742d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS). <p>Publish Date: 2017-04-12 <p>URL: <a href=https://github.com/zeit/ms/commit/305f2ddcd4eff7cc7c518aca6bb2b2d2daad8fef>WS-2017-0247</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>3.4</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/vercel/ms/pull/89">https://github.com/vercel/ms/pull/89</a></p> <p>Release Date: 2017-04-12</p> <p>Fix Resolution: 2.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws low detected in ms tgz autoclosed ws low severity vulnerability vulnerable library ms tgz tiny ms conversion utility library home page a href path to dependency file nedb package json path to vulnerable library nedb node modules ms dependency hierarchy mocha tgz root library x ms tgz vulnerable library found in head commit a href found in base branch master vulnerability details affected versions of this package are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
8,950
12,059,071,918
IssuesEvent
2020-04-15 18:36:30
googleapis/cloud-profiler-nodejs
https://api.github.com/repos/googleapis/cloud-profiler-nodejs
closed
Drop support for Node 8
api: cloudprofiler priority: p2 type: process
We'd like to drop support for Node 8 before releasing v4.0.0, because we would like to reduce the number of major version changes for this module. @bcoe and @JustinBeckwith -- At what point in time will we be able to drop Node 8?
1.0
Drop support for Node 8 - We'd like to drop support for Node 8 before releasing v4.0.0, because we would like to reduce the number of major version changes for this module. @bcoe and @JustinBeckwith -- At what point in time will we be able to drop Node 8?
process
drop support for node we d like to drop support for node before releasing because we would like to reduce the number of major version changes for this module bcoe and justinbeckwith at what point in time will we be able to drop node
1
1,926
2,521,738,882
IssuesEvent
2015-01-19 16:34:05
Trashed/MovieSuggester
https://api.github.com/repos/Trashed/MovieSuggester
opened
MainActivity
top priority UI WIP
Build the main view! This view should have: * ActionBar * Search Icon (launches search field), Settings (Icon), Help and About options * Horizontal scrolling lists for top ten suggested movies, actors and genres * These lists are aligned vertically on top of each other
1.0
MainActivity - Build the main view! This view should have: * ActionBar * Search Icon (launches search field), Settings (Icon), Help and About options * Horizontal scrolling lists for top ten suggested movies, actors and genres * These lists are aligned vertically on top of each other
non_process
mainactivity build the main view this view should have actionbar search icon launches search field settings icon help and about options horizontal scrolling lists for top ten suggested movies actors and genres these lists are aligned vertically on top of each other
0
110,514
9,458,893,663
IssuesEvent
2019-04-17 07:02:30
Students-of-the-city-of-Kostroma/Student-timetable
https://api.github.com/repos/Students-of-the-city-of-Kostroma/Student-timetable
closed
Разработать сценарии тестирования пригодных для автоматизации модульным тестированием метода Delete(Model model) сущности ВУЗ
Controller Delete(Model model) Script Unit test ВУЗ ЛР04
[План разработки.](https://docs.google.com/presentation/d/1sLkafCqJTvIAcyZ1jfNw0JrZq8o33WYKGot23z7EfzA/edit#slide=id.g4e940e7976_0_102) [Сценарий.](https://docs.google.com/spreadsheets/d/114F1wKsHoGB75gmF2p_XUR5zgbUb6IeQNX1ziO_BSIw/edit#gid=2120214548) [Ссылка на диаграмму классов.](https://docs.google.com/presentation/d/1sLkafCqJTvIAcyZ1jfNw0JrZq8o33WYKGot23z7EfzA/edit#slide=id.g4e940e7976_0_0)
1.0
Разработать сценарии тестирования пригодных для автоматизации модульным тестированием метода Delete(Model model) сущности ВУЗ - [План разработки.](https://docs.google.com/presentation/d/1sLkafCqJTvIAcyZ1jfNw0JrZq8o33WYKGot23z7EfzA/edit#slide=id.g4e940e7976_0_102) [Сценарий.](https://docs.google.com/spreadsheets/d/114F1wKsHoGB75gmF2p_XUR5zgbUb6IeQNX1ziO_BSIw/edit#gid=2120214548) [Ссылка на диаграмму классов.](https://docs.google.com/presentation/d/1sLkafCqJTvIAcyZ1jfNw0JrZq8o33WYKGot23z7EfzA/edit#slide=id.g4e940e7976_0_0)
non_process
разработать сценарии тестирования пригодных для автоматизации модульным тестированием метода delete model model сущности вуз
0
464,105
13,306,387,839
IssuesEvent
2020-08-25 20:09:37
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
Expand E2E Tests for Facility Locator (Vet Centers)
Q32020-priority QA frontend stretch-goal vsa vsa-facilities
## Background We need to expand the test suite specifically for Vet Center searches ## Tasks **1. Check to see if the following user stories are covered by existing tests** -When `Vet centers` is selected as the facility type, <details> <summary>When `Vet centers` is selected as the facility type (expand)</summary> - The "Choose a service type" is unavailable- No alert displays below search box - Results will be relevant to the geographic location and facility type specified - Search result cards contain required standard information - Alpha icon correlating to map pin - Distance - Hyperlinked name of VA Facility - Facility operating status (if normal, no status is displayed) - Address - Functional Directions hyperlink - Main Number - Mental Health Number (if available) </details> <details> <summary>When the Veteran clicks the hyperlinked facility name, a facility detail page for that facility is displayed (expand)</summary> The Facility Detail page contains the following information about the selected facility - Name of Facility - Alert box with operating status information (if normal, no alert is displayed) - Address - Main Number - Mental Health Number (if available) - Functional Directions hyperlink - List of services - Static map - Hours of Operation displayed by day </details> --- **2. Correct any existing tests if intended functionality has changed since the tests were created.** --- **3. Write additional tests to expand coverage for existing functionality.** --- **4. Conduct testing to ensure tests are correct.** --- ## Acceptance Criteria: - [ ] Tests are present for searches in which "Choose a VA facility type" is set to `Vet Centers` - [ ] Tests all pass with current functionality.
1.0
Expand E2E Tests for Facility Locator (Vet Centers) - ## Background We need to expand the test suite specifically for Vet Center searches ## Tasks **1. Check to see if the following user stories are covered by existing tests** -When `Vet centers` is selected as the facility type, <details> <summary>When `Vet centers` is selected as the facility type (expand)</summary> - The "Choose a service type" is unavailable- No alert displays below search box - Results will be relevant to the geographic location and facility type specified - Search result cards contain required standard information - Alpha icon correlating to map pin - Distance - Hyperlinked name of VA Facility - Facility operating status (if normal, no status is displayed) - Address - Functional Directions hyperlink - Main Number - Mental Health Number (if available) </details> <details> <summary>When the Veteran clicks the hyperlinked facility name, a facility detail page for that facility is displayed (expand)</summary> The Facility Detail page contains the following information about the selected facility - Name of Facility - Alert box with operating status information (if normal, no alert is displayed) - Address - Main Number - Mental Health Number (if available) - Functional Directions hyperlink - List of services - Static map - Hours of Operation displayed by day </details> --- **2. Correct any existing tests if intended functionality has changed since the tests were created.** --- **3. Write additional tests to expand coverage for existing functionality.** --- **4. Conduct testing to ensure tests are correct.** --- ## Acceptance Criteria: - [ ] Tests are present for searches in which "Choose a VA facility type" is set to `Vet Centers` - [ ] Tests all pass with current functionality.
non_process
expand tests for facility locator vet centers background we need to expand the test suite specifically for vet center searches tasks check to see if the following user stories are covered by existing tests when vet centers is selected as the facility type when vet centers is selected as the facility type expand the choose a service type is unavailable no alert displays below search box results will be relevant to the geographic location and facility type specified search result cards contain required standard information alpha icon correlating to map pin distance hyperlinked name of va facility facility operating status if normal no status is displayed address functional directions hyperlink main number mental health number if available when the veteran clicks the hyperlinked facility name a facility detail page for that facility is displayed expand the facility detail page contains the following information about the selected facility name of facility alert box with operating status information if normal no alert is displayed address main number mental health number if available functional directions hyperlink list of services static map hours of operation displayed by day correct any existing tests if intended functionality has changed since the tests were created write additional tests to expand coverage for existing functionality conduct testing to ensure tests are correct acceptance criteria tests are present for searches in which choose a va facility type is set to vet centers tests all pass with current functionality
0
17,258
23,041,373,793
IssuesEvent
2022-07-23 07:26:58
python/cpython
https://api.github.com/repos/python/cpython
closed
test_notify_all hangs forever in sparc64
type-bug tests 3.9 expert-multiprocessing
BPO | [40186](https://bugs.python.org/issue40186) --- | :--- Nosy | @isidentical <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2020-04-04.21:11:33.863> labels = ['type-bug', 'tests', '3.9'] title = 'test_notify_all hangs forever in sparc64' updated_at = <Date 2020-04-04.21:13:20.430> user = 'https://github.com/isidentical' ``` bugs.python.org fields: ```python activity = <Date 2020-04-04.21:13:20.430> actor = 'BTaskaya' assignee = 'none' closed = False closed_date = None closer = None components = ['Tests'] creation = <Date 2020-04-04.21:11:33.863> creator = 'BTaskaya' dependencies = [] files = [] hgrepos = [] issue_num = 40186 keywords = [] message_count = 2.0 messages = ['365784', '365785'] nosy_count = 1.0 nosy_names = ['BTaskaya'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue40186' versions = ['Python 3.9'] ``` </p></details>
1.0
test_notify_all hangs forever in sparc64 - BPO | [40186](https://bugs.python.org/issue40186) --- | :--- Nosy | @isidentical <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2020-04-04.21:11:33.863> labels = ['type-bug', 'tests', '3.9'] title = 'test_notify_all hangs forever in sparc64' updated_at = <Date 2020-04-04.21:13:20.430> user = 'https://github.com/isidentical' ``` bugs.python.org fields: ```python activity = <Date 2020-04-04.21:13:20.430> actor = 'BTaskaya' assignee = 'none' closed = False closed_date = None closer = None components = ['Tests'] creation = <Date 2020-04-04.21:11:33.863> creator = 'BTaskaya' dependencies = [] files = [] hgrepos = [] issue_num = 40186 keywords = [] message_count = 2.0 messages = ['365784', '365785'] nosy_count = 1.0 nosy_names = ['BTaskaya'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue40186' versions = ['Python 3.9'] ``` </p></details>
process
test notify all hangs forever in bpo nosy isidentical note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title test notify all hangs forever in updated at user bugs python org fields python activity actor btaskaya assignee none closed false closed date none closer none components creation creator btaskaya dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type behavior url versions
1
15,593
11,598,159,772
IssuesEvent
2020-02-24 22:25:47
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
opened
ERROR: Failed to download Chromium r594312!
Infrastructure
I've seen this flicker a few times recently. Don't really want to add the `PUPPETEER_SKIP_CHROMIUM_DOWNLOAD` env var but we may have to. ![image](https://user-images.githubusercontent.com/1504756/75196666-50360700-5711-11ea-8640-e2b3b9c36b02.png) > ERROR: Failed to download Chromium r594312! Set "PUPPETEER_SKIP_CHROMIUM_DOWNLOAD" env variable to skip download. > { Error: self signed certificate in certificate chain > at TLSSocket.onConnectSecure (_tls_wrap.js:1051:34) > at TLSSocket.emit (events.js:189:13) > at TLSSocket._finishInit (_tls_wrap.js:633:8) code: 'SELF_SIGNED_CERT_IN_CHAIN' } > Script cd web && yarn install handling the va:web:install event returned with error code 1 > Script @va:web:install was called via post-install-cmd
1.0
ERROR: Failed to download Chromium r594312! - I've seen this flicker a few times recently. Don't really want to add the `PUPPETEER_SKIP_CHROMIUM_DOWNLOAD` env var but we may have to. ![image](https://user-images.githubusercontent.com/1504756/75196666-50360700-5711-11ea-8640-e2b3b9c36b02.png) > ERROR: Failed to download Chromium r594312! Set "PUPPETEER_SKIP_CHROMIUM_DOWNLOAD" env variable to skip download. > { Error: self signed certificate in certificate chain > at TLSSocket.onConnectSecure (_tls_wrap.js:1051:34) > at TLSSocket.emit (events.js:189:13) > at TLSSocket._finishInit (_tls_wrap.js:633:8) code: 'SELF_SIGNED_CERT_IN_CHAIN' } > Script cd web && yarn install handling the va:web:install event returned with error code 1 > Script @va:web:install was called via post-install-cmd
non_process
error failed to download chromium i ve seen this flicker a few times recently don t really want to add the puppeteer skip chromium download env var but we may have to error failed to download chromium set puppeteer skip chromium download env variable to skip download error self signed certificate in certificate chain at tlssocket onconnectsecure tls wrap js at tlssocket emit events js at tlssocket finishinit tls wrap js code self signed cert in chain script cd web yarn install handling the va web install event returned with error code script va web install was called via post install cmd
0
17,502
23,315,752,796
IssuesEvent
2022-08-08 12:22:46
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Primary key in model using a missing column
bug/0-unknown kind/bug process/candidate team/schema topic: prisma db pull topic: Your friendly prisma developers
Hi Prisma Team! Prisma Migrate just crashed. ## Command `db pull` ## Versions | Name | Version | |-------------|--------------------| | Platform | darwin-arm64 | | Node | v16.13.1 | | Prisma CLI | 4.1.0 | | Engine | 8d8414deb360336e4698a65aa45a1fbaf1ce13d8| ## Error ``` Error: [libs/datamodel/connectors/dml/src/model.rs:494:29] Hi there! We've been seeing this error in our error reporting backend, but cannot reproduce it in our own tests. The problem is that we have a primary key in the model `CUSTOMER_BRAND_LINK` that uses the column `DATE_UPDATED` which we for some reason don't have in our internal representation. If you see this, could you please file an issue to https://github.com/prisma/prisma so we can discuss about fixing this. -- Your friendly prisma developers. ```
1.0
Primary key in model using a missing column - Hi Prisma Team! Prisma Migrate just crashed. ## Command `db pull` ## Versions | Name | Version | |-------------|--------------------| | Platform | darwin-arm64 | | Node | v16.13.1 | | Prisma CLI | 4.1.0 | | Engine | 8d8414deb360336e4698a65aa45a1fbaf1ce13d8| ## Error ``` Error: [libs/datamodel/connectors/dml/src/model.rs:494:29] Hi there! We've been seeing this error in our error reporting backend, but cannot reproduce it in our own tests. The problem is that we have a primary key in the model `CUSTOMER_BRAND_LINK` that uses the column `DATE_UPDATED` which we for some reason don't have in our internal representation. If you see this, could you please file an issue to https://github.com/prisma/prisma so we can discuss about fixing this. -- Your friendly prisma developers. ```
process
primary key in model using a missing column hi prisma team prisma migrate just crashed command db pull versions name version platform darwin node prisma cli engine error error hi there we ve been seeing this error in our error reporting backend but cannot reproduce it in our own tests the problem is that we have a primary key in the model customer brand link that uses the column date updated which we for some reason don t have in our internal representation if you see this could you please file an issue to so we can discuss about fixing this your friendly prisma developers
1
44,381
12,124,137,234
IssuesEvent
2020-04-22 13:46:44
carbon-design-system/ibm-security
https://api.github.com/repos/carbon-design-system/ibm-security
closed
Button loading prop should not disable button
Defect severity 4
## Request The `loading` prop on the `Button` component disables the button. I don't believe that this should be the default behavior. Let the consumer decide whether to disable the button or not. If they already pass in a prop to set `loading`, they can use the same one to disable if they want. Not sure if it is a bug or desired behavior, but when setting `loading` to true, the button changes to a ghost button, regardless of initial type.
1.0
Button loading prop should not disable button - ## Request The `loading` prop on the `Button` component disables the button. I don't believe that this should be the default behavior. Let the consumer decide whether to disable the button or not. If they already pass in a prop to set `loading`, they can use the same one to disable if they want. Not sure if it is a bug or desired behavior, but when setting `loading` to true, the button changes to a ghost button, regardless of initial type.
non_process
button loading prop should not disable button request the loading prop on the button component disables the button i don t believe that this should be the default behavior let the consumer decide whether to disable the button or not if they already pass in a prop to set loading they can use the same one to disable if they want not sure if it is a bug or desired behavior but when setting loading to true the button changes to a ghost button regardless of initial type
0
2,870
5,830,076,148
IssuesEvent
2017-05-08 15:56:45
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
reopened
editable function returning false is not honoured - sort on custom formatter does not work
enhancement inprocess
First of all thanks for a great component :thumbsup: - am using it extensively. Couple of issues that i have noticed 1. I use dataFormat that returns a React Component. The rendering is fine but the sort does not work at all. If this is going to be fixed it will be good if there will be an option whether to sort on the unformatted or formatted date (similar to filterFormatted). **(In my example, sort on "Name' column to see the issue)** 2. I use editable as a function to determine is a cell can be edited or not. When the editor is a cutom editor no disabling happens - we can still edit the cell. **(In my example, click on the 'Product' cell to see the issue)** 3) When we have a custom editor, we are forced to have a focus function. Ideally this should be optional 4) When a cell is made non-editable via the editable property (via a function) a click on the cell changes the cell and shows it as disabled - ideally it should not change at all. **In my example click on the id cell and click on the 'Name' cell to see the difference - both are non-editable but what is visible changes)** i.e. functionally there is no issue but it looks odd. The code below can be used to highlight these issues ```javascript class SomeEditor extends React.Component { focus(){ /* PROBLEM 3 : For a custom editor, I have to have a 'focus' function - IDEALLY THIS should be optional*/ } render() { return ( <input autoFocus type="text" /> ) } } const SomeFormatter = ({value}) => { /* Problem 1- where there is a custom formatter - sort does not seem to work */ let valToShow = value.d == 1 ? 'ONE' : 'TWO'; return ( <div> {valToShow} </div> ); }; function bsFormatter(cell, formatter) { let Formatter = formatter; return ( <Formatter value={cell} /> ); } var data = [ {id : 1, date : {d : 1}, product : 'P1'}, {id : 2, date : {d : 2}, product : 'P2'}, ]; const someEditor = (onUpdate, props) => (<SomeEditor onUpdate={ onUpdate } {...props}/>); function editable(){ /* Problem 2: Even though false is always returned the custom editor is not disabled and we can edit */ console.log("Returning FALSE always"); return false; } class ProblemTable extends React.Component { render() { const cellEditProp = { mode: 'click', }; let customEditor = {getElement : someEditor}; let formatter = (cell, row) => bsFormatter(cell, SomeFormatter); return ( <BootstrapTable cellEdit={ cellEditProp } filterFormatted striped hover data={ data }> <TableHeaderColumn dataField='id' isKey>Product ID</TableHeaderColumn> <TableHeaderColumn editable={editable} dataSort={true} dataFormat={formatter} dataField='date'> Name</TableHeaderColumn> <TableHeaderColumn editable={editable} customEditor={ customEditor } dataSort={true} dataField='product'>Product</TableHeaderColumn> </BootstrapTable> ); } }
1.0
editable function returning false is not honoured - sort on custom formatter does not work - First of all thanks for a great component :thumbsup: - am using it extensively. Couple of issues that i have noticed 1. I use dataFormat that returns a React Component. The rendering is fine but the sort does not work at all. If this is going to be fixed it will be good if there will be an option whether to sort on the unformatted or formatted date (similar to filterFormatted). **(In my example, sort on "Name' column to see the issue)** 2. I use editable as a function to determine is a cell can be edited or not. When the editor is a cutom editor no disabling happens - we can still edit the cell. **(In my example, click on the 'Product' cell to see the issue)** 3) When we have a custom editor, we are forced to have a focus function. Ideally this should be optional 4) When a cell is made non-editable via the editable property (via a function) a click on the cell changes the cell and shows it as disabled - ideally it should not change at all. **In my example click on the id cell and click on the 'Name' cell to see the difference - both are non-editable but what is visible changes)** i.e. functionally there is no issue but it looks odd. The code below can be used to highlight these issues ```javascript class SomeEditor extends React.Component { focus(){ /* PROBLEM 3 : For a custom editor, I have to have a 'focus' function - IDEALLY THIS should be optional*/ } render() { return ( <input autoFocus type="text" /> ) } } const SomeFormatter = ({value}) => { /* Problem 1- where there is a custom formatter - sort does not seem to work */ let valToShow = value.d == 1 ? 'ONE' : 'TWO'; return ( <div> {valToShow} </div> ); }; function bsFormatter(cell, formatter) { let Formatter = formatter; return ( <Formatter value={cell} /> ); } var data = [ {id : 1, date : {d : 1}, product : 'P1'}, {id : 2, date : {d : 2}, product : 'P2'}, ]; const someEditor = (onUpdate, props) => (<SomeEditor onUpdate={ onUpdate } {...props}/>); function editable(){ /* Problem 2: Even though false is always returned the custom editor is not disabled and we can edit */ console.log("Returning FALSE always"); return false; } class ProblemTable extends React.Component { render() { const cellEditProp = { mode: 'click', }; let customEditor = {getElement : someEditor}; let formatter = (cell, row) => bsFormatter(cell, SomeFormatter); return ( <BootstrapTable cellEdit={ cellEditProp } filterFormatted striped hover data={ data }> <TableHeaderColumn dataField='id' isKey>Product ID</TableHeaderColumn> <TableHeaderColumn editable={editable} dataSort={true} dataFormat={formatter} dataField='date'> Name</TableHeaderColumn> <TableHeaderColumn editable={editable} customEditor={ customEditor } dataSort={true} dataField='product'>Product</TableHeaderColumn> </BootstrapTable> ); } }
process
editable function returning false is not honoured sort on custom formatter does not work first of all thanks for a great component thumbsup am using it extensively couple of issues that i have noticed i use dataformat that returns a react component the rendering is fine but the sort does not work at all if this is going to be fixed it will be good if there will be an option whether to sort on the unformatted or formatted date similar to filterformatted in my example sort on name column to see the issue i use editable as a function to determine is a cell can be edited or not when the editor is a cutom editor no disabling happens we can still edit the cell in my example click on the product cell to see the issue when we have a custom editor we are forced to have a focus function ideally this should be optional when a cell is made non editable via the editable property via a function a click on the cell changes the cell and shows it as disabled ideally it should not change at all in my example click on the id cell and click on the name cell to see the difference both are non editable but what is visible changes i e functionally there is no issue but it looks odd the code below can be used to highlight these issues javascript class someeditor extends react component focus problem for a custom editor i have to have a focus function ideally this should be optional render return const someformatter value problem where there is a custom formatter sort does not seem to work let valtoshow value d one two return valtoshow function bsformatter cell formatter let formatter formatter return var data id date d product id date d product const someeditor onupdate props function editable problem even though false is always returned the custom editor is not disabled and we can edit console log returning false always return false class problemtable extends react component render const celleditprop mode click let customeditor getelement someeditor let formatter cell row bsformatter cell someformatter return product id name product
1
6,813
9,956,647,326
IssuesEvent
2019-07-05 14:28:06
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Issue with custom database path
log-processing
I'm trying to keep incremental logs in the btree DB. My initial command was: ```zcat /usr/local/www/usesthis.com/log/access.log.*.gz | goaccess --keep-db-files --db-path=/usr/local/share/stats/usesthis/ --log-format=COMBINED -o /usr/local/www/usesthis.com/public/stats.html -``` Which worked great, I can see a ton of data - I want to start from a solid base of old logs, etc. I want to set up a cronjob that appends the logs from `access.log`/`access.log.1` on a daily basis, so I figured I'm meant to do something like this: ```goaccess --keep-db-files --db-path=/usr/local/shared/stats/usesthis/ /usr/local/www/usesthis.com/log/access.log /usr/local/www/usesthis.com/log/access.log.1 --log-format=COMBINED -o /usr/local/www/usesthis.com/public/stats.html``` However, I get: ``` GoAccess - version 1.2 - Jun 19 2017 12:55:29 Config file: No config file used Fatal error has occurred Error occured at: src/tcbtdb.c - tc_db_set_path - 66 Unable to access database path: No such file or directory ``` A) Am I understanding how to even do the incremental data stuff correctly? B) Why can't it access the database path? Not using a config file, so what you see is all there. I can see all the .tcb files in `/usr/local/share/stats/usesthis`, and trying with or without a trailing slash doesn't seem to make a difference. What am I doing wrong?
1.0
Issue with custom database path - I'm trying to keep incremental logs in the btree DB. My initial command was: ```zcat /usr/local/www/usesthis.com/log/access.log.*.gz | goaccess --keep-db-files --db-path=/usr/local/share/stats/usesthis/ --log-format=COMBINED -o /usr/local/www/usesthis.com/public/stats.html -``` Which worked great, I can see a ton of data - I want to start from a solid base of old logs, etc. I want to set up a cronjob that appends the logs from `access.log`/`access.log.1` on a daily basis, so I figured I'm meant to do something like this: ```goaccess --keep-db-files --db-path=/usr/local/shared/stats/usesthis/ /usr/local/www/usesthis.com/log/access.log /usr/local/www/usesthis.com/log/access.log.1 --log-format=COMBINED -o /usr/local/www/usesthis.com/public/stats.html``` However, I get: ``` GoAccess - version 1.2 - Jun 19 2017 12:55:29 Config file: No config file used Fatal error has occurred Error occured at: src/tcbtdb.c - tc_db_set_path - 66 Unable to access database path: No such file or directory ``` A) Am I understanding how to even do the incremental data stuff correctly? B) Why can't it access the database path? Not using a config file, so what you see is all there. I can see all the .tcb files in `/usr/local/share/stats/usesthis`, and trying with or without a trailing slash doesn't seem to make a difference. What am I doing wrong?
process
issue with custom database path i m trying to keep incremental logs in the btree db my initial command was zcat usr local www usesthis com log access log gz goaccess keep db files db path usr local share stats usesthis log format combined o usr local www usesthis com public stats html which worked great i can see a ton of data i want to start from a solid base of old logs etc i want to set up a cronjob that appends the logs from access log access log on a daily basis so i figured i m meant to do something like this goaccess keep db files db path usr local shared stats usesthis usr local www usesthis com log access log usr local www usesthis com log access log log format combined o usr local www usesthis com public stats html however i get goaccess version jun config file no config file used fatal error has occurred error occured at src tcbtdb c tc db set path unable to access database path no such file or directory a am i understanding how to even do the incremental data stuff correctly b why can t it access the database path not using a config file so what you see is all there i can see all the tcb files in usr local share stats usesthis and trying with or without a trailing slash doesn t seem to make a difference what am i doing wrong
1
147,298
23,196,248,935
IssuesEvent
2022-08-01 16:39:15
microsoft/fluentui
https://api.github.com/repos/microsoft/fluentui
closed
[Bug]: SplitButton opens the dropdown menu instead of invoking the onClick action when used with touch.
Resolution: By Design Fluent UI react (v8) Component: SplitButton
### Library React / v8 (@fluentui/react) ### System Info ```shell System: OS: Windows 10 10.0.22000 CPU: (8) x64 Intel(R) Core(TM) i7-1065G7 CPU @ 1.30GHz Memory: 2.07 GB / 15.60 GB Browsers: Edge: Spartan (44.22000.120.0), Chromium (102.0.1245.44), ChromiumDev (104.0.1293.1) Internet Explorer: 11.0.22000.120 ``` ### Are you reporting Accessibility issue? no ### Reproduction https://developer.microsoft.com/en-us/fluentui#/controls/web/button ### Bug Description ## Actual Behavior When the SplitButton is used with touch, the dropdown menu appears even if user touches on the left portion of the split button. This is reproducible on https://developer.microsoft.com/en-us/fluentui#/controls/web/button sample controls for the SplitButton component as well. ## Expected Behavior Behavior with touch should be same as behavior with mouse click. i.e. when user touches the left portion of the split button, it should invoke the onClick function rather than opening the dropdown. ### Logs _No response_ ### Requested priority Normal ### Products/sites affected Outlook Web ### Are you willing to submit a PR to fix? no ### Validations - [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate. - [X] The provided reproduction is a minimal reproducible example of the bug.
1.0
[Bug]: SplitButton opens the dropdown menu instead of invoking the onClick action when used with touch. - ### Library React / v8 (@fluentui/react) ### System Info ```shell System: OS: Windows 10 10.0.22000 CPU: (8) x64 Intel(R) Core(TM) i7-1065G7 CPU @ 1.30GHz Memory: 2.07 GB / 15.60 GB Browsers: Edge: Spartan (44.22000.120.0), Chromium (102.0.1245.44), ChromiumDev (104.0.1293.1) Internet Explorer: 11.0.22000.120 ``` ### Are you reporting Accessibility issue? no ### Reproduction https://developer.microsoft.com/en-us/fluentui#/controls/web/button ### Bug Description ## Actual Behavior When the SplitButton is used with touch, the dropdown menu appears even if user touches on the left portion of the split button. This is reproducible on https://developer.microsoft.com/en-us/fluentui#/controls/web/button sample controls for the SplitButton component as well. ## Expected Behavior Behavior with touch should be same as behavior with mouse click. i.e. when user touches the left portion of the split button, it should invoke the onClick function rather than opening the dropdown. ### Logs _No response_ ### Requested priority Normal ### Products/sites affected Outlook Web ### Are you willing to submit a PR to fix? no ### Validations - [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate. - [X] The provided reproduction is a minimal reproducible example of the bug.
non_process
splitbutton opens the dropdown menu instead of invoking the onclick action when used with touch library react fluentui react system info shell system os windows cpu intel r core tm cpu memory gb gb browsers edge spartan chromium chromiumdev internet explorer are you reporting accessibility issue no reproduction bug description actual behavior when the splitbutton is used with touch the dropdown menu appears even if user touches on the left portion of the split button this is reproducible on sample controls for the splitbutton component as well expected behavior behavior with touch should be same as behavior with mouse click i e when user touches the left portion of the split button it should invoke the onclick function rather than opening the dropdown logs no response requested priority normal products sites affected outlook web are you willing to submit a pr to fix no validations check that there isn t already an issue that reports the same bug to avoid creating a duplicate the provided reproduction is a minimal reproducible example of the bug
0
420,406
28,263,462,804
IssuesEvent
2023-04-07 03:03:33
AY2223S2-CS2113-T13-1/tp
https://api.github.com/repos/AY2223S2-CS2113-T13-1/tp
closed
[PE-D][Tester B] Lack of example to show command with optional description
documentation priority.Medium
![image.png](https://raw.githubusercontent.com/jaredoong/ped/main/files/0cfb8d90-81d5-40c9-9843-64d9a08b06d3.png) Would be good if there was an example on how to use `add` with the optional parameter description <!--session: 1680252471869-a4ff2ce6-21c8-4fa9-ad69-6bc92f657b80--> <!--Version: Web v3.4.7--> ------------- Labels: `severity.Low` `type.DocumentationBug` original: jaredoong/ped#8
1.0
[PE-D][Tester B] Lack of example to show command with optional description - ![image.png](https://raw.githubusercontent.com/jaredoong/ped/main/files/0cfb8d90-81d5-40c9-9843-64d9a08b06d3.png) Would be good if there was an example on how to use `add` with the optional parameter description <!--session: 1680252471869-a4ff2ce6-21c8-4fa9-ad69-6bc92f657b80--> <!--Version: Web v3.4.7--> ------------- Labels: `severity.Low` `type.DocumentationBug` original: jaredoong/ped#8
non_process
lack of example to show command with optional description would be good if there was an example on how to use add with the optional parameter description labels severity low type documentationbug original jaredoong ped
0
365,623
10,790,040,692
IssuesEvent
2019-11-05 13:19:37
AY1920S1-CS2113T-F09-4/main
https://api.github.com/repos/AY1920S1-CS2113T-F09-4/main
closed
Documentation not uniform in examples, making it hard to follow.
priority.High severity.Medium status.Ongoing
![Screenshot (48).png](https://raw.githubusercontent.com/JasonLeeWeiHern/ped/master/files/606d1faf-2728-47ba-a19d-31fe86d8a381.png) Difference in example, when executed, I won't find jingle inside. Confusing for user. <hr><sub>[original: JasonLeeWeiHern/ped#7]<br/> </sub>
1.0
Documentation not uniform in examples, making it hard to follow. - ![Screenshot (48).png](https://raw.githubusercontent.com/JasonLeeWeiHern/ped/master/files/606d1faf-2728-47ba-a19d-31fe86d8a381.png) Difference in example, when executed, I won't find jingle inside. Confusing for user. <hr><sub>[original: JasonLeeWeiHern/ped#7]<br/> </sub>
non_process
documentation not uniform in examples making it hard to follow difference in example when executed i won t find jingle inside confusing for user
0
192,570
15,353,232,881
IssuesEvent
2021-03-01 08:17:37
kbuffington/Georgia
https://api.github.com/repos/kbuffington/Georgia
closed
Suggestion for Instructions
documentation
I'm super casual on this kind of thing (Not even coding, just installing) So, the roadblock I ran into was on step 8. I didn't understand how to add the Jscript Pannel, specifically as I'm ultra new to Foobar. Luckily your video helped me out there, through some reddit skimming, and happening to run across it on youtube. It may seem super trivial, but a little bit of a more direct explanation on adding a Jscript Pannel more specifically targeted at someone who had just downloaded Fubar for the first time (Like myself) could prove helpful in progressing through the install process. Simple suggestion would be View>Layout>New Scratchbox, ect.. added to the instruction flow. Awesome configuration, and beautiful design. Keep up the amazing work!
1.0
Suggestion for Instructions - I'm super casual on this kind of thing (Not even coding, just installing) So, the roadblock I ran into was on step 8. I didn't understand how to add the Jscript Pannel, specifically as I'm ultra new to Foobar. Luckily your video helped me out there, through some reddit skimming, and happening to run across it on youtube. It may seem super trivial, but a little bit of a more direct explanation on adding a Jscript Pannel more specifically targeted at someone who had just downloaded Fubar for the first time (Like myself) could prove helpful in progressing through the install process. Simple suggestion would be View>Layout>New Scratchbox, ect.. added to the instruction flow. Awesome configuration, and beautiful design. Keep up the amazing work!
non_process
suggestion for instructions i m super casual on this kind of thing not even coding just installing so the roadblock i ran into was on step i didn t understand how to add the jscript pannel specifically as i m ultra new to foobar luckily your video helped me out there through some reddit skimming and happening to run across it on youtube it may seem super trivial but a little bit of a more direct explanation on adding a jscript pannel more specifically targeted at someone who had just downloaded fubar for the first time like myself could prove helpful in progressing through the install process simple suggestion would be view layout new scratchbox ect added to the instruction flow awesome configuration and beautiful design keep up the amazing work
0
6,506
9,592,746,130
IssuesEvent
2019-05-09 09:40:56
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
opened
Obsolete 'GO:0046731 passive induction of host immune response by virus' and children?
multi-species process
Hello, GO:0046731 passive induction of host immune response by virus is defined as "The unintentional stimulation by a virus of a host defense response to viral infection, as part of the viral infectious cycle." - so that doesn't sound like any gene product is actually doing anything. Similar for the children: - GO:0046734 passive induction of host cell-mediated immune response by virus - GO:0046733 passive induction of host humoral immune response by virus - GO:0046735 passive induction of host innate immune response by virus There is a single annotation to this term by SWIE (@milarolo ) to a human protein (which is anyway inconsistent with the term's meaning. I propose to obsolete that branch, or merge it into GO:0046730 induction of host immune response by virus @mgiglio99 @pmasson55 @ValWood What do you think ? Thanks, Pascale
1.0
Obsolete 'GO:0046731 passive induction of host immune response by virus' and children? - Hello, GO:0046731 passive induction of host immune response by virus is defined as "The unintentional stimulation by a virus of a host defense response to viral infection, as part of the viral infectious cycle." - so that doesn't sound like any gene product is actually doing anything. Similar for the children: - GO:0046734 passive induction of host cell-mediated immune response by virus - GO:0046733 passive induction of host humoral immune response by virus - GO:0046735 passive induction of host innate immune response by virus There is a single annotation to this term by SWIE (@milarolo ) to a human protein (which is anyway inconsistent with the term's meaning. I propose to obsolete that branch, or merge it into GO:0046730 induction of host immune response by virus @mgiglio99 @pmasson55 @ValWood What do you think ? Thanks, Pascale
process
obsolete go passive induction of host immune response by virus and children hello go passive induction of host immune response by virus is defined as the unintentional stimulation by a virus of a host defense response to viral infection as part of the viral infectious cycle so that doesn t sound like any gene product is actually doing anything similar for the children go passive induction of host cell mediated immune response by virus go passive induction of host humoral immune response by virus go passive induction of host innate immune response by virus there is a single annotation to this term by swie milarolo to a human protein which is anyway inconsistent with the term s meaning i propose to obsolete that branch or merge it into go induction of host immune response by virus valwood what do you think thanks pascale
1
4,311
7,203,155,640
IssuesEvent
2018-02-06 08:06:08
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[processing] [FEATURE] SplitWithLines
Automatic new feature Processing
Original commit: https://github.com/qgis/QGIS/commit/0e2ef065d7e89ba65db93f3a628d46c9bb31f265 by nyalldawson Rename algorithm SplitLinesWithLines to SplitWithLines Accept polygon as input, too Use only selected lines to split with (if processing is set to use selection only) Issue log message if trying to split multi geometries Update help
1.0
[processing] [FEATURE] SplitWithLines - Original commit: https://github.com/qgis/QGIS/commit/0e2ef065d7e89ba65db93f3a628d46c9bb31f265 by nyalldawson Rename algorithm SplitLinesWithLines to SplitWithLines Accept polygon as input, too Use only selected lines to split with (if processing is set to use selection only) Issue log message if trying to split multi geometries Update help
process
splitwithlines original commit by nyalldawson rename algorithm splitlineswithlines to splitwithlines accept polygon as input too use only selected lines to split with if processing is set to use selection only issue log message if trying to split multi geometries update help
1
8,730
11,863,341,717
IssuesEvent
2020-03-25 19:32:27
w3c/webauthn
https://api.github.com/repos/w3c/webauthn
closed
Add direction to somewhere to ask for help to the contributing guidelines
type:process
I have been looking around this repo and finding answers to most of the questions I have in various issues and PR's. Does such a community exist which will accept questions for people trying to use webauthn? It would be great if there was some line in the contributing.md which said "If you are unsure if your issue belongs here please ask in X" If there is no suitable community to put in as X, it would still be helpful to point to an FAQ page or even tag on stackoverflow
1.0
Add direction to somewhere to ask for help to the contributing guidelines - I have been looking around this repo and finding answers to most of the questions I have in various issues and PR's. Does such a community exist which will accept questions for people trying to use webauthn? It would be great if there was some line in the contributing.md which said "If you are unsure if your issue belongs here please ask in X" If there is no suitable community to put in as X, it would still be helpful to point to an FAQ page or even tag on stackoverflow
process
add direction to somewhere to ask for help to the contributing guidelines i have been looking around this repo and finding answers to most of the questions i have in various issues and pr s does such a community exist which will accept questions for people trying to use webauthn it would be great if there was some line in the contributing md which said if you are unsure if your issue belongs here please ask in x if there is no suitable community to put in as x it would still be helpful to point to an faq page or even tag on stackoverflow
1
7,417
10,541,888,562
IssuesEvent
2019-10-02 11:57:34
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Run the breaking change detector on PRs
type: process
For APIs which have already gone GA, it would be useful to run our breaking change detector against the published version. Will need to look to see how tricky that is.
1.0
Run the breaking change detector on PRs - For APIs which have already gone GA, it would be useful to run our breaking change detector against the published version. Will need to look to see how tricky that is.
process
run the breaking change detector on prs for apis which have already gone ga it would be useful to run our breaking change detector against the published version will need to look to see how tricky that is
1
86,963
8,055,089,918
IssuesEvent
2018-08-02 08:12:09
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
Specific Timestamps breaks time series indexing (.loc returns wrong results)
Indexing Testing Timezones good first issue
When try to access labels (`.loc`) by using a specific list of `Timestamp` objects or a `DatetimeIndex` object (See attached csv file), the resulting index is returned in UTC offset but the original timezone is not removed. This seems to happen only in very specific cases, when the index passed to `.loc` contains labels that do not exist in the DataFrame and also contains duplicates. @yuval-jether ```python import pandas as pd import pytz idx = pd.date_range('2011-01-01', '2017-10-01 00:00:00', freq='h', tz='America/Chicago') s = pd.Series(np.random.rand(len(idx)), index=idx) i = pd.to_datetime(['2017-09-30 06:00:00', '2017-09-30 23:00:00', '2017-09-30 23:00:00', '2017-10-01 00:00:00',]).tz_localize('America/Chicago') s.loc[i] Out[39]: 2017-09-30 06:00:00-05:00 0.380138 2017-09-30 23:00:00-05:00 0.774696 2017-09-30 23:00:00-05:00 0.774696 2017-10-01 00:00:00-05:00 0.728027 dtype: float64 # Added a label that does not exist in the Series import pandas as pd import pytz idx = pd.date_range('2011-01-01', '2017-10-01 00:00:00', freq='h', tz='America/Chicago') s = pd.Series(np.random.rand(len(idx)), index=idx) i = pd.to_datetime(['2017-09-30 06:00:00', '2017-09-30 23:00:00', '2017-09-30 23:00:00', '2017-10-01 00:00:00', '2017-10-01 01:00:00',]).tz_localize('America/Chicago') s.loc[i] Out[40]: 2017-09-30 11:00:00-05:00 0.645350 2017-10-01 04:00:00-05:00 0.099323 2017-10-01 04:00:00-05:00 0.099323 2017-10-01 05:00:00-05:00 0.037136 2017-10-01 06:00:00-05:00 NaN dtype: float64 ``` #### Output of ``pd.show_versions()`` <details> pd.show_versions() 2017-10-30 03:15:08 [pip.utils] [DEBUG] lzma module is not available 2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: git 2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: hg 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_DEBUG' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'WITH_PYMALLOC' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_UNICODE_SIZE' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_DEBUG' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'WITH_PYMALLOC' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_UNICODE_SIZE' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: svn 2017-10-30 03:15:09 [pip.vcs] [DEBUG] Registered VCS backend: bzr INSTALLED VERSIONS ------------------ commit: None python: 2.7.12.final.0 python-bits: 64 OS: Windows OS-release: 8.1 machine: AMD64 processor: Intel64 Family 6 Model 60 Stepping 3, GenuineIntel byteorder: little LC_ALL: None LANG: None LOCALE: None.None pandas: 0.20.3 pytest: 3.0.7 pip: 9.0.1 setuptools: 36.5.0 Cython: None numpy: 1.13.1 scipy: 0.18.1 xarray: 0.9.6 IPython: 5.5.0 sphinx: None patsy: 0.4.1 dateutil: 2.6.1 pytz: 2017.2 blosc: None bottleneck: None tables: None numexpr: 2.6.2 feather: None matplotlib: 2.0.2 openpyxl: 2.4.8 xlrd: 1.0.0 xlwt: 1.2.0 xlsxwriter: None lxml: 3.8.0 bs4: 4.5.1 html5lib: 0.999999999 sqlalchemy: None pymysql: None psycopg2: None jinja2: 2.9.6 s3fs: None pandas_gbq: None pandas_datareader: 0.4.0 </details>
1.0
Specific Timestamps breaks time series indexing (.loc returns wrong results) - When try to access labels (`.loc`) by using a specific list of `Timestamp` objects or a `DatetimeIndex` object (See attached csv file), the resulting index is returned in UTC offset but the original timezone is not removed. This seems to happen only in very specific cases, when the index passed to `.loc` contains labels that do not exist in the DataFrame and also contains duplicates. @yuval-jether ```python import pandas as pd import pytz idx = pd.date_range('2011-01-01', '2017-10-01 00:00:00', freq='h', tz='America/Chicago') s = pd.Series(np.random.rand(len(idx)), index=idx) i = pd.to_datetime(['2017-09-30 06:00:00', '2017-09-30 23:00:00', '2017-09-30 23:00:00', '2017-10-01 00:00:00',]).tz_localize('America/Chicago') s.loc[i] Out[39]: 2017-09-30 06:00:00-05:00 0.380138 2017-09-30 23:00:00-05:00 0.774696 2017-09-30 23:00:00-05:00 0.774696 2017-10-01 00:00:00-05:00 0.728027 dtype: float64 # Added a label that does not exist in the Series import pandas as pd import pytz idx = pd.date_range('2011-01-01', '2017-10-01 00:00:00', freq='h', tz='America/Chicago') s = pd.Series(np.random.rand(len(idx)), index=idx) i = pd.to_datetime(['2017-09-30 06:00:00', '2017-09-30 23:00:00', '2017-09-30 23:00:00', '2017-10-01 00:00:00', '2017-10-01 01:00:00',]).tz_localize('America/Chicago') s.loc[i] Out[40]: 2017-09-30 11:00:00-05:00 0.645350 2017-10-01 04:00:00-05:00 0.099323 2017-10-01 04:00:00-05:00 0.099323 2017-10-01 05:00:00-05:00 0.037136 2017-10-01 06:00:00-05:00 NaN dtype: float64 ``` #### Output of ``pd.show_versions()`` <details> pd.show_versions() 2017-10-30 03:15:08 [pip.utils] [DEBUG] lzma module is not available 2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: git 2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: hg 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_DEBUG' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'WITH_PYMALLOC' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_UNICODE_SIZE' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_DEBUG' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'WITH_PYMALLOC' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_UNICODE_SIZE' is unset, Python ABI tag may be incorrect 2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: svn 2017-10-30 03:15:09 [pip.vcs] [DEBUG] Registered VCS backend: bzr INSTALLED VERSIONS ------------------ commit: None python: 2.7.12.final.0 python-bits: 64 OS: Windows OS-release: 8.1 machine: AMD64 processor: Intel64 Family 6 Model 60 Stepping 3, GenuineIntel byteorder: little LC_ALL: None LANG: None LOCALE: None.None pandas: 0.20.3 pytest: 3.0.7 pip: 9.0.1 setuptools: 36.5.0 Cython: None numpy: 1.13.1 scipy: 0.18.1 xarray: 0.9.6 IPython: 5.5.0 sphinx: None patsy: 0.4.1 dateutil: 2.6.1 pytz: 2017.2 blosc: None bottleneck: None tables: None numexpr: 2.6.2 feather: None matplotlib: 2.0.2 openpyxl: 2.4.8 xlrd: 1.0.0 xlwt: 1.2.0 xlsxwriter: None lxml: 3.8.0 bs4: 4.5.1 html5lib: 0.999999999 sqlalchemy: None pymysql: None psycopg2: None jinja2: 2.9.6 s3fs: None pandas_gbq: None pandas_datareader: 0.4.0 </details>
non_process
specific timestamps breaks time series indexing loc returns wrong results when try to access labels loc by using a specific list of timestamp objects or a datetimeindex object see attached csv file the resulting index is returned in utc offset but the original timezone is not removed this seems to happen only in very specific cases when the index passed to loc contains labels that do not exist in the dataframe and also contains duplicates yuval jether python import pandas as pd import pytz idx pd date range freq h tz america chicago s pd series np random rand len idx index idx i pd to datetime tz localize america chicago s loc out dtype added a label that does not exist in the series import pandas as pd import pytz idx pd date range freq h tz america chicago s pd series np random rand len idx index idx i pd to datetime tz localize america chicago s loc out nan dtype output of pd show versions pd show versions lzma module is not available registered vcs backend git registered vcs backend hg config variable py debug is unset python abi tag may be incorrect config variable with pymalloc is unset python abi tag may be incorrect config variable py unicode size is unset python abi tag may be incorrect config variable py debug is unset python abi tag may be incorrect config variable with pymalloc is unset python abi tag may be incorrect config variable py unicode size is unset python abi tag may be incorrect registered vcs backend svn registered vcs backend bzr installed versions commit none python final python bits os windows os release machine processor family model stepping genuineintel byteorder little lc all none lang none locale none none pandas pytest pip setuptools cython none numpy scipy xarray ipython sphinx none patsy dateutil pytz blosc none bottleneck none tables none numexpr feather none matplotlib openpyxl xlrd xlwt xlsxwriter none lxml sqlalchemy none pymysql none none none pandas gbq none pandas datareader
0
400,104
11,769,162,500
IssuesEvent
2020-03-15 13:45:07
microsoft/terraform-provider-azuredevops
https://api.github.com/repos/microsoft/terraform-provider-azuredevops
closed
Implement a Terraform data source to lookup Git repositories
good first issue priority-low
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> The provider currently only provides a way to provision Azure Git Repos, but does not allow us to reference a pre-existing repository. Having a data source for Git Repos will allow people to leverage other resources, like build definitions and repository permissions, without needing to create a new repository. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * data.azuredevops_git_repository ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azuredevops_project" "main_project" { project_name = "Test Project" description = "Test Project Description" visibility = "private" enable_tfvc = False work_item_template = "Agile" } data azuredevops_git_repository "lookup-repo" { project_id = azuredevops_project.main_project.id name = "Test Lookup Repo" } ``` ### Acceptance Criteria - [ ] A data source exists that allows for query of an Azure Git Repos based on the following fields: - [ ] `project_id` - ID of project in which the repository exists - [ ] `name` - name of repository - [ ] The `computed` fields of the resource should match the resource implementation of [Azure Git Repos](https://github.com/microsoft/terraform-provider-azuredevops/blob/master/azuredevops/resource_azure_git_repository.go). There may be an opportunity to re-use the whole resource schema! - [ ] Data source is validated with unit tests - [ ] Data source is validated with acceptance tests - [ ] Documentation is added in the client-facing docs ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? ---> * #84 * [Repositories - List](https://docs.microsoft.com/en-us/rest/api/azure/devops/git/repositories/list?view=azure-devops-rest-5.1)
1.0
Implement a Terraform data source to lookup Git repositories - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> The provider currently only provides a way to provision Azure Git Repos, but does not allow us to reference a pre-existing repository. Having a data source for Git Repos will allow people to leverage other resources, like build definitions and repository permissions, without needing to create a new repository. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * data.azuredevops_git_repository ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azuredevops_project" "main_project" { project_name = "Test Project" description = "Test Project Description" visibility = "private" enable_tfvc = False work_item_template = "Agile" } data azuredevops_git_repository "lookup-repo" { project_id = azuredevops_project.main_project.id name = "Test Lookup Repo" } ``` ### Acceptance Criteria - [ ] A data source exists that allows for query of an Azure Git Repos based on the following fields: - [ ] `project_id` - ID of project in which the repository exists - [ ] `name` - name of repository - [ ] The `computed` fields of the resource should match the resource implementation of [Azure Git Repos](https://github.com/microsoft/terraform-provider-azuredevops/blob/master/azuredevops/resource_azure_git_repository.go). There may be an opportunity to re-use the whole resource schema! - [ ] Data source is validated with unit tests - [ ] Data source is validated with acceptance tests - [ ] Documentation is added in the client-facing docs ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? ---> * #84 * [Repositories - List](https://docs.microsoft.com/en-us/rest/api/azure/devops/git/repositories/list?view=azure-devops-rest-5.1)
non_process
implement a terraform data source to lookup git repositories community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the provider currently only provides a way to provision azure git repos but does not allow us to reference a pre existing repository having a data source for git repos will allow people to leverage other resources like build definitions and repository permissions without needing to create a new repository new or affected resource s data azuredevops git repository potential terraform configuration hcl resource azuredevops project main project project name test project description test project description visibility private enable tfvc false work item template agile data azuredevops git repository lookup repo project id azuredevops project main project id name test lookup repo acceptance criteria a data source exists that allows for query of an azure git repos based on the following fields project id id of project in which the repository exists name name of repository the computed fields of the resource should match the resource implementation of there may be an opportunity to re use the whole resource schema data source is validated with unit tests data source is validated with acceptance tests documentation is added in the client facing docs references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation
0
13,762
16,521,836,538
IssuesEvent
2021-05-26 15:17:53
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
allowPartiallySucceededBuilds for deployment jobs
cba devops-cicd-process/tech devops/prod product-question
Hey Devops Gurus, Is there to allowPartiallySucceededBuilds for deployment jobs? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 5aeeaace-1c5b-a51b-e41f-f25b806155b8 * Version Independent ID: fd7ff690-b2e4-41c7-a342-e528b911c6e1 * Content: [Deployment jobs - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/deployment-jobs?view=azure-devops) * Content Source: [docs/pipelines/process/deployment-jobs.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/deployment-jobs.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
allowPartiallySucceededBuilds for deployment jobs - Hey Devops Gurus, Is there to allowPartiallySucceededBuilds for deployment jobs? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 5aeeaace-1c5b-a51b-e41f-f25b806155b8 * Version Independent ID: fd7ff690-b2e4-41c7-a342-e528b911c6e1 * Content: [Deployment jobs - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/deployment-jobs?view=azure-devops) * Content Source: [docs/pipelines/process/deployment-jobs.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/deployment-jobs.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
allowpartiallysucceededbuilds for deployment jobs hey devops gurus is there to allowpartiallysucceededbuilds for deployment jobs document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
22,625
31,847,770,452
IssuesEvent
2023-09-14 21:28:34
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Port MLv1 `hasWritePermission` to MLv2 (only for native queries)
Querying/Native .Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
There're a few MLv1 methods the editor uses that'd be nice to port to MLv2 as well. Port `hasWritePermission` (only native) to MLv2 https://github.com/metabase/metabase/blob/dbfca6c6d173294ddcf97b394750574b4ef10221/frontend/src/metabase-lib/queries/NativeQuery.ts#L200
1.0
[MLv2] Port MLv1 `hasWritePermission` to MLv2 (only for native queries) - There're a few MLv1 methods the editor uses that'd be nice to port to MLv2 as well. Port `hasWritePermission` (only native) to MLv2 https://github.com/metabase/metabase/blob/dbfca6c6d173294ddcf97b394750574b4ef10221/frontend/src/metabase-lib/queries/NativeQuery.ts#L200
process
port haswritepermission to only for native queries there re a few methods the editor uses that d be nice to port to as well port haswritepermission only native to
1
75,803
3,476,132,228
IssuesEvent
2015-12-26 14:24:59
Stephane-D/SGDK
https://api.github.com/repos/Stephane-D/SGDK
closed
recomp sprite collision
Priority-Low
Hi, rescomp readme is wrong: Collision should be set to box by default: https://github.com/Stephane-D/SGDK/blob/master/tools/rescomp/rescomp.txt#L143 but: https://github.com/Stephane-D/SGDK/blob/10f2039eacac7d72a6eb63b274af6b95c91a85b8/tools/rescomp/src/sprite.c#L72 it's set to none
1.0
recomp sprite collision - Hi, rescomp readme is wrong: Collision should be set to box by default: https://github.com/Stephane-D/SGDK/blob/master/tools/rescomp/rescomp.txt#L143 but: https://github.com/Stephane-D/SGDK/blob/10f2039eacac7d72a6eb63b274af6b95c91a85b8/tools/rescomp/src/sprite.c#L72 it's set to none
non_process
recomp sprite collision hi rescomp readme is wrong collision should be set to box by default but it s set to none
0
13,548
16,090,323,496
IssuesEvent
2021-04-26 15:55:24
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
remove depricated jenkins jobs from set pending
kind/process lifecycle/stale priority/important-longterm
for example windows virtualbox podman ,,, these still are being set as pending,even though we dont run them
1.0
remove depricated jenkins jobs from set pending - for example windows virtualbox podman ,,, these still are being set as pending,even though we dont run them
process
remove depricated jenkins jobs from set pending for example windows virtualbox podman these still are being set as pending even though we dont run them
1
393,762
11,624,364,335
IssuesEvent
2020-02-27 10:38:04
dhowe/spectre
https://api.github.com/repos/dhowe/spectre
closed
Add circular cutout to profile image in templates/email.html
med-priority needs-verification
this is static html, just need the image to look like it does for avatars on the main site you can test in templates/output.html (a generated file), then move the code to templates/email.html (the template file) when ready if finished, check for new (unassigned) tickets from @billposters - he was testing last night (post-experience is still on-hold) ![image](https://user-images.githubusercontent.com/737638/75372405-76bb9400-5903-11ea-88c3-cc198609d599.png)
1.0
Add circular cutout to profile image in templates/email.html - this is static html, just need the image to look like it does for avatars on the main site you can test in templates/output.html (a generated file), then move the code to templates/email.html (the template file) when ready if finished, check for new (unassigned) tickets from @billposters - he was testing last night (post-experience is still on-hold) ![image](https://user-images.githubusercontent.com/737638/75372405-76bb9400-5903-11ea-88c3-cc198609d599.png)
non_process
add circular cutout to profile image in templates email html this is static html just need the image to look like it does for avatars on the main site you can test in templates output html a generated file then move the code to templates email html the template file when ready if finished check for new unassigned tickets from billposters he was testing last night post experience is still on hold
0
5,863
8,682,734,453
IssuesEvent
2018-12-02 11:41:34
bitshares/bitshares-community-ui
https://api.github.com/repos/bitshares/bitshares-community-ui
closed
Backup not functioning
Privatekey Backup bug process
when clicking on Backup (header) then on 'continue to backup' (the popup) then on 'I understand' button it wont take to the next step.
1.0
Backup not functioning - when clicking on Backup (header) then on 'continue to backup' (the popup) then on 'I understand' button it wont take to the next step.
process
backup not functioning when clicking on backup header then on continue to backup the popup then on i understand button it wont take to the next step
1
7,328
10,468,918,467
IssuesEvent
2019-09-22 17:02:37
produvia/ai-platform
https://api.github.com/repos/produvia/ai-platform
closed
Text Classification
natural-language-processing task wontfix
# Goal(s) - Assign a sentence or document an appropriate category # Input(s) - Sentence or document # Output(s) - Category # Objective Function(s) - TBD
1.0
Text Classification - # Goal(s) - Assign a sentence or document an appropriate category # Input(s) - Sentence or document # Output(s) - Category # Objective Function(s) - TBD
process
text classification goal s assign a sentence or document an appropriate category input s sentence or document output s category objective function s tbd
1
12,110
14,740,468,313
IssuesEvent
2021-01-07 09:08:16
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
RE: SA Billing - Save Credit Card to Account
anc-process anp-1 ant-bug
In GitLab by @kdjstudios on Nov 9, 2018, 11:52 **Submitted by:** "Arianna Screen" <arianna.screen@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-09-49358 **Server:** Internal **Client/Site:** NA **Account:** NA **Issue:** Today I am running credit cards and following the instructions to save the credit cards. I have all the information for the payment in and there is no check box to save the credit card information.
1.0
RE: SA Billing - Save Credit Card to Account - In GitLab by @kdjstudios on Nov 9, 2018, 11:52 **Submitted by:** "Arianna Screen" <arianna.screen@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-09-49358 **Server:** Internal **Client/Site:** NA **Account:** NA **Issue:** Today I am running credit cards and following the instructions to save the credit cards. I have all the information for the payment in and there is no check box to save the credit card information.
process
re sa billing save credit card to account in gitlab by kdjstudios on nov submitted by arianna screen helpdesk server internal client site na account na issue today i am running credit cards and following the instructions to save the credit cards i have all the information for the payment in and there is no check box to save the credit card information
1
12,854
21,012,322,589
IssuesEvent
2022-03-30 07:53:21
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
opened
[maven]org.apache.maven.plugins:maven-compiler-plugin get configuration/source and configuration/target
type:feature status:requirements priority-5-triage
### What would you like Renovate to be able to do? [springshell-rce-0-day-vulnerability](https://www.cyberkendra.com/2022/03/springshell-rce-0-day-vulnerability.html) Some vulnerability in the java ecosystem are based on the JDK version, org.apache.maven.plugins:maven-compiler-plugin configuration/source and configuration/target is so important for upgraded dependencies versions I hope the JDK version can be configured in the renovate configuration file with regexManagers ### If you have any ideas on how this should be implemented, please tell us here. I hope the JDK version can be configured in the renovate configuration file with regexManagers ### Is this a feature you are interested in implementing yourself? No
1.0
[maven]org.apache.maven.plugins:maven-compiler-plugin get configuration/source and configuration/target - ### What would you like Renovate to be able to do? [springshell-rce-0-day-vulnerability](https://www.cyberkendra.com/2022/03/springshell-rce-0-day-vulnerability.html) Some vulnerability in the java ecosystem are based on the JDK version, org.apache.maven.plugins:maven-compiler-plugin configuration/source and configuration/target is so important for upgraded dependencies versions I hope the JDK version can be configured in the renovate configuration file with regexManagers ### If you have any ideas on how this should be implemented, please tell us here. I hope the JDK version can be configured in the renovate configuration file with regexManagers ### Is this a feature you are interested in implementing yourself? No
non_process
org apache maven plugins maven compiler plugin get configuration source and configuration target what would you like renovate to be able to do some vulnerability in the java ecosystem are based on the jdk version org apache maven plugins maven compiler plugin configuration source and configuration target is so important for upgraded dependencies versions i hope the jdk version can be configured in the renovate configuration file with regexmanagers if you have any ideas on how this should be implemented please tell us here i hope the jdk version can be configured in the renovate configuration file with regexmanagers is this a feature you are interested in implementing yourself no
0
13,343
15,801,515,409
IssuesEvent
2021-04-03 05:17:52
PyCQA/flake8
https://api.github.com/repos/PyCQA/flake8
closed
Flake8 3.0 does not work, throws AttributeError exception
bug:confirmed component:multiprocessing component:pyflakes priority:high
In GitLab by @akittas on Jul 25, 2016, 08:52 Python version: 3.5.2 64-bit OS: Windows 10 Pro 64 bit Installation: pip install flake8 pip version: 8.1.2 setuptools version: 25.0.1 flake 8 --version: 3.0.0 (pycodestyle: 2.0.0, mccabe: 0.5.0, pyflakes: 1.2.3) CPython 3.5.2 on Windows running method: flake8 program.py or python -m flake8 program.py Flake8 3.0 throws the following exception when executing on any file: **AttributeError: 'FlakesChecker' object has no attribute 'with_doctest'** ``` Traceback (most recent call last): File "c:\programs\python35\lib\multiprocessing\process.py", line 249, in _bootstrap self.run() File "c:\programs\python35\lib\multiprocessing\process.py", line 93, in run self._target(*self._args, **self._kwargs) File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 222, in _run_checks_from_queue checker.run_checks(self.results_queue, self.statistics_queue) File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 563, in run_checks self.run_ast_checks() File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 468, in run_ast_checks checker = self.run_check(plugin, tree=ast) File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 448, in run_check return plugin.execute(**arguments) File "c:\programs\python35\lib\site-packages\flake8\plugins\manager.py", line 126, in execute return self.plugin(*args, **kwargs) # pylint: disable=not-callable File "c:\programs\python35\lib\site-packages\flake8\plugins\pyflakes.py", line 53, in __init__ with_doctest = self.with_doctest AttributeError: 'FlakesChecker' object has no attribute 'with_doctest' ``` Flake8 2.6.2 works as expected.
1.0
Flake8 3.0 does not work, throws AttributeError exception - In GitLab by @akittas on Jul 25, 2016, 08:52 Python version: 3.5.2 64-bit OS: Windows 10 Pro 64 bit Installation: pip install flake8 pip version: 8.1.2 setuptools version: 25.0.1 flake 8 --version: 3.0.0 (pycodestyle: 2.0.0, mccabe: 0.5.0, pyflakes: 1.2.3) CPython 3.5.2 on Windows running method: flake8 program.py or python -m flake8 program.py Flake8 3.0 throws the following exception when executing on any file: **AttributeError: 'FlakesChecker' object has no attribute 'with_doctest'** ``` Traceback (most recent call last): File "c:\programs\python35\lib\multiprocessing\process.py", line 249, in _bootstrap self.run() File "c:\programs\python35\lib\multiprocessing\process.py", line 93, in run self._target(*self._args, **self._kwargs) File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 222, in _run_checks_from_queue checker.run_checks(self.results_queue, self.statistics_queue) File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 563, in run_checks self.run_ast_checks() File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 468, in run_ast_checks checker = self.run_check(plugin, tree=ast) File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 448, in run_check return plugin.execute(**arguments) File "c:\programs\python35\lib\site-packages\flake8\plugins\manager.py", line 126, in execute return self.plugin(*args, **kwargs) # pylint: disable=not-callable File "c:\programs\python35\lib\site-packages\flake8\plugins\pyflakes.py", line 53, in __init__ with_doctest = self.with_doctest AttributeError: 'FlakesChecker' object has no attribute 'with_doctest' ``` Flake8 2.6.2 works as expected.
process
does not work throws attributeerror exception in gitlab by akittas on jul python version bit os windows pro bit installation pip install pip version setuptools version flake version pycodestyle mccabe pyflakes cpython on windows running method program py or python m program py throws the following exception when executing on any file attributeerror flakeschecker object has no attribute with doctest traceback most recent call last file c programs lib multiprocessing process py line in bootstrap self run file c programs lib multiprocessing process py line in run self target self args self kwargs file c programs lib site packages checker py line in run checks from queue checker run checks self results queue self statistics queue file c programs lib site packages checker py line in run checks self run ast checks file c programs lib site packages checker py line in run ast checks checker self run check plugin tree ast file c programs lib site packages checker py line in run check return plugin execute arguments file c programs lib site packages plugins manager py line in execute return self plugin args kwargs pylint disable not callable file c programs lib site packages plugins pyflakes py line in init with doctest self with doctest attributeerror flakeschecker object has no attribute with doctest works as expected
1
452,912
32,073,573,391
IssuesEvent
2023-09-25 09:32:47
dappnode/DAppNodeDocs
https://api.github.com/repos/dappnode/DAppNodeDocs
closed
New documentation tree
documentation 2023
The new documentation will follow this tree: 1. User docs 2. Dev docs 3. Smooth 4. DAO (show warning about deprecation or remove) Inside user docs we should include: 1. Get started a. Access Dappnode (via WiFi) b. Set up Register Repository (IPFS + Ethereum) Host password WiFi password c. Discover the features: Become a validator Install blockchain node packages .... 5. Access my Dappnode a. WiFi b. VPN (local and remote) c. Local (recovery) d. ssh (advanced) 6. Install Dappnode a. ISO b. script 7. Staking a. Solo staking: Network Clients Signer MEV Boost Smoothing Pool b. Rocketpool c. Stakehouse d. Stakewise (when published) 8. Videos & Tutorials 9. FAQs (get the ones we already have from Discord) 10. Packages (highlight some like HOPR) 11. Ethical Metrics We should also: 1. Embed videos everywhere they are needed 2. Show highlighted items in documentation index 3. Enable contributions and make it easy for anybody to add new information 4. Be very verbose in clients and speak about the diversity
1.0
New documentation tree - The new documentation will follow this tree: 1. User docs 2. Dev docs 3. Smooth 4. DAO (show warning about deprecation or remove) Inside user docs we should include: 1. Get started a. Access Dappnode (via WiFi) b. Set up Register Repository (IPFS + Ethereum) Host password WiFi password c. Discover the features: Become a validator Install blockchain node packages .... 5. Access my Dappnode a. WiFi b. VPN (local and remote) c. Local (recovery) d. ssh (advanced) 6. Install Dappnode a. ISO b. script 7. Staking a. Solo staking: Network Clients Signer MEV Boost Smoothing Pool b. Rocketpool c. Stakehouse d. Stakewise (when published) 8. Videos & Tutorials 9. FAQs (get the ones we already have from Discord) 10. Packages (highlight some like HOPR) 11. Ethical Metrics We should also: 1. Embed videos everywhere they are needed 2. Show highlighted items in documentation index 3. Enable contributions and make it easy for anybody to add new information 4. Be very verbose in clients and speak about the diversity
non_process
new documentation tree the new documentation will follow this tree user docs dev docs smooth dao show warning about deprecation or remove inside user docs we should include get started a access dappnode via wifi b set up register repository ipfs ethereum host password wifi password c discover the features become a validator install blockchain node packages access my dappnode a wifi b vpn local and remote c local recovery d ssh advanced install dappnode a iso b script staking a solo staking network clients signer mev boost smoothing pool b rocketpool c stakehouse d stakewise when published videos tutorials faqs get the ones we already have from discord packages highlight some like hopr ethical metrics we should also embed videos everywhere they are needed show highlighted items in documentation index enable contributions and make it easy for anybody to add new information be very verbose in clients and speak about the diversity
0
3,210
9,213,817,246
IssuesEvent
2019-03-10 15:09:07
jimkyndemeyer/js-graphql-intellij-plugin
https://api.github.com/repos/jimkyndemeyer/js-graphql-intellij-plugin
closed
Support for multiple schemas (on multiple endpoints)
enhancement v2-alpha v2-architecture
I need to access two different graphql endpoints from a project, where each endpoint has a different schema. I'm currently keeping two different copies of graphql.config.json that I swap between as needed. Is there a better way to handle this? If not, consider this an enhancement request.
1.0
Support for multiple schemas (on multiple endpoints) - I need to access two different graphql endpoints from a project, where each endpoint has a different schema. I'm currently keeping two different copies of graphql.config.json that I swap between as needed. Is there a better way to handle this? If not, consider this an enhancement request.
non_process
support for multiple schemas on multiple endpoints i need to access two different graphql endpoints from a project where each endpoint has a different schema i m currently keeping two different copies of graphql config json that i swap between as needed is there a better way to handle this if not consider this an enhancement request
0
49,212
6,157,458,330
IssuesEvent
2017-06-28 18:59:45
mapzen/android
https://api.github.com/repos/mapzen/android
opened
Sample App Major Section: More Info
Sample App Redesign
- [ ] Links to the places demo - [ ] Release info - [ ] Feature list - [ ] Contact - [ ] Participate - [ ] Download - [ ] API keys
1.0
Sample App Major Section: More Info - - [ ] Links to the places demo - [ ] Release info - [ ] Feature list - [ ] Contact - [ ] Participate - [ ] Download - [ ] API keys
non_process
sample app major section more info links to the places demo release info feature list contact participate download api keys
0
514,457
14,939,551,731
IssuesEvent
2021-01-25 17:05:05
idaholab/raven
https://api.github.com/repos/idaholab/raven
closed
[TASK] ROMCollection.Interpolated truncated lifetime
FutureRAVENv2.1 priority_minor task
-------- Issue Description -------- For workflow debugging, an option to limit the number of cycles sampled in an Interpolated ROMCollection would be useful. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or task? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
1.0
[TASK] ROMCollection.Interpolated truncated lifetime - -------- Issue Description -------- For workflow debugging, an option to limit the number of cycles sampled in an Interpolated ROMCollection would be useful. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or task? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
non_process
romcollection interpolated truncated lifetime issue description for workflow debugging an option to limit the number of cycles sampled in an interpolated romcollection would be useful for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
0
133,660
18,299,017,408
IssuesEvent
2021-10-05 23:53:38
bsbtd/Teste
https://api.github.com/repos/bsbtd/Teste
opened
CVE-2020-11112 (High) detected in jackson-databind-2.9.5.jar, jackson-databind-2.6.7.3.jar
security vulnerability
## CVE-2020-11112 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.6.7.3.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Teste/liferay-portal/modules/etl/talend/talend-runtime/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - components-api-0.25.3.jar (Root Library) - daikon-0.27.0.jar - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.6.7.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Teste/opentelemetry-java/sdk_contrib/aws_v1_support/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.7.3/c8bad9f57a5d298280f8bc0efcb7bf8393a41534/jackson-databind-2.6.7.3.jar</p> <p> Dependency Hierarchy: - aws-java-sdk-ec2-1.11.701.jar (Root Library) - jmespath-java-1.11.701.jar - :x: **jackson-databind-2.6.7.3.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.proxy.provider.remoting.RmiProvider (aka apache/commons-proxy). <p>Publish Date: 2020-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11112>CVE-2020-11112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112</a></p> <p>Release Date: 2020-03-31</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11112 (High) detected in jackson-databind-2.9.5.jar, jackson-databind-2.6.7.3.jar - ## CVE-2020-11112 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.6.7.3.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Teste/liferay-portal/modules/etl/talend/talend-runtime/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - components-api-0.25.3.jar (Root Library) - daikon-0.27.0.jar - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.6.7.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Teste/opentelemetry-java/sdk_contrib/aws_v1_support/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.7.3/c8bad9f57a5d298280f8bc0efcb7bf8393a41534/jackson-databind-2.6.7.3.jar</p> <p> Dependency Hierarchy: - aws-java-sdk-ec2-1.11.701.jar (Root Library) - jmespath-java-1.11.701.jar - :x: **jackson-databind-2.6.7.3.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.proxy.provider.remoting.RmiProvider (aka apache/commons-proxy). <p>Publish Date: 2020-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11112>CVE-2020-11112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112</a></p> <p>Release Date: 2020-03-31</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar jackson databind jar cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file teste liferay portal modules etl talend talend runtime pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy components api jar root library daikon jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file teste opentelemetry java sdk contrib aws support build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy aws java sdk jar root library jmespath java jar x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons proxy provider remoting rmiprovider aka apache commons proxy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
21,442
29,478,655,563
IssuesEvent
2023-06-02 02:09:32
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
'webpack-preprocessor': project is missing dependency '@types/bluebird'
type: typings npm: @cypress/webpack-preprocessor stale
### Current behavior Cypress project type-checking fails if we try to type-check `cypress/plugins/index.js` and it includes `require('@cypress/webpack-preprocessor')` ![image](https://user-images.githubusercontent.com/1393142/101527078-ef0bb700-3963-11eb-9d89-c97f0cd70d24.png) ### Desired behavior Type-checking succeeds ### Test code to reproduce ```js // This function is called when a project is opened or re-opened (e.g. due to the project's config // changing). // read more: https://on.cypress.io/plugins-guide const webpackPreprocessor = require('@cypress/webpack-preprocessor'); /** @type {Cypress.PluginConfig} */ module.exports = (on, config) => { on( 'file:preprocessor', webpackPreprocessor({ webpackOptions: {}, }), ); return config; }; ``` ### Versions Introduced in https://github.com/cypress-io/cypress-webpack-preprocessor/pull/83/files#diff-dcdc3e0b3362edb8fec2a51d3fa51f8fb8af8f70247e06d9887fa934834c9122R2 (merged in v5.2.1) From looking around, it seems that `@types/bluebird` is not provided as a dependency in any of the cypress packages. It seems weird to me that users need to figure out on their own that this dependency is needed. If usage of `@cypress/webpack-preprocessor` _requires_ a `@types/` module to type-check, I would expect it to be in `dependencies` and installed automatically.
1.0
'webpack-preprocessor': project is missing dependency '@types/bluebird' - ### Current behavior Cypress project type-checking fails if we try to type-check `cypress/plugins/index.js` and it includes `require('@cypress/webpack-preprocessor')` ![image](https://user-images.githubusercontent.com/1393142/101527078-ef0bb700-3963-11eb-9d89-c97f0cd70d24.png) ### Desired behavior Type-checking succeeds ### Test code to reproduce ```js // This function is called when a project is opened or re-opened (e.g. due to the project's config // changing). // read more: https://on.cypress.io/plugins-guide const webpackPreprocessor = require('@cypress/webpack-preprocessor'); /** @type {Cypress.PluginConfig} */ module.exports = (on, config) => { on( 'file:preprocessor', webpackPreprocessor({ webpackOptions: {}, }), ); return config; }; ``` ### Versions Introduced in https://github.com/cypress-io/cypress-webpack-preprocessor/pull/83/files#diff-dcdc3e0b3362edb8fec2a51d3fa51f8fb8af8f70247e06d9887fa934834c9122R2 (merged in v5.2.1) From looking around, it seems that `@types/bluebird` is not provided as a dependency in any of the cypress packages. It seems weird to me that users need to figure out on their own that this dependency is needed. If usage of `@cypress/webpack-preprocessor` _requires_ a `@types/` module to type-check, I would expect it to be in `dependencies` and installed automatically.
process
webpack preprocessor project is missing dependency types bluebird current behavior cypress project type checking fails if we try to type check cypress plugins index js and it includes require cypress webpack preprocessor desired behavior type checking succeeds test code to reproduce js this function is called when a project is opened or re opened e g due to the project s config changing read more const webpackpreprocessor require cypress webpack preprocessor type cypress pluginconfig module exports on config on file preprocessor webpackpreprocessor webpackoptions return config versions introduced in merged in from looking around it seems that types bluebird is not provided as a dependency in any of the cypress packages it seems weird to me that users need to figure out on their own that this dependency is needed if usage of cypress webpack preprocessor requires a types module to type check i would expect it to be in dependencies and installed automatically
1
347,977
24,905,285,730
IssuesEvent
2022-10-29 06:42:41
wmo-im/wis2box
https://api.github.com/repos/wmo-im/wis2box
closed
add installation by VM instructions
documentation
Per discussion/presentation with @golfvert, add documentation for installing wis2box by way of VM (Proxmox).
1.0
add installation by VM instructions - Per discussion/presentation with @golfvert, add documentation for installing wis2box by way of VM (Proxmox).
non_process
add installation by vm instructions per discussion presentation with golfvert add documentation for installing by way of vm proxmox
0
21,144
28,121,790,020
IssuesEvent
2023-03-31 14:42:04
MPMG-DCC-UFMG/C01
https://api.github.com/repos/MPMG-DCC-UFMG/C01
closed
Post mortem DEBUG - Trace Viewer
[2] Alta Prioridade [1] Requisito [0] Desenvolvimento [3] Processamento Dinâmico
## Comportamento Esperado Desejamos incluir o "Trace Viewer" do Playwright como ferramenta de debug para coletores dinâmicos. É uma funcionalidade com uma ótima interface que permite a execução passo a passo de um coletor dinâmico, permitindo a visualização das páginas, snapshots, e inspeção do código fonte de cada instante. ## Comportamento Atual São necessárias ferramentas para debug de coletores, principalmente os que utilizam o mecanismo de passos. O Trace Viewer poderá ser utilizado pela equipe para investigar bugs do sistema. Os usuários de outras trilhas também podem se beneficiar, usando-o para depurar seus próprios coletores e para obterem mais informações sobre futuros erros a serem reportados para nossa equipe. ## Passos para reproduzir o erro Não se aplica. ## Especificações da Coleta Não se aplica ## Sistema (caso necessário) Não se aplica ## Screenshots (caso necessário) Uma amostra da interface do Trace Viewer: ![image](https://user-images.githubusercontent.com/48096245/182942132-d2acd8a9-e053-43dd-9db1-ba07a8077b9d.png) Mais explicações sobre a ferramenta aqui: https://github.com/mpmg-dcc-ufmg/c01/issues/2333
1.0
Post mortem DEBUG - Trace Viewer - ## Comportamento Esperado Desejamos incluir o "Trace Viewer" do Playwright como ferramenta de debug para coletores dinâmicos. É uma funcionalidade com uma ótima interface que permite a execução passo a passo de um coletor dinâmico, permitindo a visualização das páginas, snapshots, e inspeção do código fonte de cada instante. ## Comportamento Atual São necessárias ferramentas para debug de coletores, principalmente os que utilizam o mecanismo de passos. O Trace Viewer poderá ser utilizado pela equipe para investigar bugs do sistema. Os usuários de outras trilhas também podem se beneficiar, usando-o para depurar seus próprios coletores e para obterem mais informações sobre futuros erros a serem reportados para nossa equipe. ## Passos para reproduzir o erro Não se aplica. ## Especificações da Coleta Não se aplica ## Sistema (caso necessário) Não se aplica ## Screenshots (caso necessário) Uma amostra da interface do Trace Viewer: ![image](https://user-images.githubusercontent.com/48096245/182942132-d2acd8a9-e053-43dd-9db1-ba07a8077b9d.png) Mais explicações sobre a ferramenta aqui: https://github.com/mpmg-dcc-ufmg/c01/issues/2333
process
post mortem debug trace viewer comportamento esperado desejamos incluir o trace viewer do playwright como ferramenta de debug para coletores dinâmicos é uma funcionalidade com uma ótima interface que permite a execução passo a passo de um coletor dinâmico permitindo a visualização das páginas snapshots e inspeção do código fonte de cada instante comportamento atual são necessárias ferramentas para debug de coletores principalmente os que utilizam o mecanismo de passos o trace viewer poderá ser utilizado pela equipe para investigar bugs do sistema os usuários de outras trilhas também podem se beneficiar usando o para depurar seus próprios coletores e para obterem mais informações sobre futuros erros a serem reportados para nossa equipe passos para reproduzir o erro não se aplica especificações da coleta não se aplica sistema caso necessário não se aplica screenshots caso necessário uma amostra da interface do trace viewer mais explicações sobre a ferramenta aqui
1
313,442
26,931,362,950
IssuesEvent
2023-02-07 17:03:22
dotnet/source-build
https://api.github.com/repos/dotnet/source-build
closed
Remove security-partners-dotnet pipeline from main branch
area-ci-testing
It's my understanding that the [security-partners-dotnet.yml](https://github.com/dotnet/installer/blob/main/src/SourceBuild/content/eng/pipelines/security-partners-dotnet.yml) pipeline is not used in the main branch and only used for .NET 6/7 servicing. If so, then this file should be removed from the main branch.
1.0
Remove security-partners-dotnet pipeline from main branch - It's my understanding that the [security-partners-dotnet.yml](https://github.com/dotnet/installer/blob/main/src/SourceBuild/content/eng/pipelines/security-partners-dotnet.yml) pipeline is not used in the main branch and only used for .NET 6/7 servicing. If so, then this file should be removed from the main branch.
non_process
remove security partners dotnet pipeline from main branch it s my understanding that the pipeline is not used in the main branch and only used for net servicing if so then this file should be removed from the main branch
0
102,065
11,273,465,603
IssuesEvent
2020-01-14 16:36:40
GluuFederation/casa
https://api.github.com/repos/GluuFederation/casa
closed
Enable Plugins Isolated Way to Persist Configuration
Needs Documentation Needs QA enhancement
Plugin devs should have the possibility to store relevant configurations in the same database attribute the application uses for settings. It's good to have a single point of configuration We have to offer a safe means so that developers cannot read or write other plugins configs or core configs
1.0
Enable Plugins Isolated Way to Persist Configuration - Plugin devs should have the possibility to store relevant configurations in the same database attribute the application uses for settings. It's good to have a single point of configuration We have to offer a safe means so that developers cannot read or write other plugins configs or core configs
non_process
enable plugins isolated way to persist configuration plugin devs should have the possibility to store relevant configurations in the same database attribute the application uses for settings it s good to have a single point of configuration we have to offer a safe means so that developers cannot read or write other plugins configs or core configs
0
8,492
11,647,929,489
IssuesEvent
2020-03-01 17:49:56
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Assert failure currentProcessCpuCount == g_processAffinitySet.Count()
area-System.Diagnostics.Process untriaged
https://helix.dot.net/api/2019-06-17/jobs/5c24c550-4c42-459f-bbfd-21e69cbf8a27/workitems/System.Diagnostics.Process.Tests/console `Libraries Test Run checked coreclr Linux x64 Debug ` ``` Assert failure(PID 15210 [0x00003b6a], Thread: 15210 [0x3b6a]): currentProcessCpuCount == g_processAffinitySet.Count() File: /__w/1/s/src/coreclr/src/vm/gcenv.os.cpp Line: 123 Image: /home/helixbot/work/AB3D09A7/p/dotnet System.Diagnostics.Tests.ProcessTests.TestProcessorAffinity [FAIL] System.InvalidOperationException : Cannot process request because the process (15210) has exited. Stack Trace: System.Diagnostics.Process/src/System/Diagnostics/Process.Unix.cs(343,0): at System.Diagnostics.Process.ThrowIfExited(Boolean refresh) System.Diagnostics.Process/src/System/Diagnostics/Process.cs(932,0): at System.Diagnostics.Process.EnsureState(State state) System.Diagnostics.Process/src/System/Diagnostics/Process.Linux.cs(191,0): at System.Diagnostics.Process.set_ProcessorAffinityCore(IntPtr value) System.Diagnostics.Process/src/System/Diagnostics/Process.cs(532,0): at System.Diagnostics.Process.set_ProcessorAffinity(IntPtr value) System.Diagnostics.Process/tests/ProcessTests.cs(861,0): at System.Diagnostics.Tests.ProcessTests.TestProcessorAffinity() System.Diagnostics.Tests.ProcessTests.TestProcessRecycledPid [SKIP] Condition(s) not met: "IsStressModeEnabled" Finished: System.Diagnostics.Process.Tests ``` Seen in https://github.com/dotnet/runtime/pull/32538
1.0
Assert failure currentProcessCpuCount == g_processAffinitySet.Count() - https://helix.dot.net/api/2019-06-17/jobs/5c24c550-4c42-459f-bbfd-21e69cbf8a27/workitems/System.Diagnostics.Process.Tests/console `Libraries Test Run checked coreclr Linux x64 Debug ` ``` Assert failure(PID 15210 [0x00003b6a], Thread: 15210 [0x3b6a]): currentProcessCpuCount == g_processAffinitySet.Count() File: /__w/1/s/src/coreclr/src/vm/gcenv.os.cpp Line: 123 Image: /home/helixbot/work/AB3D09A7/p/dotnet System.Diagnostics.Tests.ProcessTests.TestProcessorAffinity [FAIL] System.InvalidOperationException : Cannot process request because the process (15210) has exited. Stack Trace: System.Diagnostics.Process/src/System/Diagnostics/Process.Unix.cs(343,0): at System.Diagnostics.Process.ThrowIfExited(Boolean refresh) System.Diagnostics.Process/src/System/Diagnostics/Process.cs(932,0): at System.Diagnostics.Process.EnsureState(State state) System.Diagnostics.Process/src/System/Diagnostics/Process.Linux.cs(191,0): at System.Diagnostics.Process.set_ProcessorAffinityCore(IntPtr value) System.Diagnostics.Process/src/System/Diagnostics/Process.cs(532,0): at System.Diagnostics.Process.set_ProcessorAffinity(IntPtr value) System.Diagnostics.Process/tests/ProcessTests.cs(861,0): at System.Diagnostics.Tests.ProcessTests.TestProcessorAffinity() System.Diagnostics.Tests.ProcessTests.TestProcessRecycledPid [SKIP] Condition(s) not met: "IsStressModeEnabled" Finished: System.Diagnostics.Process.Tests ``` Seen in https://github.com/dotnet/runtime/pull/32538
process
assert failure currentprocesscpucount g processaffinityset count libraries test run checked coreclr linux debug assert failure pid thread currentprocesscpucount g processaffinityset count file w s src coreclr src vm gcenv os cpp line image home helixbot work p dotnet system diagnostics tests processtests testprocessoraffinity system invalidoperationexception cannot process request because the process has exited stack trace system diagnostics process src system diagnostics process unix cs at system diagnostics process throwifexited boolean refresh system diagnostics process src system diagnostics process cs at system diagnostics process ensurestate state state system diagnostics process src system diagnostics process linux cs at system diagnostics process set processoraffinitycore intptr value system diagnostics process src system diagnostics process cs at system diagnostics process set processoraffinity intptr value system diagnostics process tests processtests cs at system diagnostics tests processtests testprocessoraffinity system diagnostics tests processtests testprocessrecycledpid condition s not met isstressmodeenabled finished system diagnostics process tests seen in
1
87,382
25,107,004,353
IssuesEvent
2022-11-08 17:19:58
spack/spack
https://api.github.com/repos/spack/spack
closed
Installation issue: Binary libffi fails checksum verification
build-error
### Steps to reproduce the issue ``` # On develop branch spack mirror add binary_mirror https://binaries.spack.io/releases/v0.18 spack buildcache keys --install --trust spack -vvv install /d6d3lh3 ``` ### Error message ==> Installing libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd ==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spec.json.sig gpg: Signature made Tue 31 May 2022 07:27:40 AM UTC using RSA key ID 3DB0C723 gpg: Good signature from "Spack Project Official Binaries <maintainers@spack.io>" ==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4/gcc-7.3.1/libffi-3.4.2/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spack ==> Extracting libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd from binary cache ==> Error: Failed to install libffi due to NoChecksumException: Package tarball failed checksum verification. It cannot be installed. ### Information on your system [ec2-user@ip-10-7-228-156 ~]$ spack debug report * **Spack:** 0.19.0.dev0 (b065d69136023693ace8e4d35d9fdad817d63d93) * **Python:** 3.7.10 * **Platform:** linux-amzn2-skylake_avx512 * **Concretizer:** clingo ### Additional information _No response_ ### General information - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded the build log and environment files - [X] I have searched the issues of this repo and believe this is not a duplicate
1.0
Installation issue: Binary libffi fails checksum verification - ### Steps to reproduce the issue ``` # On develop branch spack mirror add binary_mirror https://binaries.spack.io/releases/v0.18 spack buildcache keys --install --trust spack -vvv install /d6d3lh3 ``` ### Error message ==> Installing libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd ==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spec.json.sig gpg: Signature made Tue 31 May 2022 07:27:40 AM UTC using RSA key ID 3DB0C723 gpg: Good signature from "Spack Project Official Binaries <maintainers@spack.io>" ==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4/gcc-7.3.1/libffi-3.4.2/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spack ==> Extracting libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd from binary cache ==> Error: Failed to install libffi due to NoChecksumException: Package tarball failed checksum verification. It cannot be installed. ### Information on your system [ec2-user@ip-10-7-228-156 ~]$ spack debug report * **Spack:** 0.19.0.dev0 (b065d69136023693ace8e4d35d9fdad817d63d93) * **Python:** 3.7.10 * **Platform:** linux-amzn2-skylake_avx512 * **Concretizer:** clingo ### Additional information _No response_ ### General information - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded the build log and environment files - [X] I have searched the issues of this repo and believe this is not a duplicate
non_process
installation issue binary libffi fails checksum verification steps to reproduce the issue on develop branch spack mirror add binary mirror spack buildcache keys install trust spack vvv install error message installing libffi fetching gpg signature made tue may am utc using rsa key id gpg good signature from spack project official binaries fetching extracting libffi from binary cache error failed to install libffi due to nochecksumexception package tarball failed checksum verification it cannot be installed information on your system spack debug report spack python platform linux skylake concretizer clingo additional information no response general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
0
17,611
23,430,202,294
IssuesEvent
2022-08-15 00:46:00
sparc4-dev/astropop
https://api.github.com/repos/sparc4-dev/astropop
closed
Avoid breaking in image registration lists
bug image-processing
When the image registration fails for a list, it breaks. Add option to ignore and log the errors only. May be warn it on header too.
1.0
Avoid breaking in image registration lists - When the image registration fails for a list, it breaks. Add option to ignore and log the errors only. May be warn it on header too.
process
avoid breaking in image registration lists when the image registration fails for a list it breaks add option to ignore and log the errors only may be warn it on header too
1
17,554
23,368,077,419
IssuesEvent
2022-08-10 17:07:30
nghi-huynh/HPA_HuBMAP_Kaggle_competition
https://api.github.com/repos/nghi-huynh/HPA_HuBMAP_Kaggle_competition
closed
WSI Pre-processing: tiling + tissue segmentation
data preprocessing data preparation
WSI preprocessing: Tiling + Tissue Segmentation **Goal**: using simple thresholding techniques like Otsu or Triangle binarization to identify tissue/non-tissue in tiles and save it for training purpose Refer to [WSI processing: Tiling + Tissue Segmentation](https://www.kaggle.com/code/nghihuynh/wsi-preprocessing-tiling-tissue-segmentation) for more detail - [x] Rescale - [x] Thresholding - [x] Tile + Selection - [x] Save
1.0
WSI Pre-processing: tiling + tissue segmentation - WSI preprocessing: Tiling + Tissue Segmentation **Goal**: using simple thresholding techniques like Otsu or Triangle binarization to identify tissue/non-tissue in tiles and save it for training purpose Refer to [WSI processing: Tiling + Tissue Segmentation](https://www.kaggle.com/code/nghihuynh/wsi-preprocessing-tiling-tissue-segmentation) for more detail - [x] Rescale - [x] Thresholding - [x] Tile + Selection - [x] Save
process
wsi pre processing tiling tissue segmentation wsi preprocessing tiling tissue segmentation goal using simple thresholding techniques like otsu or triangle binarization to identify tissue non tissue in tiles and save it for training purpose refer to for more detail rescale thresholding tile selection save
1
8,007
7,106,350,087
IssuesEvent
2018-01-16 16:27:06
OpenLiberty/open-liberty
https://api.github.com/repos/OpenLiberty/open-liberty
opened
Generate keystore.password during server create
Epic in:Security story team:Zombie Apocalypse
Implements the resolution of design issue #1175 - During `server create` time [generate a random secure password](http://crunchify.com/java-generate-strong-random-password-securerandom/) into a `keystore.password` property in `${server.config.dir}/server.env` - We will add a new flag to server create (e.g. `server create myServer --no-password`) which will disable the default random password generation. This will be helpful for situations where tools wish to generate their own value for `keystore.password`, or simply do not wish to generate one. - Priorities to override keystore password: 1. Use what's configured in the `<keyStore password="..."/>` field in server.xml 2. Use the `keystore.password` value set in the user environment 3. Use the `keystore.password` defined in server.env as the password value for the default keystore element<br><br> - The `<keyStore>` metatype will be updated as follows: ```xml <AD id="password" ... ibm:variable="env.keystore.password"/> ```
True
Generate keystore.password during server create - Implements the resolution of design issue #1175 - During `server create` time [generate a random secure password](http://crunchify.com/java-generate-strong-random-password-securerandom/) into a `keystore.password` property in `${server.config.dir}/server.env` - We will add a new flag to server create (e.g. `server create myServer --no-password`) which will disable the default random password generation. This will be helpful for situations where tools wish to generate their own value for `keystore.password`, or simply do not wish to generate one. - Priorities to override keystore password: 1. Use what's configured in the `<keyStore password="..."/>` field in server.xml 2. Use the `keystore.password` value set in the user environment 3. Use the `keystore.password` defined in server.env as the password value for the default keystore element<br><br> - The `<keyStore>` metatype will be updated as follows: ```xml <AD id="password" ... ibm:variable="env.keystore.password"/> ```
non_process
generate keystore password during server create implements the resolution of design issue during server create time into a keystore password property in server config dir server env we will add a new flag to server create e g server create myserver no password which will disable the default random password generation this will be helpful for situations where tools wish to generate their own value for keystore password or simply do not wish to generate one priorities to override keystore password use what s configured in the field in server xml use the keystore password value set in the user environment use the keystore password defined in server env as the password value for the default keystore element the metatype will be updated as follows xml
0
6,001
8,808,922,197
IssuesEvent
2018-12-27 16:54:37
linnovate/root
https://api.github.com/repos/linnovate/root
closed
office documents from tasks bug
2.0.6 Fixed Process bug critical
after creating a task, and then going to the documents tab, clicking on manage documents create new item doesnt update the list, and after editing the document it isnt saved after refreshing the page ![image](https://user-images.githubusercontent.com/38312178/50145494-65680800-02ba-11e9-96ac-36338e64ced9.png)
1.0
office documents from tasks bug - after creating a task, and then going to the documents tab, clicking on manage documents create new item doesnt update the list, and after editing the document it isnt saved after refreshing the page ![image](https://user-images.githubusercontent.com/38312178/50145494-65680800-02ba-11e9-96ac-36338e64ced9.png)
process
office documents from tasks bug after creating a task and then going to the documents tab clicking on manage documents create new item doesnt update the list and after editing the document it isnt saved after refreshing the page
1
13,254
15,725,718,794
IssuesEvent
2021-03-29 10:20:11
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
amazon-import post-processor handles multiple disks improperly
bug post-processor/amazon-import remote-plugin/amazon
#### Overview of the Issue In my case, I'm using the hyperv-iso builder and have entries within `disk_additional_size`, when this is passed on to the amazon-import post-processor it ends up finding the wrong disk. The code in question isn't particularly smart. It iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix (in my case vhdx). The issue with this, on Windows at least, is that `vm-0.vhdx` is found before the core image without the index suffix, i.e. `vm.vhdx` when iterating over the artefacts. for _, path := range artifact.Files() { if strings.HasSuffix(path, "."+p.config.Format) { source = path break } } It would make sense to perhaps specify which image you would like to be imported, rather than attempting to guess which one. As an additional improvement, Amazon's VM Import supports multiple disk images so this could be implemented (it would still make sense to add an optional config to specify disk images, perhaps using an explicit list or regex/glob). ### Packer version 1.5.1 (master) ### Operating system and Environment details Windows 10 Pro amd64
1.0
amazon-import post-processor handles multiple disks improperly - #### Overview of the Issue In my case, I'm using the hyperv-iso builder and have entries within `disk_additional_size`, when this is passed on to the amazon-import post-processor it ends up finding the wrong disk. The code in question isn't particularly smart. It iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix (in my case vhdx). The issue with this, on Windows at least, is that `vm-0.vhdx` is found before the core image without the index suffix, i.e. `vm.vhdx` when iterating over the artefacts. for _, path := range artifact.Files() { if strings.HasSuffix(path, "."+p.config.Format) { source = path break } } It would make sense to perhaps specify which image you would like to be imported, rather than attempting to guess which one. As an additional improvement, Amazon's VM Import supports multiple disk images so this could be implemented (it would still make sense to add an optional config to specify disk images, perhaps using an explicit list or regex/glob). ### Packer version 1.5.1 (master) ### Operating system and Environment details Windows 10 Pro amd64
process
amazon import post processor handles multiple disks improperly overview of the issue in my case i m using the hyperv iso builder and have entries within disk additional size when this is passed on to the amazon import post processor it ends up finding the wrong disk the code in question isn t particularly smart it iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix in my case vhdx the issue with this on windows at least is that vm vhdx is found before the core image without the index suffix i e vm vhdx when iterating over the artefacts for path range artifact files if strings hassuffix path p config format source path break it would make sense to perhaps specify which image you would like to be imported rather than attempting to guess which one as an additional improvement amazon s vm import supports multiple disk images so this could be implemented it would still make sense to add an optional config to specify disk images perhaps using an explicit list or regex glob packer version master operating system and environment details windows pro
1
104,515
13,095,183,229
IssuesEvent
2020-08-03 13:43:07
canonical-web-and-design/certification.ubuntu.com
https://api.github.com/repos/canonical-web-and-design/certification.ubuntu.com
closed
Hide "Certification Notes" when there are no notes to display.
Priority: Medium Redesign
Every certificate includes a "Certification Notes" statement that says by default "There are no notes for this release." It would be cleaner, and make more sense, to hide the "Certification Notes" section rather than displaying it only to say there are none.
1.0
Hide "Certification Notes" when there are no notes to display. - Every certificate includes a "Certification Notes" statement that says by default "There are no notes for this release." It would be cleaner, and make more sense, to hide the "Certification Notes" section rather than displaying it only to say there are none.
non_process
hide certification notes when there are no notes to display every certificate includes a certification notes statement that says by default there are no notes for this release it would be cleaner and make more sense to hide the certification notes section rather than displaying it only to say there are none
0
131,310
12,481,013,847
IssuesEvent
2020-05-29 21:26:42
Students-of-the-city-of-Kostroma/trpo_automation
https://api.github.com/repos/Students-of-the-city-of-Kostroma/trpo_automation
closed
Реализовать unit-тесты на функцию FormListWithLetters
Sprint 11 Story Testing documentation
Реализовать unit-тесты на функцию FormListWithLetters с основой на сценарии по [ссылке ](https://docs.google.com/spreadsheets/d/1IHaHZS0_vT2SRZCWRdLPDvhNNvPcKf5QjA6GztFQgF4/edit#gid=208599158)
1.0
Реализовать unit-тесты на функцию FormListWithLetters - Реализовать unit-тесты на функцию FormListWithLetters с основой на сценарии по [ссылке ](https://docs.google.com/spreadsheets/d/1IHaHZS0_vT2SRZCWRdLPDvhNNvPcKf5QjA6GztFQgF4/edit#gid=208599158)
non_process
реализовать unit тесты на функцию formlistwithletters реализовать unit тесты на функцию formlistwithletters с основой на сценарии по
0
326,735
24,099,051,895
IssuesEvent
2022-09-19 21:48:15
marcosdosea/AgendeMe
https://api.github.com/repos/marcosdosea/AgendeMe
closed
Atualizar collaborations na seção Design Model, "ManterProfissional", "ManterOrgaoPublico", "ManterServicoPublico"
documentation 2.0
# Objetivo ## Collaborations com a Persistência Atualizar a persistência para o novo artefato AgendeMeContext. Por segurança, basta apenas usar o comando de Delete(não precisa usar o comando Delete From Model) Resultado deve ser Antes: ![Image](https://user-images.githubusercontent.com/69280619/190529813-d5739409-7179-401d-a459-309e0d4205ca.png) Depois: ![Image](https://user-images.githubusercontent.com/69280619/190529964-be2ccd44-a4f2-44ff-97a2-83695e2c081c.png) Importante lembrar de utilizar os métodos oferecidos pela classe AgendeMeContext! Então para adicionar, utilizar comando Add e na sequência o Save Changes. O mesmo para as demais alterações, sejam de Remover(Delete), Add(Alterar/Adicionar), Select(consultar). O método Save Changes sempre irá ser executado após qualquer um dos comandos acimas, exceto o "Select"
1.0
Atualizar collaborations na seção Design Model, "ManterProfissional", "ManterOrgaoPublico", "ManterServicoPublico" - # Objetivo ## Collaborations com a Persistência Atualizar a persistência para o novo artefato AgendeMeContext. Por segurança, basta apenas usar o comando de Delete(não precisa usar o comando Delete From Model) Resultado deve ser Antes: ![Image](https://user-images.githubusercontent.com/69280619/190529813-d5739409-7179-401d-a459-309e0d4205ca.png) Depois: ![Image](https://user-images.githubusercontent.com/69280619/190529964-be2ccd44-a4f2-44ff-97a2-83695e2c081c.png) Importante lembrar de utilizar os métodos oferecidos pela classe AgendeMeContext! Então para adicionar, utilizar comando Add e na sequência o Save Changes. O mesmo para as demais alterações, sejam de Remover(Delete), Add(Alterar/Adicionar), Select(consultar). O método Save Changes sempre irá ser executado após qualquer um dos comandos acimas, exceto o "Select"
non_process
atualizar collaborations na seção design model manterprofissional manterorgaopublico manterservicopublico objetivo collaborations com a persistência atualizar a persistência para o novo artefato agendemecontext por segurança basta apenas usar o comando de delete não precisa usar o comando delete from model resultado deve ser antes depois importante lembrar de utilizar os métodos oferecidos pela classe agendemecontext então para adicionar utilizar comando add e na sequência o save changes o mesmo para as demais alterações sejam de remover delete add alterar adicionar select consultar o método save changes sempre irá ser executado após qualquer um dos comandos acimas exceto o select
0
325,893
24,065,186,588
IssuesEvent
2022-09-17 11:32:33
pyxem/orix
https://api.github.com/repos/pyxem/orix
closed
Restructure documentation into examples, tutorials and an improved API reference
documentation
I plan to restructure the orix documentation in a similar fashion as [I've done in kikuchipy](https://kikuchipy.org/en/latest/#learning-resources), based on the [Diátaxis](https://diataxis.fr/) framework. The changes I plan: * Separate the user guide into *Examples* and *Tutorials*. The examples are task oriented (no/minimal explanations), while the tutorials are learning oriented. * Examples are short recipies of common tasks in orix (e.g. #370, #364, #332) represented in a [sphinx-gallery](https://sphinx-gallery.github.io/stable/index.html) from Python scripts (not notebooks), very similar to Matplotlib's or scikit-images'. Whenever we have nice plot we've made or a procedure which we think might be useful to others, we can add an example to the appropriate section (or make a new section). This is better than trying to shoehorn more stuff into the appropriate tutorial. See the [kikuchipy examples](https://kikuchipy.org/en/latest/examples/index.html) for an example. * Citing Diátaxis, tutorials are *lessons that take the reader by the hand through a series of steps to complete a project of some kind.* These will be the existing user guides, built from notebooks with `nbsphinx`. See the [kikuchipy tutorials](https://kikuchipy.org/en/latest/tutorials/index.html) for an example. * The API reference will get an overhaul, with one listing the top modules, then one page per module listing classes and functions, then one page per class listing methods and attributes, then one page per method and attribute. These pages are automatically generated when building the docs with `make html`, and using custom Sphinx templates and some smart use of Sphinx, we shouldn't have to remember to list every method and attribute manually in `reference.rst`. The result is a reference that is simpler to navigate and lists the example(s) where this particular functionality is used. See the [kikuchipy API](https://kikuchipy.org/en/latest/reference/generated/kikuchipy.data.nickel_ebsd_master_pattern_small.html#kikuchipy.data.nickel_ebsd_master_pattern_small) for an example. I've done this once for kikuchipy, so shouldn't be too difficult to do for orix.
1.0
Restructure documentation into examples, tutorials and an improved API reference - I plan to restructure the orix documentation in a similar fashion as [I've done in kikuchipy](https://kikuchipy.org/en/latest/#learning-resources), based on the [Diátaxis](https://diataxis.fr/) framework. The changes I plan: * Separate the user guide into *Examples* and *Tutorials*. The examples are task oriented (no/minimal explanations), while the tutorials are learning oriented. * Examples are short recipies of common tasks in orix (e.g. #370, #364, #332) represented in a [sphinx-gallery](https://sphinx-gallery.github.io/stable/index.html) from Python scripts (not notebooks), very similar to Matplotlib's or scikit-images'. Whenever we have nice plot we've made or a procedure which we think might be useful to others, we can add an example to the appropriate section (or make a new section). This is better than trying to shoehorn more stuff into the appropriate tutorial. See the [kikuchipy examples](https://kikuchipy.org/en/latest/examples/index.html) for an example. * Citing Diátaxis, tutorials are *lessons that take the reader by the hand through a series of steps to complete a project of some kind.* These will be the existing user guides, built from notebooks with `nbsphinx`. See the [kikuchipy tutorials](https://kikuchipy.org/en/latest/tutorials/index.html) for an example. * The API reference will get an overhaul, with one listing the top modules, then one page per module listing classes and functions, then one page per class listing methods and attributes, then one page per method and attribute. These pages are automatically generated when building the docs with `make html`, and using custom Sphinx templates and some smart use of Sphinx, we shouldn't have to remember to list every method and attribute manually in `reference.rst`. The result is a reference that is simpler to navigate and lists the example(s) where this particular functionality is used. See the [kikuchipy API](https://kikuchipy.org/en/latest/reference/generated/kikuchipy.data.nickel_ebsd_master_pattern_small.html#kikuchipy.data.nickel_ebsd_master_pattern_small) for an example. I've done this once for kikuchipy, so shouldn't be too difficult to do for orix.
non_process
restructure documentation into examples tutorials and an improved api reference i plan to restructure the orix documentation in a similar fashion as based on the framework the changes i plan separate the user guide into examples and tutorials the examples are task oriented no minimal explanations while the tutorials are learning oriented examples are short recipies of common tasks in orix e g represented in a from python scripts not notebooks very similar to matplotlib s or scikit images whenever we have nice plot we ve made or a procedure which we think might be useful to others we can add an example to the appropriate section or make a new section this is better than trying to shoehorn more stuff into the appropriate tutorial see the for an example citing diátaxis tutorials are lessons that take the reader by the hand through a series of steps to complete a project of some kind these will be the existing user guides built from notebooks with nbsphinx see the for an example the api reference will get an overhaul with one listing the top modules then one page per module listing classes and functions then one page per class listing methods and attributes then one page per method and attribute these pages are automatically generated when building the docs with make html and using custom sphinx templates and some smart use of sphinx we shouldn t have to remember to list every method and attribute manually in reference rst the result is a reference that is simpler to navigate and lists the example s where this particular functionality is used see the for an example i ve done this once for kikuchipy so shouldn t be too difficult to do for orix
0
18,277
24,357,836,694
IssuesEvent
2022-10-03 09:03:54
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
Add command to `verdi` to "revive" dormant process
topic/verdi type/accepted feature priority/nice-to-have topic/processes
There is a bug where it is possible that the process task is lost with RabbitMQ before the process is completed. This will cause that the process is never picked up by the daemon and will be dormant. It will also not respond to RPC calls, such as `verdi process play` etc. There is a way to recreate the task that can revive the task ``` from aiida.manage import get_manager get_manager().get_process_controller().continue_process(PID) ``` We have had to communicate this multiple times to users during debugging. It was never added as an official method to the CLI since it is treating a symptom, and when used incorrectly can cause bigger problems, such as the process and its children being executed by multiple workers. Still, everytime we have to use it, having users copy paste it, is not ideal.
1.0
Add command to `verdi` to "revive" dormant process - There is a bug where it is possible that the process task is lost with RabbitMQ before the process is completed. This will cause that the process is never picked up by the daemon and will be dormant. It will also not respond to RPC calls, such as `verdi process play` etc. There is a way to recreate the task that can revive the task ``` from aiida.manage import get_manager get_manager().get_process_controller().continue_process(PID) ``` We have had to communicate this multiple times to users during debugging. It was never added as an official method to the CLI since it is treating a symptom, and when used incorrectly can cause bigger problems, such as the process and its children being executed by multiple workers. Still, everytime we have to use it, having users copy paste it, is not ideal.
process
add command to verdi to revive dormant process there is a bug where it is possible that the process task is lost with rabbitmq before the process is completed this will cause that the process is never picked up by the daemon and will be dormant it will also not respond to rpc calls such as verdi process play etc there is a way to recreate the task that can revive the task from aiida manage import get manager get manager get process controller continue process pid we have had to communicate this multiple times to users during debugging it was never added as an official method to the cli since it is treating a symptom and when used incorrectly can cause bigger problems such as the process and its children being executed by multiple workers still everytime we have to use it having users copy paste it is not ideal
1
11,537
14,410,655,161
IssuesEvent
2020-12-04 05:27:30
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
closed
1D-Collection of 1D-Data does not update display when changing "Index" slider in Inspector
f - displays f - processing f - sequences stage - planning type - bug
To reproduce, open a console and run: ```python show(numpy.random.rand(2, 32)) ``` Then go to `Processing -> Redimension Data ->Redimension to Collection of 2 Spectra of Length 32` Try changing the "Index" slider in the Inspector to see the second "slice" of data. The display does not update. This is with current master branch of nionswift.
1.0
1D-Collection of 1D-Data does not update display when changing "Index" slider in Inspector - To reproduce, open a console and run: ```python show(numpy.random.rand(2, 32)) ``` Then go to `Processing -> Redimension Data ->Redimension to Collection of 2 Spectra of Length 32` Try changing the "Index" slider in the Inspector to see the second "slice" of data. The display does not update. This is with current master branch of nionswift.
process
collection of data does not update display when changing index slider in inspector to reproduce open a console and run python show numpy random rand then go to processing redimension data redimension to collection of spectra of length try changing the index slider in the inspector to see the second slice of data the display does not update this is with current master branch of nionswift
1
107,255
11,524,588,977
IssuesEvent
2020-02-15 01:26:36
10quality/wpmvc-website
https://api.github.com/repos/10quality/wpmvc-website
opened
Gulp watch documentation
documentation
Add documentation for gulp commands: ```bash gulp watch gulp watch-js gulp watch-styles ```
1.0
Gulp watch documentation - Add documentation for gulp commands: ```bash gulp watch gulp watch-js gulp watch-styles ```
non_process
gulp watch documentation add documentation for gulp commands bash gulp watch gulp watch js gulp watch styles
0
12,192
14,742,313,034
IssuesEvent
2021-01-07 12:04:37
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Payment Revert Issue
anc-process anp-1 ant-bug ant-child/secondary ant-enhancement
In GitLab by @pchaudhary on Apr 8, 2019, 07:24 The same billing cycle payment is not reverting on the invoice if there are more than 1 invoice for the same billing cycle. * Let's say there are 2 invoices * I1 = $950 * I2 = $950 * Acc Bal is `$1900` * We made a payment of $ 950. (Manual Payment) * Acc Bal is `$950` * Revert the payment. * Balance on I1 is $0 * Balance on I2 is $950 * Acc Bal is `$1900`
1.0
Payment Revert Issue - In GitLab by @pchaudhary on Apr 8, 2019, 07:24 The same billing cycle payment is not reverting on the invoice if there are more than 1 invoice for the same billing cycle. * Let's say there are 2 invoices * I1 = $950 * I2 = $950 * Acc Bal is `$1900` * We made a payment of $ 950. (Manual Payment) * Acc Bal is `$950` * Revert the payment. * Balance on I1 is $0 * Balance on I2 is $950 * Acc Bal is `$1900`
process
payment revert issue in gitlab by pchaudhary on apr the same billing cycle payment is not reverting on the invoice if there are more than invoice for the same billing cycle let s say there are invoices acc bal is we made a payment of manual payment acc bal is revert the payment balance on is balance on is acc bal is
1
760,387
26,638,329,282
IssuesEvent
2023-01-25 00:41:23
sczerwinski/wavefront-obj-intellij-plugin
https://api.github.com/repos/sczerwinski/wavefront-obj-intellij-plugin
closed
Face culling setting
type:feature resolution:done priority:low component:3d component:settings
Add face culling selection in 3D preview settings from `graphics.glimpse.FaceCullingMode` enum. - `DISABLED` – disabled - `FRONT` – front-facing facets - `BACK` – back-facing facets - `FRONT_AND_BACK` – both front- and back-facing facets (does it make sense?)
1.0
Face culling setting - Add face culling selection in 3D preview settings from `graphics.glimpse.FaceCullingMode` enum. - `DISABLED` – disabled - `FRONT` – front-facing facets - `BACK` – back-facing facets - `FRONT_AND_BACK` – both front- and back-facing facets (does it make sense?)
non_process
face culling setting add face culling selection in preview settings from graphics glimpse facecullingmode enum disabled – disabled front – front facing facets back – back facing facets front and back – both front and back facing facets does it make sense
0
199,312
22,693,285,263
IssuesEvent
2022-07-05 01:07:57
Baneeishaque/printing_press_erp-Maven
https://api.github.com/repos/Baneeishaque/printing_press_erp-Maven
opened
CVE-2020-36518 (High) detected in jackson-databind-2.10.0.jar
security vulnerability
## CVE-2020-36518 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.10.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.0/jackson-databind-2.10.0.jar</p> <p> Dependency Hierarchy: - jasperreports-6.13.0.jar (Root Library) - :x: **jackson-databind-2.10.0.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects. Mend Note: After conducting further research, Mend has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518. <p>Publish Date: 2022-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-03-11</p> <p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.12.6.1</p> <p>Direct dependency fix Resolution (net.sf.jasperreports:jasperreports): 6.19.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36518 (High) detected in jackson-databind-2.10.0.jar - ## CVE-2020-36518 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.10.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.0/jackson-databind-2.10.0.jar</p> <p> Dependency Hierarchy: - jasperreports-6.13.0.jar (Root Library) - :x: **jackson-databind-2.10.0.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects. Mend Note: After conducting further research, Mend has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518. <p>Publish Date: 2022-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-03-11</p> <p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.12.6.1</p> <p>Direct dependency fix Resolution (net.sf.jasperreports:jasperreports): 6.19.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy jasperreports jar root library x jackson databind jar vulnerable library found in base branch master vulnerability details jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects mend note after conducting further research mend has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution net sf jasperreports jasperreports step up your open source security game with mend
0
352,094
10,531,779,068
IssuesEvent
2019-10-01 09:16:51
AY1920S1-CS2103T-T13-4/main
https://api.github.com/repos/AY1920S1-CS2103T-T13-4/main
opened
Implement full autotag command
priority.High type.Story
- [ ] Include not-conditions (`nu/`, `nt/`, `nf/`) in search - [ ] Apply autotag whenever bookmarks are updated Extension of #39
1.0
Implement full autotag command - - [ ] Include not-conditions (`nu/`, `nt/`, `nf/`) in search - [ ] Apply autotag whenever bookmarks are updated Extension of #39
non_process
implement full autotag command include not conditions nu nt nf in search apply autotag whenever bookmarks are updated extension of
0
3,079
6,096,823,903
IssuesEvent
2017-06-20 00:29:20
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
vagrant-cloud lacks the ability to specify a provider causing version collisions
enhancement post-processor/vagrant
I updated my packer configs so they would use the vagrant-cloud post processor instead of the atlas post-processor, and have run into a few problems. 1) The biggest issue seems to be alternate versions of a box, intended for different providers get rejected because the version numbers collide. After glancing at the code, it appears the plugin is using the input id to determine the provider, which doesn't work well with complicated json files. II can't set the output for the every artifact to virtualbox.box... that would create quite the mess. Can we add the ability to override the provider string, in the same way we could with atlas plugin? If we're going with 'box_tag' then perhaps 'box_provider' would make the most sense for a json label. 2) Packer doesn't seem to use the ATLAS_TOKEN with the vagrant cloud plugin. Is there an equivalent environment variable? I'm reading in the environment variable manually at the moment, and then setting the token attribute, but it's clunky.
1.0
vagrant-cloud lacks the ability to specify a provider causing version collisions - I updated my packer configs so they would use the vagrant-cloud post processor instead of the atlas post-processor, and have run into a few problems. 1) The biggest issue seems to be alternate versions of a box, intended for different providers get rejected because the version numbers collide. After glancing at the code, it appears the plugin is using the input id to determine the provider, which doesn't work well with complicated json files. II can't set the output for the every artifact to virtualbox.box... that would create quite the mess. Can we add the ability to override the provider string, in the same way we could with atlas plugin? If we're going with 'box_tag' then perhaps 'box_provider' would make the most sense for a json label. 2) Packer doesn't seem to use the ATLAS_TOKEN with the vagrant cloud plugin. Is there an equivalent environment variable? I'm reading in the environment variable manually at the moment, and then setting the token attribute, but it's clunky.
process
vagrant cloud lacks the ability to specify a provider causing version collisions i updated my packer configs so they would use the vagrant cloud post processor instead of the atlas post processor and have run into a few problems the biggest issue seems to be alternate versions of a box intended for different providers get rejected because the version numbers collide after glancing at the code it appears the plugin is using the input id to determine the provider which doesn t work well with complicated json files ii can t set the output for the every artifact to virtualbox box that would create quite the mess can we add the ability to override the provider string in the same way we could with atlas plugin if we re going with box tag then perhaps box provider would make the most sense for a json label packer doesn t seem to use the atlas token with the vagrant cloud plugin is there an equivalent environment variable i m reading in the environment variable manually at the moment and then setting the token attribute but it s clunky
1
19,136
25,192,838,214
IssuesEvent
2022-11-12 05:41:14
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
ARM disassembler for ARM:LE:32:v5t:default no longer works in Ghidra 10.2
Feature: Processor/ARM Status: Internal
**Describe the bug** The ARM disassembler for language ARM:LE:32:v5t:default no longer works in Ghidra 10.2. I assume the same issue might apply for other ARM versions, but I haven't checked. **To Reproduce** Steps to reproduce the behavior: 1. Create a new project. 2. Open the CodeBrowser. 3. Import an armv5t executable appropriate for the language setting ARM:LE:32:v5t:default, run analysis, and navigate to any function. 4. Observe that instructions are not disassembled. Explicitly disassembling at the offset gives an error with the message "Error [Bad Instruction]: Unable to resolve constructor at \<address\>". **Expected behavior** ARM disassembly should work. It used to work in Ghidra 10.1.2 (I've also tested Ghidra 10.1.5, and it still works there). **Screenshots** Disassembly worked in Ghidra 10.1.2: ![Working disassembly](https://user-images.githubusercontent.com/26724430/200147708-ddac3923-c4e9-4007-b048-3b12e2293f27.png) Disassembly no longer works in Ghidra 10.2: ![Failed disassembly](https://user-images.githubusercontent.com/26724430/200147710-204a8081-1c3d-4898-a443-837dd58003a0.png) **Attachments** Here's a simple, no-op C program cross-compiled for armv5te (but the issue remains if compiled for armv5t) with `arm-none-eabi-gcc`: [a.out.zip](https://github.com/NationalSecurityAgency/ghidra/files/9944909/a.out.zip). The above screenshots are taken using this binary (`main()` is at offset 0x8000). **Environment (please complete the following information):** - OS: macOS 11.7, but I've also tested on Windows 10 and the same issue occurs - Java Version: 17.0.1 - Ghidra Version: 10.2 - Ghidra Origin: official GitHub distro
1.0
ARM disassembler for ARM:LE:32:v5t:default no longer works in Ghidra 10.2 - **Describe the bug** The ARM disassembler for language ARM:LE:32:v5t:default no longer works in Ghidra 10.2. I assume the same issue might apply for other ARM versions, but I haven't checked. **To Reproduce** Steps to reproduce the behavior: 1. Create a new project. 2. Open the CodeBrowser. 3. Import an armv5t executable appropriate for the language setting ARM:LE:32:v5t:default, run analysis, and navigate to any function. 4. Observe that instructions are not disassembled. Explicitly disassembling at the offset gives an error with the message "Error [Bad Instruction]: Unable to resolve constructor at \<address\>". **Expected behavior** ARM disassembly should work. It used to work in Ghidra 10.1.2 (I've also tested Ghidra 10.1.5, and it still works there). **Screenshots** Disassembly worked in Ghidra 10.1.2: ![Working disassembly](https://user-images.githubusercontent.com/26724430/200147708-ddac3923-c4e9-4007-b048-3b12e2293f27.png) Disassembly no longer works in Ghidra 10.2: ![Failed disassembly](https://user-images.githubusercontent.com/26724430/200147710-204a8081-1c3d-4898-a443-837dd58003a0.png) **Attachments** Here's a simple, no-op C program cross-compiled for armv5te (but the issue remains if compiled for armv5t) with `arm-none-eabi-gcc`: [a.out.zip](https://github.com/NationalSecurityAgency/ghidra/files/9944909/a.out.zip). The above screenshots are taken using this binary (`main()` is at offset 0x8000). **Environment (please complete the following information):** - OS: macOS 11.7, but I've also tested on Windows 10 and the same issue occurs - Java Version: 17.0.1 - Ghidra Version: 10.2 - Ghidra Origin: official GitHub distro
process
arm disassembler for arm le default no longer works in ghidra describe the bug the arm disassembler for language arm le default no longer works in ghidra i assume the same issue might apply for other arm versions but i haven t checked to reproduce steps to reproduce the behavior create a new project open the codebrowser import an executable appropriate for the language setting arm le default run analysis and navigate to any function observe that instructions are not disassembled explicitly disassembling at the offset gives an error with the message error unable to resolve constructor at expected behavior arm disassembly should work it used to work in ghidra i ve also tested ghidra and it still works there screenshots disassembly worked in ghidra disassembly no longer works in ghidra attachments here s a simple no op c program cross compiled for but the issue remains if compiled for with arm none eabi gcc the above screenshots are taken using this binary main is at offset environment please complete the following information os macos but i ve also tested on windows and the same issue occurs java version ghidra version ghidra origin official github distro
1
46,378
7,254,803,182
IssuesEvent
2018-02-16 12:33:10
RubenVerborgh/N3.js
https://api.github.com/repos/RubenVerborgh/N3.js
closed
Store doesn't support adding triples using prefixed IRIs
documentation question
In the documentation it says that the prefixes would support the addition **and** lookup of triples, but when adding a triple with a previously added prefix the prefixed URI is not being expanded and the triple is being store with the prefixed URI. I would have expected the triple to be stored with the expanded absolute URIs. The lookup with prefixed URIs works fine. Is this the intended behaviour?
1.0
Store doesn't support adding triples using prefixed IRIs - In the documentation it says that the prefixes would support the addition **and** lookup of triples, but when adding a triple with a previously added prefix the prefixed URI is not being expanded and the triple is being store with the prefixed URI. I would have expected the triple to be stored with the expanded absolute URIs. The lookup with prefixed URIs works fine. Is this the intended behaviour?
non_process
store doesn t support adding triples using prefixed iris in the documentation it says that the prefixes would support the addition and lookup of triples but when adding a triple with a previously added prefix the prefixed uri is not being expanded and the triple is being store with the prefixed uri i would have expected the triple to be stored with the expanded absolute uris the lookup with prefixed uris works fine is this the intended behaviour
0
74,233
3,436,560,317
IssuesEvent
2015-12-12 13:59:05
pywinauto/pywinauto
https://api.github.com/repos/pywinauto/pywinauto
opened
ActiveX controls support for desktop apps
enhancement Priority-Low
Native apps can embed ActiveX controls which text or other data cannot be recognized by pywinauto. There are few MFC samples including ActiveX controls: https://msdn.microsoft.com/en-us/library/aa728874(v=vs.71).aspx Related question on StackOverflow: http://stackoverflow.com/q/34181583/3648361
1.0
ActiveX controls support for desktop apps - Native apps can embed ActiveX controls which text or other data cannot be recognized by pywinauto. There are few MFC samples including ActiveX controls: https://msdn.microsoft.com/en-us/library/aa728874(v=vs.71).aspx Related question on StackOverflow: http://stackoverflow.com/q/34181583/3648361
non_process
activex controls support for desktop apps native apps can embed activex controls which text or other data cannot be recognized by pywinauto there are few mfc samples including activex controls related question on stackoverflow
0
15,356
19,529,633,938
IssuesEvent
2021-12-30 14:26:12
MikeKSmith/The_Lazy_Producer
https://api.github.com/repos/MikeKSmith/The_Lazy_Producer
opened
Discuss alternative / non-linear pattern sequencers like Less Concepts, Euclidean, Turing machines
process
As a process, should discuss alternative sequencer types than the standard, "linear" type with fixed length and fixed divisions. Less Concepts is random and unpredictable. Might be interesting to include this in here.
1.0
Discuss alternative / non-linear pattern sequencers like Less Concepts, Euclidean, Turing machines - As a process, should discuss alternative sequencer types than the standard, "linear" type with fixed length and fixed divisions. Less Concepts is random and unpredictable. Might be interesting to include this in here.
process
discuss alternative non linear pattern sequencers like less concepts euclidean turing machines as a process should discuss alternative sequencer types than the standard linear type with fixed length and fixed divisions less concepts is random and unpredictable might be interesting to include this in here
1
15,734
19,910,044,899
IssuesEvent
2022-01-25 16:18:53
opensearch-project/data-prepper
https://api.github.com/repos/opensearch-project/data-prepper
closed
Support parsing messages with key-value strings such as queries and properties
enhancement plugin - processor KeyValueProcessor
Data Prepper should have a processor which can parse key-value strings from one field and save the results into another field as a map. For example, on the input event: ``` { "query": "key1:value1&key2:value2&key3:value3" } ``` Could update the event to include a field: ``` { "query_params" : { "key1" : "value1", "key2" : "value2", "key3" : "value3" } } ``` This processor must allow configuration of: * The source field (`query` in the example above) * The destination field (`query_params` in the example above) * The delimiter between fields (`&` in the example above) * The delimiter between key and value (`:` in the example above) * non_match value - default `null`, the value to assign a key when it has no match * prefix - A prefix to all keys. default `""`. It should also support regex expressions for both types of delimiters.
2.0
Support parsing messages with key-value strings such as queries and properties - Data Prepper should have a processor which can parse key-value strings from one field and save the results into another field as a map. For example, on the input event: ``` { "query": "key1:value1&key2:value2&key3:value3" } ``` Could update the event to include a field: ``` { "query_params" : { "key1" : "value1", "key2" : "value2", "key3" : "value3" } } ``` This processor must allow configuration of: * The source field (`query` in the example above) * The destination field (`query_params` in the example above) * The delimiter between fields (`&` in the example above) * The delimiter between key and value (`:` in the example above) * non_match value - default `null`, the value to assign a key when it has no match * prefix - A prefix to all keys. default `""`. It should also support regex expressions for both types of delimiters.
process
support parsing messages with key value strings such as queries and properties data prepper should have a processor which can parse key value strings from one field and save the results into another field as a map for example on the input event query could update the event to include a field query params this processor must allow configuration of the source field query in the example above the destination field query params in the example above the delimiter between fields in the example above the delimiter between key and value in the example above non match value default null the value to assign a key when it has no match prefix a prefix to all keys default it should also support regex expressions for both types of delimiters
1
49,949
26,403,450,982
IssuesEvent
2023-01-13 04:53:26
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Linq enumeration of a string's characters is twice as slow as enumerating them using string.ToCharArray()
tenet-performance
### Description Enumerating characters of a string using Linq is slower than doing it on its copy created by `ToCharArray()`. Because of this behavior, people might start seeing `ToCharArray()` as a performance magic (there are already some stackoverflow entries about it), and it might get used everywhere, causing memory overhead unnecessarily. Given the benchmark below: ```csharp [MemoryDiagnoser] public partial class StringVsToCharArrayBenchmark { private static string longText = new string('a', 1000000); [Benchmark] public bool Linq_ToCharArray() { return longText.ToCharArray().Where(c => c == 'a').Count() > 100; } [Benchmark] public bool Linq_String() { return longText.Where(c => c == 'a').Count() > 100; } static void Main() { _ = BenchmarkRunner.Run<StringVsToCharArrayBenchmark>(); } } ``` The benchmark `Linq_String` is twice as slow as `Linq_ToCharArray` despite that ToCharArray() allocates two strings worth of extra memory. The relative 2x performance difference doesn't change based on the string size. Here is the BenchmarkDotNet output for the given code above: | Method | Mean | Error | StdDev | Gen0 | Gen1 | Gen2 | Allocated | |----------------- |---------:|----------:|----------:|---------:|---------:|---------:|----------:| | Linq_ToCharArray | 1.858 ms | 0.0289 ms | 0.0271 ms | 253.9063 | 253.9063 | 253.9063 | 2000935 B | | Linq_String | 4.839 ms | 0.0954 ms | 0.0980 ms | - | - | - | 92 B | (`foreach (var s in string)` is still the fastest of the bunch of course) ### Configuration ``` BenchmarkDotNet=v0.13.3, OS=Windows 11 (10.0.22623.1095) AMD Ryzen 9 5950X, 1 CPU, 32 logical and 16 physical cores .NET SDK=7.0.101 [Host] : .NET 7.0.1 (7.0.122.56804), X64 RyuJIT AVX2 [AttachedDebugger] DefaultJob : .NET 7.0.1 (7.0.122.56804), X64 RyuJIT AVX2 ``` ### Regression? The difference is still there on .NET 4.8.1 but less stark because both implementations are very slow compared to .NET 7.0. :) | Method | Job | Runtime | Mean | Error | StdDev | |----------------- |--------------------- |--------------------- |---------:|----------:|----------:| | Linq_ToCharArray | .NET 7.0 | .NET 7.0 | 2.117 ms | 0.0276 ms | 0.0258 ms | | Linq_String | .NET 7.0 | .NET 7.0 | 5.284 ms | 0.1053 ms | 0.1576 ms | | Linq_ToCharArray | .NET Framework 4.8.1 | .NET Framework 4.8.1 | 4.095 ms | 0.0239 ms | 0.0224 ms | | Linq_String | .NET Framework 4.8.1 | .NET Framework 4.8.1 | 6.799 ms | 0.0166 ms | 0.0147 ms | ### Analysis Linq's string enumeration uses `WhereEnumerableIterator<TSource>`, while `ToCharArray()` version uses `WhereArrayIterator<TSource>` which is optimized for arrays despite that string is perfectly capable of indexed element access and precalculated length. I believe the overhead comes from slow Enumerable-based iteration. ### Solution proposal I propose a `WhereStringIterator` and an accompanying `WhereSelectStringIterator` classes for strings. It might look like this (based on `WhereArrayIterator`): ```csharp /// <summary> /// An iterator that filters each character of a string. /// </summary> internal sealed partial class WhereStringIterator : Iterator<char> { private readonly string _source; private readonly Func<char, bool> _predicate; public WhereStringIterator(string source, Func<char, bool> predicate) { Debug.Assert(source != null && source.Length > 0); Debug.Assert(predicate != null); _source = source; _predicate = predicate; } public override Iterator<char> Clone() => new WhereStringIterator(_source, _predicate); public override bool MoveNext() { int index = _state - 1; string source = _source; while (unchecked((uint)index < (uint)source.Length)) { char item = source[index]; index = _state++; if (_predicate(item)) { _current = item; return true; } } Dispose(); return false; } public override IEnumerable<TResult> Select<TResult>(Func<TSource, TResult> selector) => new WhereSelectStringIterator<TResult>(_source, _predicate, selector); public override IEnumerable<TSource> Where(Func<TSource, bool> predicate) => new WhereStringIterator(_source, CombinePredicates(_predicate, predicate)); } ``` The catch is that `Enumerable.Where<TSource>()` implementation in `Where.cs` will have an additional type check like this: ```csharp public static IEnumerable<TSource> Where<TSource>(this IEnumerable<TSource> source, Func<TSource, bool> predicate) { // ... irrelevant code removed // @ssg: these consecutive ifs below can also benefit from being converted // to a switch expression, not sure if there's any performance difference though. if (source is Iterator<TSource> iterator) { return iterator.Where(predicate); } if (source is TSource[] array) { return array.Length == 0 ? Empty<TSource>() : new WhereArrayIterator<TSource>(array, predicate); } if (source is List<TSource> list) { return new WhereListIterator<TSource>(list, predicate); } // @ssg: This is what we're adding. if (source is string str) { return str.Length == 0 ? Empty<TSource>() : new WhereStringIterator(str, predicate); } return new WhereEnumerableIterator<TSource>(source, predicate); } ``` If you can confirm that this is way to go, I can go ahead and try creating a PR for this on System.Linq. It would make string enumeration as fast as array enumeration and would remove the superstition around using `ToCharArray()` unnecessarily. It might also be worthwhile to investigate why `WhereEnumerableIterator<TSource>` is twice as slow as the array iterator. My guess would be because of the overhead of an extra `MoveNext()` call every iteration instead of just an indexed memory access.
True
Linq enumeration of a string's characters is twice as slow as enumerating them using string.ToCharArray() - ### Description Enumerating characters of a string using Linq is slower than doing it on its copy created by `ToCharArray()`. Because of this behavior, people might start seeing `ToCharArray()` as a performance magic (there are already some stackoverflow entries about it), and it might get used everywhere, causing memory overhead unnecessarily. Given the benchmark below: ```csharp [MemoryDiagnoser] public partial class StringVsToCharArrayBenchmark { private static string longText = new string('a', 1000000); [Benchmark] public bool Linq_ToCharArray() { return longText.ToCharArray().Where(c => c == 'a').Count() > 100; } [Benchmark] public bool Linq_String() { return longText.Where(c => c == 'a').Count() > 100; } static void Main() { _ = BenchmarkRunner.Run<StringVsToCharArrayBenchmark>(); } } ``` The benchmark `Linq_String` is twice as slow as `Linq_ToCharArray` despite that ToCharArray() allocates two strings worth of extra memory. The relative 2x performance difference doesn't change based on the string size. Here is the BenchmarkDotNet output for the given code above: | Method | Mean | Error | StdDev | Gen0 | Gen1 | Gen2 | Allocated | |----------------- |---------:|----------:|----------:|---------:|---------:|---------:|----------:| | Linq_ToCharArray | 1.858 ms | 0.0289 ms | 0.0271 ms | 253.9063 | 253.9063 | 253.9063 | 2000935 B | | Linq_String | 4.839 ms | 0.0954 ms | 0.0980 ms | - | - | - | 92 B | (`foreach (var s in string)` is still the fastest of the bunch of course) ### Configuration ``` BenchmarkDotNet=v0.13.3, OS=Windows 11 (10.0.22623.1095) AMD Ryzen 9 5950X, 1 CPU, 32 logical and 16 physical cores .NET SDK=7.0.101 [Host] : .NET 7.0.1 (7.0.122.56804), X64 RyuJIT AVX2 [AttachedDebugger] DefaultJob : .NET 7.0.1 (7.0.122.56804), X64 RyuJIT AVX2 ``` ### Regression? The difference is still there on .NET 4.8.1 but less stark because both implementations are very slow compared to .NET 7.0. :) | Method | Job | Runtime | Mean | Error | StdDev | |----------------- |--------------------- |--------------------- |---------:|----------:|----------:| | Linq_ToCharArray | .NET 7.0 | .NET 7.0 | 2.117 ms | 0.0276 ms | 0.0258 ms | | Linq_String | .NET 7.0 | .NET 7.0 | 5.284 ms | 0.1053 ms | 0.1576 ms | | Linq_ToCharArray | .NET Framework 4.8.1 | .NET Framework 4.8.1 | 4.095 ms | 0.0239 ms | 0.0224 ms | | Linq_String | .NET Framework 4.8.1 | .NET Framework 4.8.1 | 6.799 ms | 0.0166 ms | 0.0147 ms | ### Analysis Linq's string enumeration uses `WhereEnumerableIterator<TSource>`, while `ToCharArray()` version uses `WhereArrayIterator<TSource>` which is optimized for arrays despite that string is perfectly capable of indexed element access and precalculated length. I believe the overhead comes from slow Enumerable-based iteration. ### Solution proposal I propose a `WhereStringIterator` and an accompanying `WhereSelectStringIterator` classes for strings. It might look like this (based on `WhereArrayIterator`): ```csharp /// <summary> /// An iterator that filters each character of a string. /// </summary> internal sealed partial class WhereStringIterator : Iterator<char> { private readonly string _source; private readonly Func<char, bool> _predicate; public WhereStringIterator(string source, Func<char, bool> predicate) { Debug.Assert(source != null && source.Length > 0); Debug.Assert(predicate != null); _source = source; _predicate = predicate; } public override Iterator<char> Clone() => new WhereStringIterator(_source, _predicate); public override bool MoveNext() { int index = _state - 1; string source = _source; while (unchecked((uint)index < (uint)source.Length)) { char item = source[index]; index = _state++; if (_predicate(item)) { _current = item; return true; } } Dispose(); return false; } public override IEnumerable<TResult> Select<TResult>(Func<TSource, TResult> selector) => new WhereSelectStringIterator<TResult>(_source, _predicate, selector); public override IEnumerable<TSource> Where(Func<TSource, bool> predicate) => new WhereStringIterator(_source, CombinePredicates(_predicate, predicate)); } ``` The catch is that `Enumerable.Where<TSource>()` implementation in `Where.cs` will have an additional type check like this: ```csharp public static IEnumerable<TSource> Where<TSource>(this IEnumerable<TSource> source, Func<TSource, bool> predicate) { // ... irrelevant code removed // @ssg: these consecutive ifs below can also benefit from being converted // to a switch expression, not sure if there's any performance difference though. if (source is Iterator<TSource> iterator) { return iterator.Where(predicate); } if (source is TSource[] array) { return array.Length == 0 ? Empty<TSource>() : new WhereArrayIterator<TSource>(array, predicate); } if (source is List<TSource> list) { return new WhereListIterator<TSource>(list, predicate); } // @ssg: This is what we're adding. if (source is string str) { return str.Length == 0 ? Empty<TSource>() : new WhereStringIterator(str, predicate); } return new WhereEnumerableIterator<TSource>(source, predicate); } ``` If you can confirm that this is way to go, I can go ahead and try creating a PR for this on System.Linq. It would make string enumeration as fast as array enumeration and would remove the superstition around using `ToCharArray()` unnecessarily. It might also be worthwhile to investigate why `WhereEnumerableIterator<TSource>` is twice as slow as the array iterator. My guess would be because of the overhead of an extra `MoveNext()` call every iteration instead of just an indexed memory access.
non_process
linq enumeration of a string s characters is twice as slow as enumerating them using string tochararray description enumerating characters of a string using linq is slower than doing it on its copy created by tochararray because of this behavior people might start seeing tochararray as a performance magic there are already some stackoverflow entries about it and it might get used everywhere causing memory overhead unnecessarily given the benchmark below csharp public partial class stringvstochararraybenchmark private static string longtext new string a public bool linq tochararray return longtext tochararray where c c a count public bool linq string return longtext where c c a count static void main benchmarkrunner run the benchmark linq string is twice as slow as linq tochararray despite that tochararray allocates two strings worth of extra memory the relative performance difference doesn t change based on the string size here is the benchmarkdotnet output for the given code above method mean error stddev allocated linq tochararray ms ms ms b linq string ms ms ms b foreach var s in string is still the fastest of the bunch of course configuration benchmarkdotnet os windows amd ryzen cpu logical and physical cores net sdk net ryujit defaultjob net ryujit regression the difference is still there on net but less stark because both implementations are very slow compared to net method job runtime mean error stddev linq tochararray net net ms ms ms linq string net net ms ms ms linq tochararray net framework net framework ms ms ms linq string net framework net framework ms ms ms analysis linq s string enumeration uses whereenumerableiterator while tochararray version uses wherearrayiterator which is optimized for arrays despite that string is perfectly capable of indexed element access and precalculated length i believe the overhead comes from slow enumerable based iteration solution proposal i propose a wherestringiterator and an accompanying whereselectstringiterator classes for strings it might look like this based on wherearrayiterator csharp an iterator that filters each character of a string internal sealed partial class wherestringiterator iterator private readonly string source private readonly func predicate public wherestringiterator string source func predicate debug assert source null source length debug assert predicate null source source predicate predicate public override iterator clone new wherestringiterator source predicate public override bool movenext int index state string source source while unchecked uint index uint source length char item source index state if predicate item current item return true dispose return false public override ienumerable select func selector new whereselectstringiterator source predicate selector public override ienumerable where func predicate new wherestringiterator source combinepredicates predicate predicate the catch is that enumerable where implementation in where cs will have an additional type check like this csharp public static ienumerable where this ienumerable source func predicate irrelevant code removed ssg these consecutive ifs below can also benefit from being converted to a switch expression not sure if there s any performance difference though if source is iterator iterator return iterator where predicate if source is tsource array return array length empty new wherearrayiterator array predicate if source is list list return new wherelistiterator list predicate ssg this is what we re adding if source is string str return str length empty new wherestringiterator str predicate return new whereenumerableiterator source predicate if you can confirm that this is way to go i can go ahead and try creating a pr for this on system linq it would make string enumeration as fast as array enumeration and would remove the superstition around using tochararray unnecessarily it might also be worthwhile to investigate why whereenumerableiterator is twice as slow as the array iterator my guess would be because of the overhead of an extra movenext call every iteration instead of just an indexed memory access
0
25,266
11,164,769,086
IssuesEvent
2019-12-27 06:38:31
soumya132/java-code
https://api.github.com/repos/soumya132/java-code
opened
CVE-2018-14720 (High) detected in jackson-databind-2.8.1.jar
security vulnerability
## CVE-2018-14720 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/java-code/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/soumya132/java-code/commit/323d775d03827a3c9662d0ba50b640fe7addea5a">323d775d03827a3c9662d0ba50b640fe7addea5a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.7 might allow attackers to conduct external XML entity (XXE) attacks by leveraging failure to block unspecified JDK classes from polymorphic deserialization. <p>Publish Date: 2019-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14720>CVE-2018-14720</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14720">https://nvd.nist.gov/vuln/detail/CVE-2018-14720</a></p> <p>Release Date: 2019-01-02</p> <p>Fix Resolution: 2.9.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-14720 (High) detected in jackson-databind-2.8.1.jar - ## CVE-2018-14720 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/java-code/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/soumya132/java-code/commit/323d775d03827a3c9662d0ba50b640fe7addea5a">323d775d03827a3c9662d0ba50b640fe7addea5a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.7 might allow attackers to conduct external XML entity (XXE) attacks by leveraging failure to block unspecified JDK classes from polymorphic deserialization. <p>Publish Date: 2019-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14720>CVE-2018-14720</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14720">https://nvd.nist.gov/vuln/detail/CVE-2018-14720</a></p> <p>Release Date: 2019-01-02</p> <p>Fix Resolution: 2.9.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm java code pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter jersey release jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow attackers to conduct external xml entity xxe attacks by leveraging failure to block unspecified jdk classes from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
19,714
26,054,756,728
IssuesEvent
2022-12-22 23:19:36
bitfocus/companion-module-requests
https://api.github.com/repos/bitfocus/companion-module-requests
closed
Ma Lighting gMA3
NOT YET PROCESSED
On Bitfocus companion I only find ma lighting grandma 2, but does someone have companion module for grandMA 3? Cheers
1.0
Ma Lighting gMA3 - On Bitfocus companion I only find ma lighting grandma 2, but does someone have companion module for grandMA 3? Cheers
process
ma lighting on bitfocus companion i only find ma lighting grandma but does someone have companion module for grandma cheers
1
452
2,892,965,138
IssuesEvent
2015-06-15 15:34:55
cfpb/hmda-viz-prototype
https://api.github.com/repos/cfpb/hmda-viz-prototype
closed
Jekyll support for disclosure reports
Processing
Write scripts to power Jekyll @awolfe76 Can you create a list of the requirements?
1.0
Jekyll support for disclosure reports - Write scripts to power Jekyll @awolfe76 Can you create a list of the requirements?
process
jekyll support for disclosure reports write scripts to power jekyll can you create a list of the requirements
1
2,724
5,612,332,704
IssuesEvent
2017-04-03 04:27:40
alexrj/Slic3r
https://api.github.com/repos/alexrj/Slic3r
closed
Feature request- belt stretch/frame wobble compensation for tall prints
Feature request Fixable with post-process script Won't Change
My most recent build is a coreXY printer with belt-lifted Z axis. When the bed is loaded, the belts stretch. I have measured the amount of stretch in my machine at about 42 um/kg load. That means that as the print gets heavier, the print layers are going to thicken while getting narrower. If there was a way to enter the belt stretch value, say 42 um/kg in my case, and have slic3r compensate for the belt stretch by adjusting the layer thickness while maintaining constant extruder flow, even large, heavy prints would come out with uniform layer thickness/extrusion, and accurate height. In tall printers with beds moving in the Z axis, when the bed is high it is closely coupled to the XY stage that tends to throw the printer around as the parts move. As the print gets taller, the bed drops and becomes more decoupled from the wobbling top of the printer, producing artifacts in the print surface. If the speed/acceleration could be adjusted as a function of print height the print artifacts could be minimized. The same could apply to i3 type machines lifting the X axis in Z.
1.0
Feature request- belt stretch/frame wobble compensation for tall prints - My most recent build is a coreXY printer with belt-lifted Z axis. When the bed is loaded, the belts stretch. I have measured the amount of stretch in my machine at about 42 um/kg load. That means that as the print gets heavier, the print layers are going to thicken while getting narrower. If there was a way to enter the belt stretch value, say 42 um/kg in my case, and have slic3r compensate for the belt stretch by adjusting the layer thickness while maintaining constant extruder flow, even large, heavy prints would come out with uniform layer thickness/extrusion, and accurate height. In tall printers with beds moving in the Z axis, when the bed is high it is closely coupled to the XY stage that tends to throw the printer around as the parts move. As the print gets taller, the bed drops and becomes more decoupled from the wobbling top of the printer, producing artifacts in the print surface. If the speed/acceleration could be adjusted as a function of print height the print artifacts could be minimized. The same could apply to i3 type machines lifting the X axis in Z.
process
feature request belt stretch frame wobble compensation for tall prints my most recent build is a corexy printer with belt lifted z axis when the bed is loaded the belts stretch i have measured the amount of stretch in my machine at about um kg load that means that as the print gets heavier the print layers are going to thicken while getting narrower if there was a way to enter the belt stretch value say um kg in my case and have compensate for the belt stretch by adjusting the layer thickness while maintaining constant extruder flow even large heavy prints would come out with uniform layer thickness extrusion and accurate height in tall printers with beds moving in the z axis when the bed is high it is closely coupled to the xy stage that tends to throw the printer around as the parts move as the print gets taller the bed drops and becomes more decoupled from the wobbling top of the printer producing artifacts in the print surface if the speed acceleration could be adjusted as a function of print height the print artifacts could be minimized the same could apply to type machines lifting the x axis in z
1
127,971
10,509,099,144
IssuesEvent
2019-09-27 10:08:33
ampproject/amp-wp
https://api.github.com/repos/ampproject/amp-wp
closed
Error message: "this block contains unexpected or invalid content"
AMP Stories Feedback Needs Testing Status: Duplicate
Repro steps: 1. I created a story last week 1. My role was changed from author to editor (not sure if this step is required to repro) 1. I opened the story today, and I see the following error messages. ![image](https://user-images.githubusercontent.com/38660603/60518770-2c204400-9cb0-11e9-84da-e0db1fa5dcde.png)
1.0
Error message: "this block contains unexpected or invalid content" - Repro steps: 1. I created a story last week 1. My role was changed from author to editor (not sure if this step is required to repro) 1. I opened the story today, and I see the following error messages. ![image](https://user-images.githubusercontent.com/38660603/60518770-2c204400-9cb0-11e9-84da-e0db1fa5dcde.png)
non_process
error message this block contains unexpected or invalid content repro steps i created a story last week my role was changed from author to editor not sure if this step is required to repro i opened the story today and i see the following error messages
0
267,905
20,250,374,068
IssuesEvent
2022-02-14 17:17:23
openfeatureflags/spec
https://api.github.com/repos/openfeatureflags/spec
opened
Add architecture summary
documentation
There are some architecture sketches which could be added to the repository. It would be nice to do so
1.0
Add architecture summary - There are some architecture sketches which could be added to the repository. It would be nice to do so
non_process
add architecture summary there are some architecture sketches which could be added to the repository it would be nice to do so
0