Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
242,994
| 26,277,893,409
|
IssuesEvent
|
2023-01-07 01:26:03
|
turkdevops/snyk
|
https://api.github.com/repos/turkdevops/snyk
|
closed
|
CVE-2011-4969 (Low) detected in yiisoft/yii-1.1.14 - autoclosed
|
security vulnerability
|
## CVE-2011-4969 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yiisoft/yii-1.1.14</b></p></summary>
<p>Yii Web Programming Framework</p>
<p>Library home page: <a href="https://api.github.com/repos/yiisoft/yii/zipball/f0fee98ee84f70f1f3652f65562c9670e919cb4e">https://api.github.com/repos/yiisoft/yii/zipball/f0fee98ee84f70f1f3652f65562c9670e919cb4e</a></p>
<p>
Dependency Hierarchy:
- :x: **yiisoft/yii-1.1.14** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/snyk/commit/9505f4ca92405cc9273dc3726c2d274ce28a4407">9505f4ca92405cc9273dc3726c2d274ce28a4407</a></p>
<p>Found in base branch: <b>ALL_HANDS/major-secrets</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.
<p>Publish Date: 2013-03-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-4969>CVE-2011-4969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p>
<p>Release Date: 2013-03-08</p>
<p>Fix Resolution: 1.6.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2011-4969 (Low) detected in yiisoft/yii-1.1.14 - autoclosed - ## CVE-2011-4969 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yiisoft/yii-1.1.14</b></p></summary>
<p>Yii Web Programming Framework</p>
<p>Library home page: <a href="https://api.github.com/repos/yiisoft/yii/zipball/f0fee98ee84f70f1f3652f65562c9670e919cb4e">https://api.github.com/repos/yiisoft/yii/zipball/f0fee98ee84f70f1f3652f65562c9670e919cb4e</a></p>
<p>
Dependency Hierarchy:
- :x: **yiisoft/yii-1.1.14** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/snyk/commit/9505f4ca92405cc9273dc3726c2d274ce28a4407">9505f4ca92405cc9273dc3726c2d274ce28a4407</a></p>
<p>Found in base branch: <b>ALL_HANDS/major-secrets</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.
<p>Publish Date: 2013-03-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-4969>CVE-2011-4969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p>
<p>Release Date: 2013-03-08</p>
<p>Fix Resolution: 1.6.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve low detected in yiisoft yii autoclosed cve low severity vulnerability vulnerable library yiisoft yii yii web programming framework library home page a href dependency hierarchy x yiisoft yii vulnerable library found in head commit a href found in base branch all hands major secrets vulnerability details cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
105,051
| 22,832,341,291
|
IssuesEvent
|
2022-07-12 13:55:48
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
closed
|
[Accessibility]: `N/A` hard to understand when read aloud
|
team/code-intelligence accessibility team/frontend-platform wcag/2.1/fixing wcag/2.1
|
### Audit type
Screen reader navigation
### User journey audit issue
#33516, #33518
### Problem description
<img width="884" alt="n:a" src="https://user-images.githubusercontent.com/103087/167504500-6c228e31-688b-4262-ba26-14d18129ffd1.png">
I couldn't understand the text for `N/A` here.
### Expected behavior
Rename text or add an aria-label that can say `not applicable` or some other long-form version that can be understood.
### Additional details
_No response_
### Assigning labels
- [X] Please give this issue an estimate by applying a label like `estimate/Xd`, where X is the estimated number of days it will take to complete.
- [X] If this issue is specific to a specific Sourcegraph product, please assign the appropriate team label to this issue.
- [X] If this issue will require input from designers in order to complete, please assign the label `needs-design`.
- [X] If you are confident that this issue should be fixed by GitStart, please assign the label `gitstart`.
### Owner
This issue will be fixed by my team, I have assigned a relevant member to it, or I will do so in the near future.
|
1.0
|
[Accessibility]: `N/A` hard to understand when read aloud - ### Audit type
Screen reader navigation
### User journey audit issue
#33516, #33518
### Problem description
<img width="884" alt="n:a" src="https://user-images.githubusercontent.com/103087/167504500-6c228e31-688b-4262-ba26-14d18129ffd1.png">
I couldn't understand the text for `N/A` here.
### Expected behavior
Rename text or add an aria-label that can say `not applicable` or some other long-form version that can be understood.
### Additional details
_No response_
### Assigning labels
- [X] Please give this issue an estimate by applying a label like `estimate/Xd`, where X is the estimated number of days it will take to complete.
- [X] If this issue is specific to a specific Sourcegraph product, please assign the appropriate team label to this issue.
- [X] If this issue will require input from designers in order to complete, please assign the label `needs-design`.
- [X] If you are confident that this issue should be fixed by GitStart, please assign the label `gitstart`.
### Owner
This issue will be fixed by my team, I have assigned a relevant member to it, or I will do so in the near future.
|
non_process
|
n a hard to understand when read aloud audit type screen reader navigation user journey audit issue problem description img width alt n a src i couldn t understand the text for n a here expected behavior rename text or add an aria label that can say not applicable or some other long form version that can be understood additional details no response assigning labels please give this issue an estimate by applying a label like estimate xd where x is the estimated number of days it will take to complete if this issue is specific to a specific sourcegraph product please assign the appropriate team label to this issue if this issue will require input from designers in order to complete please assign the label needs design if you are confident that this issue should be fixed by gitstart please assign the label gitstart owner this issue will be fixed by my team i have assigned a relevant member to it or i will do so in the near future
| 0
|
6,790
| 9,921,948,721
|
IssuesEvent
|
2019-06-30 22:59:12
|
GroceriStar/fetch-constants
|
https://api.github.com/repos/GroceriStar/fetch-constants
|
closed
|
#### [Recipe Search][URLS][part1]
|
in-process
|
https://chickenkyiv.github.io/search-api-documentation/docs/db-schema
https://chickenkyiv.github.io/search-api-documentation/docs/db-schema
By using names on URLS from this [page](https://chickenkyiv.github.io/search-api-documentation/docs/database-tables-models/attribute/allergy)
In order to make it better, we'll create set of constants, each for a different method.
Example:
*http://localhost:3000/api/attribute?filter[where][type]=allergy*
will became `export const ATTRIBUTE_FILTER_TYPE_ALLERGY = "ATTRIBUTE_FILTER_TYPE_ALLERGY";`
|
1.0
|
#### [Recipe Search][URLS][part1] - https://chickenkyiv.github.io/search-api-documentation/docs/db-schema
https://chickenkyiv.github.io/search-api-documentation/docs/db-schema
By using names on URLS from this [page](https://chickenkyiv.github.io/search-api-documentation/docs/database-tables-models/attribute/allergy)
In order to make it better, we'll create set of constants, each for a different method.
Example:
*http://localhost:3000/api/attribute?filter[where][type]=allergy*
will became `export const ATTRIBUTE_FILTER_TYPE_ALLERGY = "ATTRIBUTE_FILTER_TYPE_ALLERGY";`
|
process
|
by using names on urls from this in order to make it better we ll create set of constants each for a different method example allergy will became export const attribute filter type allergy attribute filter type allergy
| 1
|
4,300
| 7,194,970,240
|
IssuesEvent
|
2018-02-04 12:13:39
|
w3c/html
|
https://api.github.com/repos/w3c/html
|
closed
|
Add attribution of WHATWG HTML, as required by the CC-BY license
|
editorial process
|
WHATWG HTML now has the following copyright notice:
> Copyright © 2018 WHATWG (Apple, Google, Mozilla, Microsoft). This work is licensed under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/).
Unlike the previous license, this license explicitly requires attribution. Although previous copying can be interpreted as under the previous license, newer copying must follow the CC-BY 4.0 license and therefore requires attribution.
The following commits appears to copy text from WHATWG HTML without attribution:
https://github.com/w3c/html/commit/25aaa34bc4b8fbbad11a125d57a441f21256206d
https://github.com/w3c/html/commit/6d380f2936c0b08ca4a2dc64e334c029780f1e2f
Copying this text is allowed by the copyright license. However, the license requires proper attribution.
Please add the attribution required by the new copyright license.
Here is an example of an appropriate attribution CC-BY:
> Portions derived from [HTML Living Standard](https://html.spec.whatwg.org), Copyright © 2018 WHATWG (Apple, Google, Mozilla, Microsoft) under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/).
|
1.0
|
Add attribution of WHATWG HTML, as required by the CC-BY license - WHATWG HTML now has the following copyright notice:
> Copyright © 2018 WHATWG (Apple, Google, Mozilla, Microsoft). This work is licensed under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/).
Unlike the previous license, this license explicitly requires attribution. Although previous copying can be interpreted as under the previous license, newer copying must follow the CC-BY 4.0 license and therefore requires attribution.
The following commits appears to copy text from WHATWG HTML without attribution:
https://github.com/w3c/html/commit/25aaa34bc4b8fbbad11a125d57a441f21256206d
https://github.com/w3c/html/commit/6d380f2936c0b08ca4a2dc64e334c029780f1e2f
Copying this text is allowed by the copyright license. However, the license requires proper attribution.
Please add the attribution required by the new copyright license.
Here is an example of an appropriate attribution CC-BY:
> Portions derived from [HTML Living Standard](https://html.spec.whatwg.org), Copyright © 2018 WHATWG (Apple, Google, Mozilla, Microsoft) under a [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/).
|
process
|
add attribution of whatwg html as required by the cc by license whatwg html now has the following copyright notice copyright © whatwg apple google mozilla microsoft this work is licensed under a unlike the previous license this license explicitly requires attribution although previous copying can be interpreted as under the previous license newer copying must follow the cc by license and therefore requires attribution the following commits appears to copy text from whatwg html without attribution copying this text is allowed by the copyright license however the license requires proper attribution please add the attribution required by the new copyright license here is an example of an appropriate attribution cc by portions derived from copyright © whatwg apple google mozilla microsoft under a
| 1
|
3,523
| 6,564,753,246
|
IssuesEvent
|
2017-09-08 03:58:05
|
zero-os/0-Disk
|
https://api.github.com/repos/zero-os/0-Disk
|
closed
|
zeroctl rebalance command
|
process_wontfix type_feature
|
Part of the self-healing as described in https://github.com/zero-os/0-orchestrator/blob/master/specs/selfhealing/storage.md.
TODO:
+ describe parameters and how the command would look like;
+ describe what the command would do and to which storage types it would apply (if specific at all);
|
1.0
|
zeroctl rebalance command - Part of the self-healing as described in https://github.com/zero-os/0-orchestrator/blob/master/specs/selfhealing/storage.md.
TODO:
+ describe parameters and how the command would look like;
+ describe what the command would do and to which storage types it would apply (if specific at all);
|
process
|
zeroctl rebalance command part of the self healing as described in todo describe parameters and how the command would look like describe what the command would do and to which storage types it would apply if specific at all
| 1
|
6,044
| 8,854,377,592
|
IssuesEvent
|
2019-01-09 01:04:16
|
NottingHack/hms2
|
https://api.github.com/repos/NottingHack/hms2
|
opened
|
member.ex audit
|
Process
|
After 6 years we should remove ex member records
but to keep DB integrity we likely obfuscate the data
need to look carefully at what tables are have already that need cleaning and as we add more this audit will constantly need updating
going to be best to break this down in `Entity` base files in say `app/HMS/Obfuscate/`
so when adding a new entity we add a new `EntityObfuscate.php` file and the command runs uses auto discovery to find them all and call handle/run methods
|
1.0
|
member.ex audit - After 6 years we should remove ex member records
but to keep DB integrity we likely obfuscate the data
need to look carefully at what tables are have already that need cleaning and as we add more this audit will constantly need updating
going to be best to break this down in `Entity` base files in say `app/HMS/Obfuscate/`
so when adding a new entity we add a new `EntityObfuscate.php` file and the command runs uses auto discovery to find them all and call handle/run methods
|
process
|
member ex audit after years we should remove ex member records but to keep db integrity we likely obfuscate the data need to look carefully at what tables are have already that need cleaning and as we add more this audit will constantly need updating going to be best to break this down in entity base files in say app hms obfuscate so when adding a new entity we add a new entityobfuscate php file and the command runs uses auto discovery to find them all and call handle run methods
| 1
|
247,773
| 26,728,863,316
|
IssuesEvent
|
2023-01-30 01:13:34
|
Yash-Handa/GitHub-Org-Geographics
|
https://api.github.com/repos/Yash-Handa/GitHub-Org-Geographics
|
opened
|
CVE-2022-48285 (Medium) detected in jszip-3.2.1.tgz
|
security vulnerability
|
## CVE-2022-48285 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jszip-3.2.1.tgz</b></p></summary>
<p>Create, read and edit .zip files with JavaScript http://stuartk.com/jszip</p>
<p>Library home page: <a href="https://registry.npmjs.org/jszip/-/jszip-3.2.1.tgz">https://registry.npmjs.org/jszip/-/jszip-3.2.1.tgz</a></p>
<p>Path to dependency file: /GitHub-Org-Geographics/package.json</p>
<p>Path to vulnerable library: /node_modules/jszip/package.json</p>
<p>
Dependency Hierarchy:
- protractor-5.4.2.tgz (Root Library)
- selenium-webdriver-3.6.0.tgz
- :x: **jszip-3.2.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
loadAsync in JSZip before 3.8.0 allows Directory Traversal via a crafted ZIP archive.
<p>Publish Date: 2023-01-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-48285>CVE-2022-48285</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-48285">https://www.cve.org/CVERecord?id=CVE-2022-48285</a></p>
<p>Release Date: 2023-01-29</p>
<p>Fix Resolution (jszip): 3.8.0</p>
<p>Direct dependency fix Resolution (protractor): 5.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-48285 (Medium) detected in jszip-3.2.1.tgz - ## CVE-2022-48285 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jszip-3.2.1.tgz</b></p></summary>
<p>Create, read and edit .zip files with JavaScript http://stuartk.com/jszip</p>
<p>Library home page: <a href="https://registry.npmjs.org/jszip/-/jszip-3.2.1.tgz">https://registry.npmjs.org/jszip/-/jszip-3.2.1.tgz</a></p>
<p>Path to dependency file: /GitHub-Org-Geographics/package.json</p>
<p>Path to vulnerable library: /node_modules/jszip/package.json</p>
<p>
Dependency Hierarchy:
- protractor-5.4.2.tgz (Root Library)
- selenium-webdriver-3.6.0.tgz
- :x: **jszip-3.2.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
loadAsync in JSZip before 3.8.0 allows Directory Traversal via a crafted ZIP archive.
<p>Publish Date: 2023-01-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-48285>CVE-2022-48285</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-48285">https://www.cve.org/CVERecord?id=CVE-2022-48285</a></p>
<p>Release Date: 2023-01-29</p>
<p>Fix Resolution (jszip): 3.8.0</p>
<p>Direct dependency fix Resolution (protractor): 5.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jszip tgz cve medium severity vulnerability vulnerable library jszip tgz create read and edit zip files with javascript library home page a href path to dependency file github org geographics package json path to vulnerable library node modules jszip package json dependency hierarchy protractor tgz root library selenium webdriver tgz x jszip tgz vulnerable library found in base branch master vulnerability details loadasync in jszip before allows directory traversal via a crafted zip archive publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jszip direct dependency fix resolution protractor step up your open source security game with mend
| 0
|
40,398
| 6,822,351,597
|
IssuesEvent
|
2017-11-07 19:46:46
|
flyve-mdm/flyve-mdm-glpi-demo
|
https://api.github.com/repos/flyve-mdm/flyve-mdm-glpi-demo
|
closed
|
Move Wiki articles to project site
|
documentation
|
Hi, @Naylin15
Could you move this article to the Wiki in the project site?
https://github.com/flyve-mdm/flyve-mdm-glpi-demo/wiki/Self_create_an_user_account
Thank you.
|
1.0
|
Move Wiki articles to project site - Hi, @Naylin15
Could you move this article to the Wiki in the project site?
https://github.com/flyve-mdm/flyve-mdm-glpi-demo/wiki/Self_create_an_user_account
Thank you.
|
non_process
|
move wiki articles to project site hi could you move this article to the wiki in the project site thank you
| 0
|
9,172
| 12,225,607,296
|
IssuesEvent
|
2020-05-03 06:25:24
|
labnote-ant/labnote
|
https://api.github.com/repos/labnote-ant/labnote
|
closed
|
Make checkbox for conditions in process
|
process-view
|
Make checkbox for True/False conditions (ex. gradually, heating and etc.) in process.
|
1.0
|
Make checkbox for conditions in process - Make checkbox for True/False conditions (ex. gradually, heating and etc.) in process.
|
process
|
make checkbox for conditions in process make checkbox for true false conditions ex gradually heating and etc in process
| 1
|
10,700
| 13,495,112,155
|
IssuesEvent
|
2020-09-11 22:59:34
|
googleapis/java-secretmanager
|
https://api.github.com/repos/googleapis/java-secretmanager
|
closed
|
Promote to GA
|
api: secretmanager type: process
|
Package name: **google-cloud-secretmanager**
Current release: **beta**
Proposed release: **GA**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [ ] 28 days elapsed since last beta release with new API surface
- [ ] Server API is GA
- [ ] Package API is stable, and we can commit to backward compatibility
- [ ] All dependencies are GA
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [ ] Manual code has been reviewed by API producer
- [ ] Manual code has been reviewed by a DPE responsible for samples
- [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
1.0
|
Promote to GA - Package name: **google-cloud-secretmanager**
Current release: **beta**
Proposed release: **GA**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [ ] 28 days elapsed since last beta release with new API surface
- [ ] Server API is GA
- [ ] Package API is stable, and we can commit to backward compatibility
- [ ] All dependencies are GA
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [ ] Manual code has been reviewed by API producer
- [ ] Manual code has been reviewed by a DPE responsible for samples
- [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
process
|
promote to ga package name google cloud secretmanager current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
| 1
|
15,118
| 18,851,809,719
|
IssuesEvent
|
2021-11-11 22:00:34
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
closed
|
TablesClient: make predict no longer call `get_model` and remove related arguments [breaking change]
|
type: process
|
Currently, for every [`predict`](https://github.com/googleapis/google-cloud-python/blob/master/automl/google/cloud/automl_v1beta1/tables/tables_client.py#L2593-L2603) call, two API requests are made:
1. [`get_model`](https://github.com/googleapis/google-cloud-python/blob/master/automl/google/cloud/automl_v1beta1/tables/tables_client.py#L221-L232)
2. `predict`
`get_model` is costly, often 0.5s, but sometimes 10s in my tests (when the project contains huge number of models).
Planning to never do `get_model` call in `predict`, and the user has to do a manual `get_model` call before making (repeated) predictions. So the cost is explicit and the user will never be caught off guard.
This will be a breaking change, we will need to bump the major version.
CC: @busunkim96 @sirtorry
|
1.0
|
TablesClient: make predict no longer call `get_model` and remove related arguments [breaking change] - Currently, for every [`predict`](https://github.com/googleapis/google-cloud-python/blob/master/automl/google/cloud/automl_v1beta1/tables/tables_client.py#L2593-L2603) call, two API requests are made:
1. [`get_model`](https://github.com/googleapis/google-cloud-python/blob/master/automl/google/cloud/automl_v1beta1/tables/tables_client.py#L221-L232)
2. `predict`
`get_model` is costly, often 0.5s, but sometimes 10s in my tests (when the project contains huge number of models).
Planning to never do `get_model` call in `predict`, and the user has to do a manual `get_model` call before making (repeated) predictions. So the cost is explicit and the user will never be caught off guard.
This will be a breaking change, we will need to bump the major version.
CC: @busunkim96 @sirtorry
|
process
|
tablesclient make predict no longer call get model and remove related arguments currently for every call two api requests are made predict get model is costly often but sometimes in my tests when the project contains huge number of models planning to never do get model call in predict and the user has to do a manual get model call before making repeated predictions so the cost is explicit and the user will never be caught off guard this will be a breaking change we will need to bump the major version cc sirtorry
| 1
|
21,308
| 28,502,253,305
|
IssuesEvent
|
2023-04-18 18:18:03
|
daviddrysdale/python-phonenumbers
|
https://api.github.com/repos/daviddrysdale/python-phonenumbers
|
closed
|
United States area code 557 is not working in python-phonenumbers
|
process
|
This is a python specific issue. The parent repository shows this as a valid phone number:
Python Version: v3.10.8
Library Version: v8.13.9
Phone number area code not working: 557
Parent Repo Validation:

|
1.0
|
United States area code 557 is not working in python-phonenumbers - This is a python specific issue. The parent repository shows this as a valid phone number:
Python Version: v3.10.8
Library Version: v8.13.9
Phone number area code not working: 557
Parent Repo Validation:

|
process
|
united states area code is not working in python phonenumbers this is a python specific issue the parent repository shows this as a valid phone number python version library version phone number area code not working parent repo validation
| 1
|
5,924
| 8,744,130,340
|
IssuesEvent
|
2018-12-12 21:16:43
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
ESS modifes built-in variables in ess-noweb-mode
|
process:eval
|
ESS modifes built-in variables in:
https://github.com/emacs-ess/ESS/blob/d947e772639fb8b5957ea3afbbebff583e9a46cf/lisp/ess-noweb-mode.el#L411-L426
That can break other modes, see
https://github.com/abo-abo/swiper/issues/1660
|
1.0
|
ESS modifes built-in variables in ess-noweb-mode -
ESS modifes built-in variables in:
https://github.com/emacs-ess/ESS/blob/d947e772639fb8b5957ea3afbbebff583e9a46cf/lisp/ess-noweb-mode.el#L411-L426
That can break other modes, see
https://github.com/abo-abo/swiper/issues/1660
|
process
|
ess modifes built in variables in ess noweb mode ess modifes built in variables in that can break other modes see
| 1
|
435,193
| 30,491,596,754
|
IssuesEvent
|
2023-07-18 08:05:20
|
LiskHQ/app-registry
|
https://api.github.com/repos/LiskHQ/app-registry
|
closed
|
Add contributions information to the README
|
documentation type: improvement
|
### Description
Currently, the README doesn't provide instructions for the users to submit a PR with the off-chain information for the sidechain applications.
### Acceptance Criteria
- The contributions section in the README is updated with proper instructions for users to submit a PR
### Additional Information
Ref: https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request-from-a-fork
|
1.0
|
Add contributions information to the README - ### Description
Currently, the README doesn't provide instructions for the users to submit a PR with the off-chain information for the sidechain applications.
### Acceptance Criteria
- The contributions section in the README is updated with proper instructions for users to submit a PR
### Additional Information
Ref: https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request-from-a-fork
|
non_process
|
add contributions information to the readme description currently the readme doesn t provide instructions for the users to submit a pr with the off chain information for the sidechain applications acceptance criteria the contributions section in the readme is updated with proper instructions for users to submit a pr additional information ref
| 0
|
21,800
| 11,387,477,327
|
IssuesEvent
|
2020-01-29 15:05:21
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
opened
|
CustomPaint does not respect isComplex/willChange if painters are null
|
customer: dream (g3) perf: speed severe: performance
|
# Background and reproduction
One of our customer wants to use `CustomPaint` to trigger `RasterCache` by setting `isComplex = true, willChange = false` as below:
```
import 'package:flutter/material.dart';
class CostlyToRasterize extends StatelessWidget {
@override
Widget build(BuildContext context) => Container(color: Colors.blue);
}
void main() => runApp(
MaterialApp(
title: 'RootView',
checkerboardRasterCacheImages: true,
home: PageView.builder(
itemBuilder: (BuildContext _, int index) {
return CustomPaint(
isComplex: true,
willChange: false,
child: CostlyToRasterize(),
);
},
),
),
);
```
(A more realistic app is available in [this gist](https://gist.github.com/liyuqian/8699bf9300ed2c749b4e69710ee43ef4).)
However, the `isComplex` isn't sent to engine and the raster cache isn't triggered as expected.
# Cause and solution
The root cause is the following snippet of code (https://github.com/flutter/flutter/blob/master/packages/flutter/lib/src/rendering/custom_paint.dart#L571)
```
if (_painter != null) {
_paintWithPainter(context.canvas, offset, _painter);
_setRasterCacheHints(context);
}
```
which ignores the raster cache hints if painter is null.
To fix this, we could do either of the following:
1. Update our API doc to specify that raster cache hints are ignored if painters aren't provided. Hence the user needs to provide at least a dummy painter to trigger the raster cache hints.
2. Change our code to always `_setRasterCacheHints` no matter whether painter exists.
CC @Hixie and @goderbauer for suggestions about which solution should be adopted.
@ignatz: for the time being, you should be able to trigger the raster cache by providing a dummy painter like the following
```
import 'package:flutter/material.dart';
class CostlyToRasterize extends StatelessWidget {
@override
Widget build(BuildContext context) => Container(color: Colors.blue);
}
void main() => runApp(
MaterialApp(
title: 'RootView',
checkerboardRasterCacheImages: true,
home: PageView.builder(
itemBuilder: (BuildContext _, int index) {
return CustomPaint(
painter: EmptyPainter(),
isComplex: true,
willChange: false,
child: CostlyToRasterize(),
);
},
),
),
);
class EmptyPainter extends CustomPainter {
@override
void paint(Canvas canvas, Size size) {}
@override
bool shouldRepaint(CustomPainter oldDelegate) => false;
}
```
|
True
|
CustomPaint does not respect isComplex/willChange if painters are null - # Background and reproduction
One of our customer wants to use `CustomPaint` to trigger `RasterCache` by setting `isComplex = true, willChange = false` as below:
```
import 'package:flutter/material.dart';
class CostlyToRasterize extends StatelessWidget {
@override
Widget build(BuildContext context) => Container(color: Colors.blue);
}
void main() => runApp(
MaterialApp(
title: 'RootView',
checkerboardRasterCacheImages: true,
home: PageView.builder(
itemBuilder: (BuildContext _, int index) {
return CustomPaint(
isComplex: true,
willChange: false,
child: CostlyToRasterize(),
);
},
),
),
);
```
(A more realistic app is available in [this gist](https://gist.github.com/liyuqian/8699bf9300ed2c749b4e69710ee43ef4).)
However, the `isComplex` isn't sent to engine and the raster cache isn't triggered as expected.
# Cause and solution
The root cause is the following snippet of code (https://github.com/flutter/flutter/blob/master/packages/flutter/lib/src/rendering/custom_paint.dart#L571)
```
if (_painter != null) {
_paintWithPainter(context.canvas, offset, _painter);
_setRasterCacheHints(context);
}
```
which ignores the raster cache hints if painter is null.
To fix this, we could do either of the following:
1. Update our API doc to specify that raster cache hints are ignored if painters aren't provided. Hence the user needs to provide at least a dummy painter to trigger the raster cache hints.
2. Change our code to always `_setRasterCacheHints` no matter whether painter exists.
CC @Hixie and @goderbauer for suggestions about which solution should be adopted.
@ignatz: for the time being, you should be able to trigger the raster cache by providing a dummy painter like the following
```
import 'package:flutter/material.dart';
class CostlyToRasterize extends StatelessWidget {
@override
Widget build(BuildContext context) => Container(color: Colors.blue);
}
void main() => runApp(
MaterialApp(
title: 'RootView',
checkerboardRasterCacheImages: true,
home: PageView.builder(
itemBuilder: (BuildContext _, int index) {
return CustomPaint(
painter: EmptyPainter(),
isComplex: true,
willChange: false,
child: CostlyToRasterize(),
);
},
),
),
);
class EmptyPainter extends CustomPainter {
@override
void paint(Canvas canvas, Size size) {}
@override
bool shouldRepaint(CustomPainter oldDelegate) => false;
}
```
|
non_process
|
custompaint does not respect iscomplex willchange if painters are null background and reproduction one of our customer wants to use custompaint to trigger rastercache by setting iscomplex true willchange false as below import package flutter material dart class costlytorasterize extends statelesswidget override widget build buildcontext context container color colors blue void main runapp materialapp title rootview checkerboardrastercacheimages true home pageview builder itembuilder buildcontext int index return custompaint iscomplex true willchange false child costlytorasterize a more realistic app is available in however the iscomplex isn t sent to engine and the raster cache isn t triggered as expected cause and solution the root cause is the following snippet of code if painter null paintwithpainter context canvas offset painter setrastercachehints context which ignores the raster cache hints if painter is null to fix this we could do either of the following update our api doc to specify that raster cache hints are ignored if painters aren t provided hence the user needs to provide at least a dummy painter to trigger the raster cache hints change our code to always setrastercachehints no matter whether painter exists cc hixie and goderbauer for suggestions about which solution should be adopted ignatz for the time being you should be able to trigger the raster cache by providing a dummy painter like the following import package flutter material dart class costlytorasterize extends statelesswidget override widget build buildcontext context container color colors blue void main runapp materialapp title rootview checkerboardrastercacheimages true home pageview builder itembuilder buildcontext int index return custompaint painter emptypainter iscomplex true willchange false child costlytorasterize class emptypainter extends custompainter override void paint canvas canvas size size override bool shouldrepaint custompainter olddelegate false
| 0
|
12,298
| 14,854,473,458
|
IssuesEvent
|
2021-01-18 11:21:51
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
Error messages should be updated as per document
|
Bug P1 Participant manager Process: Fixed Process: Tested dev
|
Error messages should be updated as per following document
https://docs.google.com/spreadsheets/d/1CEcgbG2Et3FCc5qbavewlNQapxNHG5kTSNMFYzdz4nc/edit?usp=sharing
|
2.0
|
Error messages should be updated as per document - Error messages should be updated as per following document
https://docs.google.com/spreadsheets/d/1CEcgbG2Et3FCc5qbavewlNQapxNHG5kTSNMFYzdz4nc/edit?usp=sharing
|
process
|
error messages should be updated as per document error messages should be updated as per following document
| 1
|
27,062
| 4,867,181,025
|
IssuesEvent
|
2016-11-15 03:03:29
|
TNGSB/eWallet
|
https://api.github.com/repos/TNGSB/eWallet
|
closed
|
e-Wallet_WebAdmin 07112016
|
Defect - High (Sev-2)
|
As for today testing, there are 3 defects found in web admin module.
-Card Management
-Customer Service
Kindly refer the attached defects document for your perusal.
[Sev_2.zip](https://github.com/TNGSB/eWallet/files/574946/Sev_2.zip)
|
1.0
|
e-Wallet_WebAdmin 07112016 - As for today testing, there are 3 defects found in web admin module.
-Card Management
-Customer Service
Kindly refer the attached defects document for your perusal.
[Sev_2.zip](https://github.com/TNGSB/eWallet/files/574946/Sev_2.zip)
|
non_process
|
e wallet webadmin as for today testing there are defects found in web admin module card management customer service kindly refer the attached defects document for your perusal
| 0
|
21,947
| 30,448,636,913
|
IssuesEvent
|
2023-07-16 02:00:07
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Fri, 14 Jul 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### Temporal Label-Refinement for Weakly-Supervised Audio-Visual Event Localization
- **Authors:** Kalyan Ramakrishnan
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Sound (cs.SD); Audio and Speech Processing (eess.AS)
- **Arxiv link:** https://arxiv.org/abs/2307.06385
- **Pdf link:** https://arxiv.org/pdf/2307.06385
- **Abstract**
Audio-Visual Event Localization (AVEL) is the task of temporally localizing and classifying \emph{audio-visual events}, i.e., events simultaneously visible and audible in a video. In this paper, we solve AVEL in a weakly-supervised setting, where only video-level event labels (their presence/absence, but not their locations in time) are available as supervision for training. Our idea is to use a base model to estimate labels on the training data at a finer temporal resolution than at the video level and re-train the model with these labels. I.e., we determine the subset of labels for each \emph{slice} of frames in a training video by (i) replacing the frames outside the slice with those from a second video having no overlap in video-level labels, and (ii) feeding this synthetic video into the base model to extract labels for just the slice in question. To handle the out-of-distribution nature of our synthetic videos, we propose an auxiliary objective for the base model that induces more reliable predictions of the localized event labels as desired. Our three-stage pipeline outperforms several existing AVEL methods with no architectural changes and improves performance on a related weakly-supervised task as well.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### ConvNeXt-ChARM: ConvNeXt-based Transform for Efficient Neural Image Compression
- **Authors:** Ahmed Ghorbel, Wassim Hamidouche, Luce Morin
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.06342
- **Pdf link:** https://arxiv.org/pdf/2307.06342
- **Abstract**
Over the last few years, neural image compression has gained wide attention from research and industry, yielding promising end-to-end deep neural codecs outperforming their conventional counterparts in rate-distortion performance. Despite significant advancement, current methods, including attention-based transform coding, still need to be improved in reducing the coding rate while preserving the reconstruction fidelity, especially in non-homogeneous textured image areas. Those models also require more parameters and a higher decoding time. To tackle the above challenges, we propose ConvNeXt-ChARM, an efficient ConvNeXt-based transform coding framework, paired with a compute-efficient channel-wise auto-regressive prior to capturing both global and local contexts from the hyper and quantized latent representations. The proposed architecture can be optimized end-to-end to fully exploit the context information and extract compact latent representation while reconstructing higher-quality images. Experimental results on four widely-used datasets showed that ConvNeXt-ChARM brings consistent and significant BD-rate (PSNR) reductions estimated on average to 5.24% and 1.22% over the versatile video coding (VVC) reference encoder (VTM-18.0) and the state-of-the-art learned image compression method SwinT-ChARM, respectively. Moreover, we provide model scaling studies to verify the computational efficiency of our approach and conduct several objective and subjective analyses to bring to the fore the performance gap between the next generation ConvNet, namely ConvNeXt, and Swin Transformer.
## Keyword: RAW
### Image Transformation Sequence Retrieval with General Reinforcement Learning
- **Authors:** Enrique Mas-Candela, Antonio Ríos-Vila, Jorge Calvo-Zaragoza
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.06630
- **Pdf link:** https://arxiv.org/pdf/2307.06630
- **Abstract**
In this work, the novel Image Transformation Sequence Retrieval (ITSR) task is presented, in which a model must retrieve the sequence of transformations between two given images that act as source and target, respectively. Given certain characteristics of the challenge such as the multiplicity of a correct sequence or the correlation between consecutive steps of the process, we propose a solution to ITSR using a general model-based Reinforcement Learning such as Monte Carlo Tree Search (MCTS), which is combined with a deep neural network. Our experiments provide a benchmark in both synthetic and real domains, where the proposed approach is compared with supervised training. The results report that a model trained with MCTS is able to outperform its supervised counterpart in both the simplest and the most complex cases. Our work draws interesting conclusions about the nature of ITSR and its associated challenges.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Fri, 14 Jul 23 - ## Keyword: events
### Temporal Label-Refinement for Weakly-Supervised Audio-Visual Event Localization
- **Authors:** Kalyan Ramakrishnan
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Sound (cs.SD); Audio and Speech Processing (eess.AS)
- **Arxiv link:** https://arxiv.org/abs/2307.06385
- **Pdf link:** https://arxiv.org/pdf/2307.06385
- **Abstract**
Audio-Visual Event Localization (AVEL) is the task of temporally localizing and classifying \emph{audio-visual events}, i.e., events simultaneously visible and audible in a video. In this paper, we solve AVEL in a weakly-supervised setting, where only video-level event labels (their presence/absence, but not their locations in time) are available as supervision for training. Our idea is to use a base model to estimate labels on the training data at a finer temporal resolution than at the video level and re-train the model with these labels. I.e., we determine the subset of labels for each \emph{slice} of frames in a training video by (i) replacing the frames outside the slice with those from a second video having no overlap in video-level labels, and (ii) feeding this synthetic video into the base model to extract labels for just the slice in question. To handle the out-of-distribution nature of our synthetic videos, we propose an auxiliary objective for the base model that induces more reliable predictions of the localized event labels as desired. Our three-stage pipeline outperforms several existing AVEL methods with no architectural changes and improves performance on a related weakly-supervised task as well.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### ConvNeXt-ChARM: ConvNeXt-based Transform for Efficient Neural Image Compression
- **Authors:** Ahmed Ghorbel, Wassim Hamidouche, Luce Morin
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.06342
- **Pdf link:** https://arxiv.org/pdf/2307.06342
- **Abstract**
Over the last few years, neural image compression has gained wide attention from research and industry, yielding promising end-to-end deep neural codecs outperforming their conventional counterparts in rate-distortion performance. Despite significant advancement, current methods, including attention-based transform coding, still need to be improved in reducing the coding rate while preserving the reconstruction fidelity, especially in non-homogeneous textured image areas. Those models also require more parameters and a higher decoding time. To tackle the above challenges, we propose ConvNeXt-ChARM, an efficient ConvNeXt-based transform coding framework, paired with a compute-efficient channel-wise auto-regressive prior to capturing both global and local contexts from the hyper and quantized latent representations. The proposed architecture can be optimized end-to-end to fully exploit the context information and extract compact latent representation while reconstructing higher-quality images. Experimental results on four widely-used datasets showed that ConvNeXt-ChARM brings consistent and significant BD-rate (PSNR) reductions estimated on average to 5.24% and 1.22% over the versatile video coding (VVC) reference encoder (VTM-18.0) and the state-of-the-art learned image compression method SwinT-ChARM, respectively. Moreover, we provide model scaling studies to verify the computational efficiency of our approach and conduct several objective and subjective analyses to bring to the fore the performance gap between the next generation ConvNet, namely ConvNeXt, and Swin Transformer.
## Keyword: RAW
### Image Transformation Sequence Retrieval with General Reinforcement Learning
- **Authors:** Enrique Mas-Candela, Antonio Ríos-Vila, Jorge Calvo-Zaragoza
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.06630
- **Pdf link:** https://arxiv.org/pdf/2307.06630
- **Abstract**
In this work, the novel Image Transformation Sequence Retrieval (ITSR) task is presented, in which a model must retrieve the sequence of transformations between two given images that act as source and target, respectively. Given certain characteristics of the challenge such as the multiplicity of a correct sequence or the correlation between consecutive steps of the process, we propose a solution to ITSR using a general model-based Reinforcement Learning such as Monte Carlo Tree Search (MCTS), which is combined with a deep neural network. Our experiments provide a benchmark in both synthetic and real domains, where the proposed approach is compared with supervised training. The results report that a model trained with MCTS is able to outperform its supervised counterpart in both the simplest and the most complex cases. Our work draws interesting conclusions about the nature of ITSR and its associated challenges.
## Keyword: raw image
There is no result
|
process
|
new submissions for fri jul keyword events temporal label refinement for weakly supervised audio visual event localization authors kalyan ramakrishnan subjects computer vision and pattern recognition cs cv machine learning cs lg sound cs sd audio and speech processing eess as arxiv link pdf link abstract audio visual event localization avel is the task of temporally localizing and classifying emph audio visual events i e events simultaneously visible and audible in a video in this paper we solve avel in a weakly supervised setting where only video level event labels their presence absence but not their locations in time are available as supervision for training our idea is to use a base model to estimate labels on the training data at a finer temporal resolution than at the video level and re train the model with these labels i e we determine the subset of labels for each emph slice of frames in a training video by i replacing the frames outside the slice with those from a second video having no overlap in video level labels and ii feeding this synthetic video into the base model to extract labels for just the slice in question to handle the out of distribution nature of our synthetic videos we propose an auxiliary objective for the base model that induces more reliable predictions of the localized event labels as desired our three stage pipeline outperforms several existing avel methods with no architectural changes and improves performance on a related weakly supervised task as well keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression convnext charm convnext based transform for efficient neural image compression authors ahmed ghorbel wassim hamidouche luce morin subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract over the last few years neural image compression has gained wide attention from research and industry yielding promising end to end deep neural codecs outperforming their conventional counterparts in rate distortion performance despite significant advancement current methods including attention based transform coding still need to be improved in reducing the coding rate while preserving the reconstruction fidelity especially in non homogeneous textured image areas those models also require more parameters and a higher decoding time to tackle the above challenges we propose convnext charm an efficient convnext based transform coding framework paired with a compute efficient channel wise auto regressive prior to capturing both global and local contexts from the hyper and quantized latent representations the proposed architecture can be optimized end to end to fully exploit the context information and extract compact latent representation while reconstructing higher quality images experimental results on four widely used datasets showed that convnext charm brings consistent and significant bd rate psnr reductions estimated on average to and over the versatile video coding vvc reference encoder vtm and the state of the art learned image compression method swint charm respectively moreover we provide model scaling studies to verify the computational efficiency of our approach and conduct several objective and subjective analyses to bring to the fore the performance gap between the next generation convnet namely convnext and swin transformer keyword raw image transformation sequence retrieval with general reinforcement learning authors enrique mas candela antonio ríos vila jorge calvo zaragoza subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract in this work the novel image transformation sequence retrieval itsr task is presented in which a model must retrieve the sequence of transformations between two given images that act as source and target respectively given certain characteristics of the challenge such as the multiplicity of a correct sequence or the correlation between consecutive steps of the process we propose a solution to itsr using a general model based reinforcement learning such as monte carlo tree search mcts which is combined with a deep neural network our experiments provide a benchmark in both synthetic and real domains where the proposed approach is compared with supervised training the results report that a model trained with mcts is able to outperform its supervised counterpart in both the simplest and the most complex cases our work draws interesting conclusions about the nature of itsr and its associated challenges keyword raw image there is no result
| 1
|
8,892
| 11,986,967,840
|
IssuesEvent
|
2020-04-07 20:18:32
|
googleapis/google-cloud-go
|
https://api.github.com/repos/googleapis/google-cloud-go
|
closed
|
internal/gapicgen: Go-based binary dependencies are using master instead of latest release
|
type: process
|
We've noticed that the `internal/gapicgen` system is pulling `master` of Go-based binary dependencies i.e. `protoc-gen-go_gapic` and `goimports`, rather than the version specified in `internal/gapicgen/go.mod`.
We need to ensure that the pinned versions are respected by the generation tooling so that we don't introduce unintended changes.
|
1.0
|
internal/gapicgen: Go-based binary dependencies are using master instead of latest release - We've noticed that the `internal/gapicgen` system is pulling `master` of Go-based binary dependencies i.e. `protoc-gen-go_gapic` and `goimports`, rather than the version specified in `internal/gapicgen/go.mod`.
We need to ensure that the pinned versions are respected by the generation tooling so that we don't introduce unintended changes.
|
process
|
internal gapicgen go based binary dependencies are using master instead of latest release we ve noticed that the internal gapicgen system is pulling master of go based binary dependencies i e protoc gen go gapic and goimports rather than the version specified in internal gapicgen go mod we need to ensure that the pinned versions are respected by the generation tooling so that we don t introduce unintended changes
| 1
|
40,120
| 6,800,157,238
|
IssuesEvent
|
2017-11-02 13:04:50
|
onury/geolocator
|
https://api.github.com/repos/onury/geolocator
|
closed
|
Demo pages not using HTTPS
|
documentation
|
Hello,
As stated in your own README:
> Make sure you're calling Geolocation APIs (such as geolocator.locate() and geolocator.watch()) from a secure origin (i.e. an HTTPS page). In Chrome 50, Geolocation API is removed from unsecured origins. Other browsers are expected to follow.
Since the demo is not using HTTPS, the demo doesn't ask for location in Google Chrome.
Please put the demo pages under HTTPS, its free using Let's Encrypt now!
|
1.0
|
Demo pages not using HTTPS - Hello,
As stated in your own README:
> Make sure you're calling Geolocation APIs (such as geolocator.locate() and geolocator.watch()) from a secure origin (i.e. an HTTPS page). In Chrome 50, Geolocation API is removed from unsecured origins. Other browsers are expected to follow.
Since the demo is not using HTTPS, the demo doesn't ask for location in Google Chrome.
Please put the demo pages under HTTPS, its free using Let's Encrypt now!
|
non_process
|
demo pages not using https hello as stated in your own readme make sure you re calling geolocation apis such as geolocator locate and geolocator watch from a secure origin i e an https page in chrome geolocation api is removed from unsecured origins other browsers are expected to follow since the demo is not using https the demo doesn t ask for location in google chrome please put the demo pages under https its free using let s encrypt now
| 0
|
12,482
| 14,950,090,914
|
IssuesEvent
|
2021-01-26 12:33:46
|
panther-labs/panther
|
https://api.github.com/repos/panther-labs/panther
|
opened
|
Expand AWS ALB log parser to include Desync mitigation mode fields
|
data inputs story team:data processing
|
### Description
[AWS ALB access logs](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/load-balancer-access-logs.html#access-log-file-format) include 4 additional fields that the current [log parser](https://github.com/panther-labs/panther/blob/2a27d25374cde5ab88d67142d3ebc42568d2dd94/internal/log_analysis/log_processor/parsers/awslogs/alb.go#L37) seems to ignore:
- `"target:port_list"`
- `"target_status_code_list"`
- `"classification"`
- `"classification_reason"`
The last two fields seem particularly important in the security context since they are specified by the [Desync mitigation mode](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/application-load-balancers.html#desync-mitigation-mode) protection.
> Desync mitigation mode protects your application from issues due to HTTP Desync. The load balancer classifies each request based on its threat level, allows safe requests, and then mitigates risk as specified by the mitigation mode that you specify. The desync mitigation modes are monitor, defensive, and strictest.
### Additional References
- https://github.com/aws/http-desync-guardian
- https://aws.amazon.com/about-aws/whats-new/2020/08/application-and-classic-load-balancers-adding-defense-in-depth-with-introduction-of-desync-mitigation-mode/
### Related Services
- Log Processing / ALB Log Parser
- AWS Glue metadata for ALB access log table
### Acceptance Criteria
- Parse AWS ALB log records with the 4 additional fields.
- Output the additional fields accordingly and make them available for rules.
|
1.0
|
Expand AWS ALB log parser to include Desync mitigation mode fields - ### Description
[AWS ALB access logs](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/load-balancer-access-logs.html#access-log-file-format) include 4 additional fields that the current [log parser](https://github.com/panther-labs/panther/blob/2a27d25374cde5ab88d67142d3ebc42568d2dd94/internal/log_analysis/log_processor/parsers/awslogs/alb.go#L37) seems to ignore:
- `"target:port_list"`
- `"target_status_code_list"`
- `"classification"`
- `"classification_reason"`
The last two fields seem particularly important in the security context since they are specified by the [Desync mitigation mode](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/application-load-balancers.html#desync-mitigation-mode) protection.
> Desync mitigation mode protects your application from issues due to HTTP Desync. The load balancer classifies each request based on its threat level, allows safe requests, and then mitigates risk as specified by the mitigation mode that you specify. The desync mitigation modes are monitor, defensive, and strictest.
### Additional References
- https://github.com/aws/http-desync-guardian
- https://aws.amazon.com/about-aws/whats-new/2020/08/application-and-classic-load-balancers-adding-defense-in-depth-with-introduction-of-desync-mitigation-mode/
### Related Services
- Log Processing / ALB Log Parser
- AWS Glue metadata for ALB access log table
### Acceptance Criteria
- Parse AWS ALB log records with the 4 additional fields.
- Output the additional fields accordingly and make them available for rules.
|
process
|
expand aws alb log parser to include desync mitigation mode fields description include additional fields that the current seems to ignore target port list target status code list classification classification reason the last two fields seem particularly important in the security context since they are specified by the protection desync mitigation mode protects your application from issues due to http desync the load balancer classifies each request based on its threat level allows safe requests and then mitigates risk as specified by the mitigation mode that you specify the desync mitigation modes are monitor defensive and strictest additional references related services log processing alb log parser aws glue metadata for alb access log table acceptance criteria parse aws alb log records with the additional fields output the additional fields accordingly and make them available for rules
| 1
|
1,193
| 3,690,399,345
|
IssuesEvent
|
2016-02-25 19:51:34
|
hoodiehq/editorial
|
https://api.github.com/repos/hoodiehq/editorial
|
closed
|
Define a formal method in which contributors are added to the Editorial team
|
process
|
I updated the `team-roles.md` file with descriptions in https://github.com/hoodiehq/editorial/pull/23. In addition to adding the descriptions, I think we need to define a formal process in which contributors are added to the Editorials team.
Having this early on makes the transition from one (or few) contributor to many slide through with much less friction than when it is developed later.
Tasks for getting this done:
- [x] Separate file or put in `team-roles.md`?
- [x] Different path to contributor for each of the defined roles in `team-roles.md` or a unified path?
- [x] ~~Define what `active` and `contribution` mean.~~
- [ ] Link to CoC, `CONTRIBUTING.md`, and anything else?
- [ ] {{other points?}}
|
1.0
|
Define a formal method in which contributors are added to the Editorial team - I updated the `team-roles.md` file with descriptions in https://github.com/hoodiehq/editorial/pull/23. In addition to adding the descriptions, I think we need to define a formal process in which contributors are added to the Editorials team.
Having this early on makes the transition from one (or few) contributor to many slide through with much less friction than when it is developed later.
Tasks for getting this done:
- [x] Separate file or put in `team-roles.md`?
- [x] Different path to contributor for each of the defined roles in `team-roles.md` or a unified path?
- [x] ~~Define what `active` and `contribution` mean.~~
- [ ] Link to CoC, `CONTRIBUTING.md`, and anything else?
- [ ] {{other points?}}
|
process
|
define a formal method in which contributors are added to the editorial team i updated the team roles md file with descriptions in in addition to adding the descriptions i think we need to define a formal process in which contributors are added to the editorials team having this early on makes the transition from one or few contributor to many slide through with much less friction than when it is developed later tasks for getting this done separate file or put in team roles md different path to contributor for each of the defined roles in team roles md or a unified path define what active and contribution mean link to coc contributing md and anything else other points
| 1
|
3,559
| 3,955,259,057
|
IssuesEvent
|
2016-04-29 20:12:21
|
commercialhaskell/stack
|
https://api.github.com/repos/commercialhaskell/stack
|
closed
|
Avoid unnecessarily loading the hackage index
|
awaiting pr help wanted type: performance
|
Loading the hackage is expensive, because the encoded index cache is 18Mb. On my fairly fast computer, it takes 0.3 seconds to decode it. Search `getPackageCaches` to find the spots it's used.
There are two cases where loading the hackage currently happens unnecessarily:
1) In `loadSourceMap`, because the user might have specified a package target on the commandline which is not in the resolver. In this case, it uses the latest from hackage.
2) In `constructPlan`, because plan construction errors get annotated with "latest applicable" versions.
For both of these, loading the hackage index is only necessary under specific and abormal circumstances. However, currently this expensive operation is being done regardless of need.
It's also used in `resolveBuildPlan`, but that only seems to be used by the tests.
|
True
|
Avoid unnecessarily loading the hackage index - Loading the hackage is expensive, because the encoded index cache is 18Mb. On my fairly fast computer, it takes 0.3 seconds to decode it. Search `getPackageCaches` to find the spots it's used.
There are two cases where loading the hackage currently happens unnecessarily:
1) In `loadSourceMap`, because the user might have specified a package target on the commandline which is not in the resolver. In this case, it uses the latest from hackage.
2) In `constructPlan`, because plan construction errors get annotated with "latest applicable" versions.
For both of these, loading the hackage index is only necessary under specific and abormal circumstances. However, currently this expensive operation is being done regardless of need.
It's also used in `resolveBuildPlan`, but that only seems to be used by the tests.
|
non_process
|
avoid unnecessarily loading the hackage index loading the hackage is expensive because the encoded index cache is on my fairly fast computer it takes seconds to decode it search getpackagecaches to find the spots it s used there are two cases where loading the hackage currently happens unnecessarily in loadsourcemap because the user might have specified a package target on the commandline which is not in the resolver in this case it uses the latest from hackage in constructplan because plan construction errors get annotated with latest applicable versions for both of these loading the hackage index is only necessary under specific and abormal circumstances however currently this expensive operation is being done regardless of need it s also used in resolvebuildplan but that only seems to be used by the tests
| 0
|
248,990
| 26,870,749,324
|
IssuesEvent
|
2023-02-04 12:35:24
|
turkdevops/electron-api-demos
|
https://api.github.com/repos/turkdevops/electron-api-demos
|
closed
|
CVE-2022-25881 (Medium) detected in http-cache-semantics-4.1.0.tgz - autoclosed
|
security vulnerability
|
## CVE-2022-25881 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-cache-semantics-4.1.0.tgz</b></p></summary>
<p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz</a></p>
<p>
Dependency Hierarchy:
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/electron-api-demos/commit/8b3c67fde2016f47e681b745f49afdea23a50ed4">8b3c67fde2016f47e681b745f49afdea23a50ed4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library.
<p>Publish Date: 2023-01-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p>
<p>Release Date: 2023-01-31</p>
<p>Fix Resolution: http-cache-semantics - 4.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-25881 (Medium) detected in http-cache-semantics-4.1.0.tgz - autoclosed - ## CVE-2022-25881 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-cache-semantics-4.1.0.tgz</b></p></summary>
<p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz</a></p>
<p>
Dependency Hierarchy:
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/electron-api-demos/commit/8b3c67fde2016f47e681b745f49afdea23a50ed4">8b3c67fde2016f47e681b745f49afdea23a50ed4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library.
<p>Publish Date: 2023-01-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p>
<p>Release Date: 2023-01-31</p>
<p>Fix Resolution: http-cache-semantics - 4.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in http cache semantics tgz autoclosed cve medium severity vulnerability vulnerable library http cache semantics tgz parses cache control and other headers helps building correct http caches and proxies library home page a href dependency hierarchy found in head commit a href found in base branch master vulnerability details this affects versions of the package http cache semantics before the issue can be exploited via malicious request header values sent to a server when that server reads the cache policy from the request using this library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution http cache semantics step up your open source security game with mend
| 0
|
16,413
| 21,191,507,959
|
IssuesEvent
|
2022-04-08 17:59:19
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Webkit: Run internal tests across Webkit browser
|
process: tests type: chore browser: webkit stage: icebox
|
In addition to Chrome and Firefox, run our internal tests across Webkit browser.
Skip any tests in the Webkit browser that do not currently pass.
|
1.0
|
Webkit: Run internal tests across Webkit browser - In addition to Chrome and Firefox, run our internal tests across Webkit browser.
Skip any tests in the Webkit browser that do not currently pass.
|
process
|
webkit run internal tests across webkit browser in addition to chrome and firefox run our internal tests across webkit browser skip any tests in the webkit browser that do not currently pass
| 1
|
299,408
| 25,901,504,710
|
IssuesEvent
|
2022-12-15 06:18:48
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: liquibase failed
|
C-test-failure O-robot O-roachtest T-sql-sessions branch-release-22.2.0
|
roachtest.liquibase [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6784087?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6784087?buildTab=artifacts#/liquibase) on release-22.2.0 @ [9bb8a7faf624dcd23ce60e2a8a805ef863b10f72](https://github.com/cockroachdb/cockroach/commits/9bb8a7faf624dcd23ce60e2a8a805ef863b10f72):
```
test artifacts and logs in: /artifacts/liquibase/run_1
orm_helpers.go:191,orm_helpers.go:117,java_helpers.go:220,liquibase.go:123,liquibase.go:140,test_runner.go:930:
Tests run on Cockroach v22.2.0-beta.2-106-g9bb8a7faf6
Tests run against liquibase 1790ddef2d0339c5c96839ac60ac424c130dadd8
47 Total Tests Run
38 tests passed
9 tests failed
17 tests skipped
0 tests ignored
0 tests passed unexpectedly
6 tests failed unexpectedly
0 tests expected failed but skipped
0 tests expected failed but not run
---
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply createSequence against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply addDefaultValueSequenceNext against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply alterSequence against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply addAutoIncrement against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply dropSequence against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply renameSequence against cockroachdb 20.2 - unknown (unexpected)
For a full summary look at the liquibase artifacts
An updated blocklist (liquibaseBlocklist22_2) is available in the artifacts' liquibase log
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #87586 roachtest: liquibase failed [C-test-failure O-roachtest O-robot branch-release-22.2]
- #83313 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.2]
- #83050 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-master]
- #83048 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*liquibase.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-20236
|
2.0
|
roachtest: liquibase failed - roachtest.liquibase [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6784087?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6784087?buildTab=artifacts#/liquibase) on release-22.2.0 @ [9bb8a7faf624dcd23ce60e2a8a805ef863b10f72](https://github.com/cockroachdb/cockroach/commits/9bb8a7faf624dcd23ce60e2a8a805ef863b10f72):
```
test artifacts and logs in: /artifacts/liquibase/run_1
orm_helpers.go:191,orm_helpers.go:117,java_helpers.go:220,liquibase.go:123,liquibase.go:140,test_runner.go:930:
Tests run on Cockroach v22.2.0-beta.2-106-g9bb8a7faf6
Tests run against liquibase 1790ddef2d0339c5c96839ac60ac424c130dadd8
47 Total Tests Run
38 tests passed
9 tests failed
17 tests skipped
0 tests ignored
0 tests passed unexpectedly
6 tests failed unexpectedly
0 tests expected failed but skipped
0 tests expected failed but not run
---
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply createSequence against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply addDefaultValueSequenceNext against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply alterSequence against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply addAutoIncrement against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply dropSequence against cockroachdb 20.2 - unknown (unexpected)
--- FAIL: liquibase.harness.change.ChangeObjectTests.apply renameSequence against cockroachdb 20.2 - unknown (unexpected)
For a full summary look at the liquibase artifacts
An updated blocklist (liquibaseBlocklist22_2) is available in the artifacts' liquibase log
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #87586 roachtest: liquibase failed [C-test-failure O-roachtest O-robot branch-release-22.2]
- #83313 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.2]
- #83050 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-master]
- #83048 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*liquibase.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-20236
|
non_process
|
roachtest liquibase failed roachtest liquibase with on release test artifacts and logs in artifacts liquibase run orm helpers go orm helpers go java helpers go liquibase go liquibase go test runner go tests run on cockroach beta tests run against liquibase total tests run tests passed tests failed tests skipped tests ignored tests passed unexpectedly tests failed unexpectedly tests expected failed but skipped tests expected failed but not run fail liquibase harness change changeobjecttests apply createsequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply adddefaultvaluesequencenext against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply altersequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply addautoincrement against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply dropsequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply renamesequence against cockroachdb unknown unexpected for a full summary look at the liquibase artifacts an updated blocklist is available in the artifacts liquibase log parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest liquibase failed roachtest liquibase failed roachtest liquibase failed roachtest liquibase failed cc cockroachdb sql experience jira issue crdb
| 0
|
378,623
| 11,205,407,359
|
IssuesEvent
|
2020-01-05 14:04:23
|
kubernetes-sigs/kubefed
|
https://api.github.com/repos/kubernetes-sigs/kubefed
|
closed
|
federation controller disaster recovery Support
|
kind/feature lifecycle/rotten priority/backlog
|
<!-- Please only use this template for submitting enhancement requests -->
**What would you like to be added**:
Currently, we can only have one host cluster, but if the host cluster goes down, then the federation controller will also goes down, and then the federation will not work.
**Why is this needed**:
It is better to enable the federation controller can support disaster recovery mode, such as we can introduce active-passive mode for federation controller, if the primary controller goes down, then another controller will take over to manage all of the clusters.
<!-- DO NOT EDIT BELOW THIS LINE -->
/kind feature
/cc @marun @kubernetes-sigs/federation-wg
|
1.0
|
federation controller disaster recovery Support - <!-- Please only use this template for submitting enhancement requests -->
**What would you like to be added**:
Currently, we can only have one host cluster, but if the host cluster goes down, then the federation controller will also goes down, and then the federation will not work.
**Why is this needed**:
It is better to enable the federation controller can support disaster recovery mode, such as we can introduce active-passive mode for federation controller, if the primary controller goes down, then another controller will take over to manage all of the clusters.
<!-- DO NOT EDIT BELOW THIS LINE -->
/kind feature
/cc @marun @kubernetes-sigs/federation-wg
|
non_process
|
federation controller disaster recovery support what would you like to be added currently we can only have one host cluster but if the host cluster goes down then the federation controller will also goes down and then the federation will not work why is this needed it is better to enable the federation controller can support disaster recovery mode such as we can introduce active passive mode for federation controller if the primary controller goes down then another controller will take over to manage all of the clusters kind feature cc marun kubernetes sigs federation wg
| 0
|
700
| 3,197,184,575
|
IssuesEvent
|
2015-10-01 02:00:35
|
18F/CMS.gov-developer
|
https://api.github.com/repos/18F/CMS.gov-developer
|
closed
|
Plan training dates
|
process
|
* [ ] CMS to consider what kind of trainings and with you.
* [ ] CMS to pick tentative dates and times.
* [ ] Add these to the [scheduled trainings document](https://github.com/18F/CMS.gov-developer/blob/master/deliverables/draft/training_schedule.md).
|
1.0
|
Plan training dates - * [ ] CMS to consider what kind of trainings and with you.
* [ ] CMS to pick tentative dates and times.
* [ ] Add these to the [scheduled trainings document](https://github.com/18F/CMS.gov-developer/blob/master/deliverables/draft/training_schedule.md).
|
process
|
plan training dates cms to consider what kind of trainings and with you cms to pick tentative dates and times add these to the
| 1
|
174,844
| 21,300,487,115
|
IssuesEvent
|
2022-04-15 01:59:18
|
YJSoft/nedb
|
https://api.github.com/repos/YJSoft/nedb
|
closed
|
WS-2017-0247 (Low) detected in ms-0.3.0.tgz - autoclosed
|
security vulnerability
|
## WS-2017-0247 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ms-0.3.0.tgz</b></p></summary>
<p>Tiny ms conversion utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.3.0.tgz">https://registry.npmjs.org/ms/-/ms-0.3.0.tgz</a></p>
<p>Path to dependency file: nedb/package.json</p>
<p>Path to vulnerable library: nedb/node_modules/ms</p>
<p>
Dependency Hierarchy:
- mocha-1.4.3.tgz (Root Library)
- :x: **ms-0.3.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/YJSoft/nedb/commit/c3b8d81a383e4da24efcf18e552e70fb79fe742d">c3b8d81a383e4da24efcf18e552e70fb79fe742d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS).
<p>Publish Date: 2017-04-12
<p>URL: <a href=https://github.com/zeit/ms/commit/305f2ddcd4eff7cc7c518aca6bb2b2d2daad8fef>WS-2017-0247</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>3.4</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/vercel/ms/pull/89">https://github.com/vercel/ms/pull/89</a></p>
<p>Release Date: 2017-04-12</p>
<p>Fix Resolution: 2.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2017-0247 (Low) detected in ms-0.3.0.tgz - autoclosed - ## WS-2017-0247 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ms-0.3.0.tgz</b></p></summary>
<p>Tiny ms conversion utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.3.0.tgz">https://registry.npmjs.org/ms/-/ms-0.3.0.tgz</a></p>
<p>Path to dependency file: nedb/package.json</p>
<p>Path to vulnerable library: nedb/node_modules/ms</p>
<p>
Dependency Hierarchy:
- mocha-1.4.3.tgz (Root Library)
- :x: **ms-0.3.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/YJSoft/nedb/commit/c3b8d81a383e4da24efcf18e552e70fb79fe742d">c3b8d81a383e4da24efcf18e552e70fb79fe742d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS).
<p>Publish Date: 2017-04-12
<p>URL: <a href=https://github.com/zeit/ms/commit/305f2ddcd4eff7cc7c518aca6bb2b2d2daad8fef>WS-2017-0247</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>3.4</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/vercel/ms/pull/89">https://github.com/vercel/ms/pull/89</a></p>
<p>Release Date: 2017-04-12</p>
<p>Fix Resolution: 2.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws low detected in ms tgz autoclosed ws low severity vulnerability vulnerable library ms tgz tiny ms conversion utility library home page a href path to dependency file nedb package json path to vulnerable library nedb node modules ms dependency hierarchy mocha tgz root library x ms tgz vulnerable library found in head commit a href found in base branch master vulnerability details affected versions of this package are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
8,950
| 12,059,071,918
|
IssuesEvent
|
2020-04-15 18:36:30
|
googleapis/cloud-profiler-nodejs
|
https://api.github.com/repos/googleapis/cloud-profiler-nodejs
|
closed
|
Drop support for Node 8
|
api: cloudprofiler priority: p2 type: process
|
We'd like to drop support for Node 8 before releasing v4.0.0, because we would like to reduce the number of major version changes for this module.
@bcoe and @JustinBeckwith -- At what point in time will we be able to drop Node 8?
|
1.0
|
Drop support for Node 8 - We'd like to drop support for Node 8 before releasing v4.0.0, because we would like to reduce the number of major version changes for this module.
@bcoe and @JustinBeckwith -- At what point in time will we be able to drop Node 8?
|
process
|
drop support for node we d like to drop support for node before releasing because we would like to reduce the number of major version changes for this module bcoe and justinbeckwith at what point in time will we be able to drop node
| 1
|
1,926
| 2,521,738,882
|
IssuesEvent
|
2015-01-19 16:34:05
|
Trashed/MovieSuggester
|
https://api.github.com/repos/Trashed/MovieSuggester
|
opened
|
MainActivity
|
top priority UI WIP
|
Build the main view!
This view should have:
* ActionBar
* Search Icon (launches search field), Settings (Icon), Help and About options
* Horizontal scrolling lists for top ten suggested movies, actors and genres
* These lists are aligned vertically on top of each other
|
1.0
|
MainActivity - Build the main view!
This view should have:
* ActionBar
* Search Icon (launches search field), Settings (Icon), Help and About options
* Horizontal scrolling lists for top ten suggested movies, actors and genres
* These lists are aligned vertically on top of each other
|
non_process
|
mainactivity build the main view this view should have actionbar search icon launches search field settings icon help and about options horizontal scrolling lists for top ten suggested movies actors and genres these lists are aligned vertically on top of each other
| 0
|
110,514
| 9,458,893,663
|
IssuesEvent
|
2019-04-17 07:02:30
|
Students-of-the-city-of-Kostroma/Student-timetable
|
https://api.github.com/repos/Students-of-the-city-of-Kostroma/Student-timetable
|
closed
|
Разработать сценарии тестирования пригодных для автоматизации модульным тестированием метода Delete(Model model) сущности ВУЗ
|
Controller Delete(Model model) Script Unit test ВУЗ ЛР04
|
[План разработки.](https://docs.google.com/presentation/d/1sLkafCqJTvIAcyZ1jfNw0JrZq8o33WYKGot23z7EfzA/edit#slide=id.g4e940e7976_0_102)
[Сценарий.](https://docs.google.com/spreadsheets/d/114F1wKsHoGB75gmF2p_XUR5zgbUb6IeQNX1ziO_BSIw/edit#gid=2120214548)
[Ссылка на диаграмму классов.](https://docs.google.com/presentation/d/1sLkafCqJTvIAcyZ1jfNw0JrZq8o33WYKGot23z7EfzA/edit#slide=id.g4e940e7976_0_0)
|
1.0
|
Разработать сценарии тестирования пригодных для автоматизации модульным тестированием метода Delete(Model model) сущности ВУЗ - [План разработки.](https://docs.google.com/presentation/d/1sLkafCqJTvIAcyZ1jfNw0JrZq8o33WYKGot23z7EfzA/edit#slide=id.g4e940e7976_0_102)
[Сценарий.](https://docs.google.com/spreadsheets/d/114F1wKsHoGB75gmF2p_XUR5zgbUb6IeQNX1ziO_BSIw/edit#gid=2120214548)
[Ссылка на диаграмму классов.](https://docs.google.com/presentation/d/1sLkafCqJTvIAcyZ1jfNw0JrZq8o33WYKGot23z7EfzA/edit#slide=id.g4e940e7976_0_0)
|
non_process
|
разработать сценарии тестирования пригодных для автоматизации модульным тестированием метода delete model model сущности вуз
| 0
|
464,105
| 13,306,387,839
|
IssuesEvent
|
2020-08-25 20:09:37
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
Expand E2E Tests for Facility Locator (Vet Centers)
|
Q32020-priority QA frontend stretch-goal vsa vsa-facilities
|
## Background
We need to expand the test suite specifically for Vet Center searches
## Tasks
**1. Check to see if the following user stories are covered by existing tests**
-When `Vet centers` is selected as the facility type,
<details>
<summary>When `Vet centers` is selected as the facility type (expand)</summary>
- The "Choose a service type" is unavailable- No alert displays below search box
- Results will be relevant to the geographic location and facility type specified
- Search result cards contain required standard information
- Alpha icon correlating to map pin
- Distance
- Hyperlinked name of VA Facility
- Facility operating status (if normal, no status is displayed)
- Address
- Functional Directions hyperlink
- Main Number
- Mental Health Number (if available)
</details>
<details>
<summary>When the Veteran clicks the hyperlinked facility name, a facility detail page for that facility is displayed (expand)</summary>
The Facility Detail page contains the following information about the selected facility
- Name of Facility
- Alert box with operating status information (if normal, no alert is displayed)
- Address
- Main Number
- Mental Health Number (if available)
- Functional Directions hyperlink
- List of services
- Static map
- Hours of Operation displayed by day
</details>
---
**2. Correct any existing tests if intended functionality has changed since the tests were created.**
---
**3. Write additional tests to expand coverage for existing functionality.**
---
**4. Conduct testing to ensure tests are correct.**
---
## Acceptance Criteria:
- [ ] Tests are present for searches in which "Choose a VA facility type" is set to `Vet Centers`
- [ ] Tests all pass with current functionality.
|
1.0
|
Expand E2E Tests for Facility Locator (Vet Centers) - ## Background
We need to expand the test suite specifically for Vet Center searches
## Tasks
**1. Check to see if the following user stories are covered by existing tests**
-When `Vet centers` is selected as the facility type,
<details>
<summary>When `Vet centers` is selected as the facility type (expand)</summary>
- The "Choose a service type" is unavailable- No alert displays below search box
- Results will be relevant to the geographic location and facility type specified
- Search result cards contain required standard information
- Alpha icon correlating to map pin
- Distance
- Hyperlinked name of VA Facility
- Facility operating status (if normal, no status is displayed)
- Address
- Functional Directions hyperlink
- Main Number
- Mental Health Number (if available)
</details>
<details>
<summary>When the Veteran clicks the hyperlinked facility name, a facility detail page for that facility is displayed (expand)</summary>
The Facility Detail page contains the following information about the selected facility
- Name of Facility
- Alert box with operating status information (if normal, no alert is displayed)
- Address
- Main Number
- Mental Health Number (if available)
- Functional Directions hyperlink
- List of services
- Static map
- Hours of Operation displayed by day
</details>
---
**2. Correct any existing tests if intended functionality has changed since the tests were created.**
---
**3. Write additional tests to expand coverage for existing functionality.**
---
**4. Conduct testing to ensure tests are correct.**
---
## Acceptance Criteria:
- [ ] Tests are present for searches in which "Choose a VA facility type" is set to `Vet Centers`
- [ ] Tests all pass with current functionality.
|
non_process
|
expand tests for facility locator vet centers background we need to expand the test suite specifically for vet center searches tasks check to see if the following user stories are covered by existing tests when vet centers is selected as the facility type when vet centers is selected as the facility type expand the choose a service type is unavailable no alert displays below search box results will be relevant to the geographic location and facility type specified search result cards contain required standard information alpha icon correlating to map pin distance hyperlinked name of va facility facility operating status if normal no status is displayed address functional directions hyperlink main number mental health number if available when the veteran clicks the hyperlinked facility name a facility detail page for that facility is displayed expand the facility detail page contains the following information about the selected facility name of facility alert box with operating status information if normal no alert is displayed address main number mental health number if available functional directions hyperlink list of services static map hours of operation displayed by day correct any existing tests if intended functionality has changed since the tests were created write additional tests to expand coverage for existing functionality conduct testing to ensure tests are correct acceptance criteria tests are present for searches in which choose a va facility type is set to vet centers tests all pass with current functionality
| 0
|
17,258
| 23,041,373,793
|
IssuesEvent
|
2022-07-23 07:26:58
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
test_notify_all hangs forever in sparc64
|
type-bug tests 3.9 expert-multiprocessing
|
BPO | [40186](https://bugs.python.org/issue40186)
--- | :---
Nosy | @isidentical
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = None
closed_at = None
created_at = <Date 2020-04-04.21:11:33.863>
labels = ['type-bug', 'tests', '3.9']
title = 'test_notify_all hangs forever in sparc64'
updated_at = <Date 2020-04-04.21:13:20.430>
user = 'https://github.com/isidentical'
```
bugs.python.org fields:
```python
activity = <Date 2020-04-04.21:13:20.430>
actor = 'BTaskaya'
assignee = 'none'
closed = False
closed_date = None
closer = None
components = ['Tests']
creation = <Date 2020-04-04.21:11:33.863>
creator = 'BTaskaya'
dependencies = []
files = []
hgrepos = []
issue_num = 40186
keywords = []
message_count = 2.0
messages = ['365784', '365785']
nosy_count = 1.0
nosy_names = ['BTaskaya']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = 'behavior'
url = 'https://bugs.python.org/issue40186'
versions = ['Python 3.9']
```
</p></details>
|
1.0
|
test_notify_all hangs forever in sparc64 - BPO | [40186](https://bugs.python.org/issue40186)
--- | :---
Nosy | @isidentical
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = None
closed_at = None
created_at = <Date 2020-04-04.21:11:33.863>
labels = ['type-bug', 'tests', '3.9']
title = 'test_notify_all hangs forever in sparc64'
updated_at = <Date 2020-04-04.21:13:20.430>
user = 'https://github.com/isidentical'
```
bugs.python.org fields:
```python
activity = <Date 2020-04-04.21:13:20.430>
actor = 'BTaskaya'
assignee = 'none'
closed = False
closed_date = None
closer = None
components = ['Tests']
creation = <Date 2020-04-04.21:11:33.863>
creator = 'BTaskaya'
dependencies = []
files = []
hgrepos = []
issue_num = 40186
keywords = []
message_count = 2.0
messages = ['365784', '365785']
nosy_count = 1.0
nosy_names = ['BTaskaya']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = 'behavior'
url = 'https://bugs.python.org/issue40186'
versions = ['Python 3.9']
```
</p></details>
|
process
|
test notify all hangs forever in bpo nosy isidentical note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title test notify all hangs forever in updated at user bugs python org fields python activity actor btaskaya assignee none closed false closed date none closer none components creation creator btaskaya dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type behavior url versions
| 1
|
15,593
| 11,598,159,772
|
IssuesEvent
|
2020-02-24 22:25:47
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
opened
|
ERROR: Failed to download Chromium r594312!
|
Infrastructure
|
I've seen this flicker a few times recently. Don't really want to add the `PUPPETEER_SKIP_CHROMIUM_DOWNLOAD` env var but we may have to.

> ERROR: Failed to download Chromium r594312! Set "PUPPETEER_SKIP_CHROMIUM_DOWNLOAD" env variable to skip download.
> { Error: self signed certificate in certificate chain
> at TLSSocket.onConnectSecure (_tls_wrap.js:1051:34)
> at TLSSocket.emit (events.js:189:13)
> at TLSSocket._finishInit (_tls_wrap.js:633:8) code: 'SELF_SIGNED_CERT_IN_CHAIN' }
> Script cd web && yarn install handling the va:web:install event returned with error code 1
> Script @va:web:install was called via post-install-cmd
|
1.0
|
ERROR: Failed to download Chromium r594312! - I've seen this flicker a few times recently. Don't really want to add the `PUPPETEER_SKIP_CHROMIUM_DOWNLOAD` env var but we may have to.

> ERROR: Failed to download Chromium r594312! Set "PUPPETEER_SKIP_CHROMIUM_DOWNLOAD" env variable to skip download.
> { Error: self signed certificate in certificate chain
> at TLSSocket.onConnectSecure (_tls_wrap.js:1051:34)
> at TLSSocket.emit (events.js:189:13)
> at TLSSocket._finishInit (_tls_wrap.js:633:8) code: 'SELF_SIGNED_CERT_IN_CHAIN' }
> Script cd web && yarn install handling the va:web:install event returned with error code 1
> Script @va:web:install was called via post-install-cmd
|
non_process
|
error failed to download chromium i ve seen this flicker a few times recently don t really want to add the puppeteer skip chromium download env var but we may have to error failed to download chromium set puppeteer skip chromium download env variable to skip download error self signed certificate in certificate chain at tlssocket onconnectsecure tls wrap js at tlssocket emit events js at tlssocket finishinit tls wrap js code self signed cert in chain script cd web yarn install handling the va web install event returned with error code script va web install was called via post install cmd
| 0
|
17,502
| 23,315,752,796
|
IssuesEvent
|
2022-08-08 12:22:46
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Primary key in model using a missing column
|
bug/0-unknown kind/bug process/candidate team/schema topic: prisma db pull topic: Your friendly prisma developers
|
Hi Prisma Team! Prisma Migrate just crashed.
## Command
`db pull`
## Versions
| Name | Version |
|-------------|--------------------|
| Platform | darwin-arm64 |
| Node | v16.13.1 |
| Prisma CLI | 4.1.0 |
| Engine | 8d8414deb360336e4698a65aa45a1fbaf1ce13d8|
## Error
```
Error: [libs/datamodel/connectors/dml/src/model.rs:494:29] Hi there! We've been seeing this error in our error reporting backend, but cannot reproduce it in our own tests. The problem is that we have a primary key in the model `CUSTOMER_BRAND_LINK` that uses the column `DATE_UPDATED` which we for some reason don't have in our internal representation. If you see this, could you please file an issue to https://github.com/prisma/prisma so we can discuss about fixing this. -- Your friendly prisma developers.
```
|
1.0
|
Primary key in model using a missing column - Hi Prisma Team! Prisma Migrate just crashed.
## Command
`db pull`
## Versions
| Name | Version |
|-------------|--------------------|
| Platform | darwin-arm64 |
| Node | v16.13.1 |
| Prisma CLI | 4.1.0 |
| Engine | 8d8414deb360336e4698a65aa45a1fbaf1ce13d8|
## Error
```
Error: [libs/datamodel/connectors/dml/src/model.rs:494:29] Hi there! We've been seeing this error in our error reporting backend, but cannot reproduce it in our own tests. The problem is that we have a primary key in the model `CUSTOMER_BRAND_LINK` that uses the column `DATE_UPDATED` which we for some reason don't have in our internal representation. If you see this, could you please file an issue to https://github.com/prisma/prisma so we can discuss about fixing this. -- Your friendly prisma developers.
```
|
process
|
primary key in model using a missing column hi prisma team prisma migrate just crashed command db pull versions name version platform darwin node prisma cli engine error error hi there we ve been seeing this error in our error reporting backend but cannot reproduce it in our own tests the problem is that we have a primary key in the model customer brand link that uses the column date updated which we for some reason don t have in our internal representation if you see this could you please file an issue to so we can discuss about fixing this your friendly prisma developers
| 1
|
44,381
| 12,124,137,234
|
IssuesEvent
|
2020-04-22 13:46:44
|
carbon-design-system/ibm-security
|
https://api.github.com/repos/carbon-design-system/ibm-security
|
closed
|
Button loading prop should not disable button
|
Defect severity 4
|
## Request
The `loading` prop on the `Button` component disables the button. I don't believe that this should be the default behavior. Let the consumer decide whether to disable the button or not. If they already pass in a prop to set `loading`, they can use the same one to disable if they want.
Not sure if it is a bug or desired behavior, but when setting `loading` to true, the button changes to a ghost button, regardless of initial type.
|
1.0
|
Button loading prop should not disable button - ## Request
The `loading` prop on the `Button` component disables the button. I don't believe that this should be the default behavior. Let the consumer decide whether to disable the button or not. If they already pass in a prop to set `loading`, they can use the same one to disable if they want.
Not sure if it is a bug or desired behavior, but when setting `loading` to true, the button changes to a ghost button, regardless of initial type.
|
non_process
|
button loading prop should not disable button request the loading prop on the button component disables the button i don t believe that this should be the default behavior let the consumer decide whether to disable the button or not if they already pass in a prop to set loading they can use the same one to disable if they want not sure if it is a bug or desired behavior but when setting loading to true the button changes to a ghost button regardless of initial type
| 0
|
2,870
| 5,830,076,148
|
IssuesEvent
|
2017-05-08 15:56:45
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
reopened
|
editable function returning false is not honoured - sort on custom formatter does not work
|
enhancement inprocess
|
First of all thanks for a great component :thumbsup: - am using it extensively.
Couple of issues that i have noticed
1. I use dataFormat that returns a React Component. The rendering is fine but the sort does not work at all. If this is going to be fixed it will be good if there will be an option whether to sort on the unformatted or formatted date (similar to filterFormatted). **(In my example, sort on "Name' column to see the issue)**
2. I use editable as a function to determine is a cell can be edited or not. When the editor is a cutom editor no disabling happens - we can still edit the cell. **(In my example, click on the 'Product' cell to see the issue)**
3) When we have a custom editor, we are forced to have a focus function. Ideally this should be optional
4) When a cell is made non-editable via the editable property (via a function) a click on the cell changes the cell and shows it as disabled - ideally it should not change at all. **In my example click on the id cell and click on the 'Name' cell to see the difference - both are non-editable but what is visible changes)** i.e. functionally there is no issue but it looks odd.
The code below can be used to highlight these issues
```javascript
class SomeEditor extends React.Component {
focus(){
/* PROBLEM 3 : For a custom editor, I have to have a 'focus' function - IDEALLY THIS should be optional*/
}
render() {
return (
<input autoFocus type="text" />
)
}
}
const SomeFormatter = ({value}) => {
/* Problem 1- where there is a custom formatter - sort does not seem to work */
let valToShow = value.d == 1 ? 'ONE' : 'TWO';
return (
<div> {valToShow} </div>
);
};
function bsFormatter(cell, formatter) {
let Formatter = formatter;
return (
<Formatter value={cell} />
);
}
var data = [
{id : 1, date : {d : 1}, product : 'P1'},
{id : 2, date : {d : 2}, product : 'P2'},
];
const someEditor = (onUpdate, props) => (<SomeEditor onUpdate={ onUpdate } {...props}/>);
function editable(){
/* Problem 2: Even though false is always returned the custom editor is not disabled and we can edit */
console.log("Returning FALSE always");
return false;
}
class ProblemTable extends React.Component {
render() {
const cellEditProp = {
mode: 'click',
};
let customEditor = {getElement : someEditor};
let formatter = (cell, row) => bsFormatter(cell, SomeFormatter);
return (
<BootstrapTable cellEdit={ cellEditProp } filterFormatted striped hover data={ data }>
<TableHeaderColumn dataField='id' isKey>Product ID</TableHeaderColumn>
<TableHeaderColumn editable={editable} dataSort={true} dataFormat={formatter} dataField='date'> Name</TableHeaderColumn>
<TableHeaderColumn editable={editable} customEditor={ customEditor } dataSort={true} dataField='product'>Product</TableHeaderColumn>
</BootstrapTable>
);
}
}
|
1.0
|
editable function returning false is not honoured - sort on custom formatter does not work - First of all thanks for a great component :thumbsup: - am using it extensively.
Couple of issues that i have noticed
1. I use dataFormat that returns a React Component. The rendering is fine but the sort does not work at all. If this is going to be fixed it will be good if there will be an option whether to sort on the unformatted or formatted date (similar to filterFormatted). **(In my example, sort on "Name' column to see the issue)**
2. I use editable as a function to determine is a cell can be edited or not. When the editor is a cutom editor no disabling happens - we can still edit the cell. **(In my example, click on the 'Product' cell to see the issue)**
3) When we have a custom editor, we are forced to have a focus function. Ideally this should be optional
4) When a cell is made non-editable via the editable property (via a function) a click on the cell changes the cell and shows it as disabled - ideally it should not change at all. **In my example click on the id cell and click on the 'Name' cell to see the difference - both are non-editable but what is visible changes)** i.e. functionally there is no issue but it looks odd.
The code below can be used to highlight these issues
```javascript
class SomeEditor extends React.Component {
focus(){
/* PROBLEM 3 : For a custom editor, I have to have a 'focus' function - IDEALLY THIS should be optional*/
}
render() {
return (
<input autoFocus type="text" />
)
}
}
const SomeFormatter = ({value}) => {
/* Problem 1- where there is a custom formatter - sort does not seem to work */
let valToShow = value.d == 1 ? 'ONE' : 'TWO';
return (
<div> {valToShow} </div>
);
};
function bsFormatter(cell, formatter) {
let Formatter = formatter;
return (
<Formatter value={cell} />
);
}
var data = [
{id : 1, date : {d : 1}, product : 'P1'},
{id : 2, date : {d : 2}, product : 'P2'},
];
const someEditor = (onUpdate, props) => (<SomeEditor onUpdate={ onUpdate } {...props}/>);
function editable(){
/* Problem 2: Even though false is always returned the custom editor is not disabled and we can edit */
console.log("Returning FALSE always");
return false;
}
class ProblemTable extends React.Component {
render() {
const cellEditProp = {
mode: 'click',
};
let customEditor = {getElement : someEditor};
let formatter = (cell, row) => bsFormatter(cell, SomeFormatter);
return (
<BootstrapTable cellEdit={ cellEditProp } filterFormatted striped hover data={ data }>
<TableHeaderColumn dataField='id' isKey>Product ID</TableHeaderColumn>
<TableHeaderColumn editable={editable} dataSort={true} dataFormat={formatter} dataField='date'> Name</TableHeaderColumn>
<TableHeaderColumn editable={editable} customEditor={ customEditor } dataSort={true} dataField='product'>Product</TableHeaderColumn>
</BootstrapTable>
);
}
}
|
process
|
editable function returning false is not honoured sort on custom formatter does not work first of all thanks for a great component thumbsup am using it extensively couple of issues that i have noticed i use dataformat that returns a react component the rendering is fine but the sort does not work at all if this is going to be fixed it will be good if there will be an option whether to sort on the unformatted or formatted date similar to filterformatted in my example sort on name column to see the issue i use editable as a function to determine is a cell can be edited or not when the editor is a cutom editor no disabling happens we can still edit the cell in my example click on the product cell to see the issue when we have a custom editor we are forced to have a focus function ideally this should be optional when a cell is made non editable via the editable property via a function a click on the cell changes the cell and shows it as disabled ideally it should not change at all in my example click on the id cell and click on the name cell to see the difference both are non editable but what is visible changes i e functionally there is no issue but it looks odd the code below can be used to highlight these issues javascript class someeditor extends react component focus problem for a custom editor i have to have a focus function ideally this should be optional render return const someformatter value problem where there is a custom formatter sort does not seem to work let valtoshow value d one two return valtoshow function bsformatter cell formatter let formatter formatter return var data id date d product id date d product const someeditor onupdate props function editable problem even though false is always returned the custom editor is not disabled and we can edit console log returning false always return false class problemtable extends react component render const celleditprop mode click let customeditor getelement someeditor let formatter cell row bsformatter cell someformatter return product id name product
| 1
|
6,813
| 9,956,647,326
|
IssuesEvent
|
2019-07-05 14:28:06
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
Issue with custom database path
|
log-processing
|
I'm trying to keep incremental logs in the btree DB. My initial command was:
```zcat /usr/local/www/usesthis.com/log/access.log.*.gz | goaccess --keep-db-files --db-path=/usr/local/share/stats/usesthis/ --log-format=COMBINED -o /usr/local/www/usesthis.com/public/stats.html -```
Which worked great, I can see a ton of data - I want to start from a solid base of old logs, etc. I want to set up a cronjob that appends the logs from `access.log`/`access.log.1` on a daily basis, so I figured I'm meant to do something like this:
```goaccess --keep-db-files --db-path=/usr/local/shared/stats/usesthis/ /usr/local/www/usesthis.com/log/access.log /usr/local/www/usesthis.com/log/access.log.1 --log-format=COMBINED -o /usr/local/www/usesthis.com/public/stats.html```
However, I get:
```
GoAccess - version 1.2 - Jun 19 2017 12:55:29
Config file: No config file used
Fatal error has occurred
Error occured at: src/tcbtdb.c - tc_db_set_path - 66
Unable to access database path: No such file or directory
```
A) Am I understanding how to even do the incremental data stuff correctly?
B) Why can't it access the database path? Not using a config file, so what you see is all there. I can see all the .tcb files in `/usr/local/share/stats/usesthis`, and trying with or without a trailing slash doesn't seem to make a difference.
What am I doing wrong?
|
1.0
|
Issue with custom database path - I'm trying to keep incremental logs in the btree DB. My initial command was:
```zcat /usr/local/www/usesthis.com/log/access.log.*.gz | goaccess --keep-db-files --db-path=/usr/local/share/stats/usesthis/ --log-format=COMBINED -o /usr/local/www/usesthis.com/public/stats.html -```
Which worked great, I can see a ton of data - I want to start from a solid base of old logs, etc. I want to set up a cronjob that appends the logs from `access.log`/`access.log.1` on a daily basis, so I figured I'm meant to do something like this:
```goaccess --keep-db-files --db-path=/usr/local/shared/stats/usesthis/ /usr/local/www/usesthis.com/log/access.log /usr/local/www/usesthis.com/log/access.log.1 --log-format=COMBINED -o /usr/local/www/usesthis.com/public/stats.html```
However, I get:
```
GoAccess - version 1.2 - Jun 19 2017 12:55:29
Config file: No config file used
Fatal error has occurred
Error occured at: src/tcbtdb.c - tc_db_set_path - 66
Unable to access database path: No such file or directory
```
A) Am I understanding how to even do the incremental data stuff correctly?
B) Why can't it access the database path? Not using a config file, so what you see is all there. I can see all the .tcb files in `/usr/local/share/stats/usesthis`, and trying with or without a trailing slash doesn't seem to make a difference.
What am I doing wrong?
|
process
|
issue with custom database path i m trying to keep incremental logs in the btree db my initial command was zcat usr local www usesthis com log access log gz goaccess keep db files db path usr local share stats usesthis log format combined o usr local www usesthis com public stats html which worked great i can see a ton of data i want to start from a solid base of old logs etc i want to set up a cronjob that appends the logs from access log access log on a daily basis so i figured i m meant to do something like this goaccess keep db files db path usr local shared stats usesthis usr local www usesthis com log access log usr local www usesthis com log access log log format combined o usr local www usesthis com public stats html however i get goaccess version jun config file no config file used fatal error has occurred error occured at src tcbtdb c tc db set path unable to access database path no such file or directory a am i understanding how to even do the incremental data stuff correctly b why can t it access the database path not using a config file so what you see is all there i can see all the tcb files in usr local share stats usesthis and trying with or without a trailing slash doesn t seem to make a difference what am i doing wrong
| 1
|
147,298
| 23,196,248,935
|
IssuesEvent
|
2022-08-01 16:39:15
|
microsoft/fluentui
|
https://api.github.com/repos/microsoft/fluentui
|
closed
|
[Bug]: SplitButton opens the dropdown menu instead of invoking the onClick action when used with touch.
|
Resolution: By Design Fluent UI react (v8) Component: SplitButton
|
### Library
React / v8 (@fluentui/react)
### System Info
```shell
System:
OS: Windows 10 10.0.22000
CPU: (8) x64 Intel(R) Core(TM) i7-1065G7 CPU @ 1.30GHz
Memory: 2.07 GB / 15.60 GB
Browsers:
Edge: Spartan (44.22000.120.0), Chromium (102.0.1245.44), ChromiumDev (104.0.1293.1)
Internet Explorer: 11.0.22000.120
```
### Are you reporting Accessibility issue?
no
### Reproduction
https://developer.microsoft.com/en-us/fluentui#/controls/web/button
### Bug Description
## Actual Behavior
When the SplitButton is used with touch, the dropdown menu appears even if user touches on the left portion of the split button. This is reproducible on https://developer.microsoft.com/en-us/fluentui#/controls/web/button sample controls for the SplitButton component as well.
## Expected Behavior
Behavior with touch should be same as behavior with mouse click. i.e. when user touches the left portion of the split button, it should invoke the onClick function rather than opening the dropdown.
### Logs
_No response_
### Requested priority
Normal
### Products/sites affected
Outlook Web
### Are you willing to submit a PR to fix?
no
### Validations
- [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate.
- [X] The provided reproduction is a minimal reproducible example of the bug.
|
1.0
|
[Bug]: SplitButton opens the dropdown menu instead of invoking the onClick action when used with touch. - ### Library
React / v8 (@fluentui/react)
### System Info
```shell
System:
OS: Windows 10 10.0.22000
CPU: (8) x64 Intel(R) Core(TM) i7-1065G7 CPU @ 1.30GHz
Memory: 2.07 GB / 15.60 GB
Browsers:
Edge: Spartan (44.22000.120.0), Chromium (102.0.1245.44), ChromiumDev (104.0.1293.1)
Internet Explorer: 11.0.22000.120
```
### Are you reporting Accessibility issue?
no
### Reproduction
https://developer.microsoft.com/en-us/fluentui#/controls/web/button
### Bug Description
## Actual Behavior
When the SplitButton is used with touch, the dropdown menu appears even if user touches on the left portion of the split button. This is reproducible on https://developer.microsoft.com/en-us/fluentui#/controls/web/button sample controls for the SplitButton component as well.
## Expected Behavior
Behavior with touch should be same as behavior with mouse click. i.e. when user touches the left portion of the split button, it should invoke the onClick function rather than opening the dropdown.
### Logs
_No response_
### Requested priority
Normal
### Products/sites affected
Outlook Web
### Are you willing to submit a PR to fix?
no
### Validations
- [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate.
- [X] The provided reproduction is a minimal reproducible example of the bug.
|
non_process
|
splitbutton opens the dropdown menu instead of invoking the onclick action when used with touch library react fluentui react system info shell system os windows cpu intel r core tm cpu memory gb gb browsers edge spartan chromium chromiumdev internet explorer are you reporting accessibility issue no reproduction bug description actual behavior when the splitbutton is used with touch the dropdown menu appears even if user touches on the left portion of the split button this is reproducible on sample controls for the splitbutton component as well expected behavior behavior with touch should be same as behavior with mouse click i e when user touches the left portion of the split button it should invoke the onclick function rather than opening the dropdown logs no response requested priority normal products sites affected outlook web are you willing to submit a pr to fix no validations check that there isn t already an issue that reports the same bug to avoid creating a duplicate the provided reproduction is a minimal reproducible example of the bug
| 0
|
420,406
| 28,263,462,804
|
IssuesEvent
|
2023-04-07 03:03:33
|
AY2223S2-CS2113-T13-1/tp
|
https://api.github.com/repos/AY2223S2-CS2113-T13-1/tp
|
closed
|
[PE-D][Tester B] Lack of example to show command with optional description
|
documentation priority.Medium
|

Would be good if there was an example on how to use `add` with the optional parameter description
<!--session: 1680252471869-a4ff2ce6-21c8-4fa9-ad69-6bc92f657b80-->
<!--Version: Web v3.4.7-->
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: jaredoong/ped#8
|
1.0
|
[PE-D][Tester B] Lack of example to show command with optional description - 
Would be good if there was an example on how to use `add` with the optional parameter description
<!--session: 1680252471869-a4ff2ce6-21c8-4fa9-ad69-6bc92f657b80-->
<!--Version: Web v3.4.7-->
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: jaredoong/ped#8
|
non_process
|
lack of example to show command with optional description would be good if there was an example on how to use add with the optional parameter description labels severity low type documentationbug original jaredoong ped
| 0
|
365,623
| 10,790,040,692
|
IssuesEvent
|
2019-11-05 13:19:37
|
AY1920S1-CS2113T-F09-4/main
|
https://api.github.com/repos/AY1920S1-CS2113T-F09-4/main
|
closed
|
Documentation not uniform in examples, making it hard to follow.
|
priority.High severity.Medium status.Ongoing
|

Difference in example, when executed, I won't find jingle inside. Confusing for user.
<hr><sub>[original: JasonLeeWeiHern/ped#7]<br/>
</sub>
|
1.0
|
Documentation not uniform in examples, making it hard to follow. - 
Difference in example, when executed, I won't find jingle inside. Confusing for user.
<hr><sub>[original: JasonLeeWeiHern/ped#7]<br/>
</sub>
|
non_process
|
documentation not uniform in examples making it hard to follow difference in example when executed i won t find jingle inside confusing for user
| 0
|
192,570
| 15,353,232,881
|
IssuesEvent
|
2021-03-01 08:17:37
|
kbuffington/Georgia
|
https://api.github.com/repos/kbuffington/Georgia
|
closed
|
Suggestion for Instructions
|
documentation
|
I'm super casual on this kind of thing (Not even coding, just installing)
So, the roadblock I ran into was on step 8. I didn't understand how to add the Jscript Pannel, specifically as I'm ultra new to Foobar.
Luckily your video helped me out there, through some reddit skimming, and happening to run across it on youtube.
It may seem super trivial, but a little bit of a more direct explanation on adding a Jscript Pannel more specifically targeted at someone who had just downloaded Fubar for the first time (Like myself) could prove helpful in progressing through the install process.
Simple suggestion would be View>Layout>New Scratchbox, ect.. added to the instruction flow.
Awesome configuration, and beautiful design. Keep up the amazing work!
|
1.0
|
Suggestion for Instructions - I'm super casual on this kind of thing (Not even coding, just installing)
So, the roadblock I ran into was on step 8. I didn't understand how to add the Jscript Pannel, specifically as I'm ultra new to Foobar.
Luckily your video helped me out there, through some reddit skimming, and happening to run across it on youtube.
It may seem super trivial, but a little bit of a more direct explanation on adding a Jscript Pannel more specifically targeted at someone who had just downloaded Fubar for the first time (Like myself) could prove helpful in progressing through the install process.
Simple suggestion would be View>Layout>New Scratchbox, ect.. added to the instruction flow.
Awesome configuration, and beautiful design. Keep up the amazing work!
|
non_process
|
suggestion for instructions i m super casual on this kind of thing not even coding just installing so the roadblock i ran into was on step i didn t understand how to add the jscript pannel specifically as i m ultra new to foobar luckily your video helped me out there through some reddit skimming and happening to run across it on youtube it may seem super trivial but a little bit of a more direct explanation on adding a jscript pannel more specifically targeted at someone who had just downloaded fubar for the first time like myself could prove helpful in progressing through the install process simple suggestion would be view layout new scratchbox ect added to the instruction flow awesome configuration and beautiful design keep up the amazing work
| 0
|
6,506
| 9,592,746,130
|
IssuesEvent
|
2019-05-09 09:40:56
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
opened
|
Obsolete 'GO:0046731 passive induction of host immune response by virus' and children?
|
multi-species process
|
Hello,
GO:0046731 passive induction of host immune response by virus
is defined as "The unintentional stimulation by a virus of a host defense response to viral infection, as part of the viral infectious cycle." - so that doesn't sound like any gene product is actually doing anything.
Similar for the children:
- GO:0046734 passive induction of host cell-mediated immune response by virus
- GO:0046733 passive induction of host humoral immune response by virus
- GO:0046735 passive induction of host innate immune response by virus
There is a single annotation to this term by SWIE (@milarolo ) to a human protein (which is anyway inconsistent with the term's meaning.
I propose to obsolete that branch, or merge it into GO:0046730 induction of host immune response by virus
@mgiglio99 @pmasson55 @ValWood
What do you think ?
Thanks, Pascale
|
1.0
|
Obsolete 'GO:0046731 passive induction of host immune response by virus' and children? - Hello,
GO:0046731 passive induction of host immune response by virus
is defined as "The unintentional stimulation by a virus of a host defense response to viral infection, as part of the viral infectious cycle." - so that doesn't sound like any gene product is actually doing anything.
Similar for the children:
- GO:0046734 passive induction of host cell-mediated immune response by virus
- GO:0046733 passive induction of host humoral immune response by virus
- GO:0046735 passive induction of host innate immune response by virus
There is a single annotation to this term by SWIE (@milarolo ) to a human protein (which is anyway inconsistent with the term's meaning.
I propose to obsolete that branch, or merge it into GO:0046730 induction of host immune response by virus
@mgiglio99 @pmasson55 @ValWood
What do you think ?
Thanks, Pascale
|
process
|
obsolete go passive induction of host immune response by virus and children hello go passive induction of host immune response by virus is defined as the unintentional stimulation by a virus of a host defense response to viral infection as part of the viral infectious cycle so that doesn t sound like any gene product is actually doing anything similar for the children go passive induction of host cell mediated immune response by virus go passive induction of host humoral immune response by virus go passive induction of host innate immune response by virus there is a single annotation to this term by swie milarolo to a human protein which is anyway inconsistent with the term s meaning i propose to obsolete that branch or merge it into go induction of host immune response by virus valwood what do you think thanks pascale
| 1
|
4,311
| 7,203,155,640
|
IssuesEvent
|
2018-02-06 08:06:08
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[processing] [FEATURE] SplitWithLines
|
Automatic new feature Processing
|
Original commit: https://github.com/qgis/QGIS/commit/0e2ef065d7e89ba65db93f3a628d46c9bb31f265 by nyalldawson
Rename algorithm SplitLinesWithLines to SplitWithLines
Accept polygon as input, too
Use only selected lines to split with (if processing is set to use selection only)
Issue log message if trying to split multi geometries
Update help
|
1.0
|
[processing] [FEATURE] SplitWithLines - Original commit: https://github.com/qgis/QGIS/commit/0e2ef065d7e89ba65db93f3a628d46c9bb31f265 by nyalldawson
Rename algorithm SplitLinesWithLines to SplitWithLines
Accept polygon as input, too
Use only selected lines to split with (if processing is set to use selection only)
Issue log message if trying to split multi geometries
Update help
|
process
|
splitwithlines original commit by nyalldawson rename algorithm splitlineswithlines to splitwithlines accept polygon as input too use only selected lines to split with if processing is set to use selection only issue log message if trying to split multi geometries update help
| 1
|
8,730
| 11,863,341,717
|
IssuesEvent
|
2020-03-25 19:32:27
|
w3c/webauthn
|
https://api.github.com/repos/w3c/webauthn
|
closed
|
Add direction to somewhere to ask for help to the contributing guidelines
|
type:process
|
I have been looking around this repo and finding answers to most of the questions I have in various issues and PR's.
Does such a community exist which will accept questions for people trying to use webauthn?
It would be great if there was some line in the contributing.md which said
"If you are unsure if your issue belongs here please ask in X"
If there is no suitable community to put in as X, it would still be helpful to point to an FAQ page or even tag on stackoverflow
|
1.0
|
Add direction to somewhere to ask for help to the contributing guidelines - I have been looking around this repo and finding answers to most of the questions I have in various issues and PR's.
Does such a community exist which will accept questions for people trying to use webauthn?
It would be great if there was some line in the contributing.md which said
"If you are unsure if your issue belongs here please ask in X"
If there is no suitable community to put in as X, it would still be helpful to point to an FAQ page or even tag on stackoverflow
|
process
|
add direction to somewhere to ask for help to the contributing guidelines i have been looking around this repo and finding answers to most of the questions i have in various issues and pr s does such a community exist which will accept questions for people trying to use webauthn it would be great if there was some line in the contributing md which said if you are unsure if your issue belongs here please ask in x if there is no suitable community to put in as x it would still be helpful to point to an faq page or even tag on stackoverflow
| 1
|
7,417
| 10,541,888,562
|
IssuesEvent
|
2019-10-02 11:57:34
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
closed
|
Run the breaking change detector on PRs
|
type: process
|
For APIs which have already gone GA, it would be useful to run our breaking change detector against the published version.
Will need to look to see how tricky that is.
|
1.0
|
Run the breaking change detector on PRs - For APIs which have already gone GA, it would be useful to run our breaking change detector against the published version.
Will need to look to see how tricky that is.
|
process
|
run the breaking change detector on prs for apis which have already gone ga it would be useful to run our breaking change detector against the published version will need to look to see how tricky that is
| 1
|
86,963
| 8,055,089,918
|
IssuesEvent
|
2018-08-02 08:12:09
|
pandas-dev/pandas
|
https://api.github.com/repos/pandas-dev/pandas
|
closed
|
Specific Timestamps breaks time series indexing (.loc returns wrong results)
|
Indexing Testing Timezones good first issue
|
When try to access labels (`.loc`) by using a specific list of `Timestamp` objects or a `DatetimeIndex` object (See attached csv file), the resulting index is returned in UTC offset but the original timezone is not removed.
This seems to happen only in very specific cases, when the index passed to `.loc` contains labels that do not exist in the DataFrame and also contains duplicates.
@yuval-jether
```python
import pandas as pd
import pytz
idx = pd.date_range('2011-01-01', '2017-10-01 00:00:00', freq='h', tz='America/Chicago')
s = pd.Series(np.random.rand(len(idx)), index=idx)
i = pd.to_datetime(['2017-09-30 06:00:00', '2017-09-30 23:00:00', '2017-09-30 23:00:00', '2017-10-01 00:00:00',]).tz_localize('America/Chicago')
s.loc[i]
Out[39]:
2017-09-30 06:00:00-05:00 0.380138
2017-09-30 23:00:00-05:00 0.774696
2017-09-30 23:00:00-05:00 0.774696
2017-10-01 00:00:00-05:00 0.728027
dtype: float64
# Added a label that does not exist in the Series
import pandas as pd
import pytz
idx = pd.date_range('2011-01-01', '2017-10-01 00:00:00', freq='h', tz='America/Chicago')
s = pd.Series(np.random.rand(len(idx)), index=idx)
i = pd.to_datetime(['2017-09-30 06:00:00', '2017-09-30 23:00:00', '2017-09-30 23:00:00', '2017-10-01 00:00:00', '2017-10-01 01:00:00',]).tz_localize('America/Chicago')
s.loc[i]
Out[40]:
2017-09-30 11:00:00-05:00 0.645350
2017-10-01 04:00:00-05:00 0.099323
2017-10-01 04:00:00-05:00 0.099323
2017-10-01 05:00:00-05:00 0.037136
2017-10-01 06:00:00-05:00 NaN
dtype: float64
```
#### Output of ``pd.show_versions()``
<details>
pd.show_versions()
2017-10-30 03:15:08 [pip.utils] [DEBUG] lzma module is not available
2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: git
2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: hg
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_DEBUG' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'WITH_PYMALLOC' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_UNICODE_SIZE' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_DEBUG' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'WITH_PYMALLOC' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_UNICODE_SIZE' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: svn
2017-10-30 03:15:09 [pip.vcs] [DEBUG] Registered VCS backend: bzr
INSTALLED VERSIONS
------------------
commit: None
python: 2.7.12.final.0
python-bits: 64
OS: Windows
OS-release: 8.1
machine: AMD64
processor: Intel64 Family 6 Model 60 Stepping 3, GenuineIntel
byteorder: little
LC_ALL: None
LANG: None
LOCALE: None.None
pandas: 0.20.3
pytest: 3.0.7
pip: 9.0.1
setuptools: 36.5.0
Cython: None
numpy: 1.13.1
scipy: 0.18.1
xarray: 0.9.6
IPython: 5.5.0
sphinx: None
patsy: 0.4.1
dateutil: 2.6.1
pytz: 2017.2
blosc: None
bottleneck: None
tables: None
numexpr: 2.6.2
feather: None
matplotlib: 2.0.2
openpyxl: 2.4.8
xlrd: 1.0.0
xlwt: 1.2.0
xlsxwriter: None
lxml: 3.8.0
bs4: 4.5.1
html5lib: 0.999999999
sqlalchemy: None
pymysql: None
psycopg2: None
jinja2: 2.9.6
s3fs: None
pandas_gbq: None
pandas_datareader: 0.4.0
</details>
|
1.0
|
Specific Timestamps breaks time series indexing (.loc returns wrong results) - When try to access labels (`.loc`) by using a specific list of `Timestamp` objects or a `DatetimeIndex` object (See attached csv file), the resulting index is returned in UTC offset but the original timezone is not removed.
This seems to happen only in very specific cases, when the index passed to `.loc` contains labels that do not exist in the DataFrame and also contains duplicates.
@yuval-jether
```python
import pandas as pd
import pytz
idx = pd.date_range('2011-01-01', '2017-10-01 00:00:00', freq='h', tz='America/Chicago')
s = pd.Series(np.random.rand(len(idx)), index=idx)
i = pd.to_datetime(['2017-09-30 06:00:00', '2017-09-30 23:00:00', '2017-09-30 23:00:00', '2017-10-01 00:00:00',]).tz_localize('America/Chicago')
s.loc[i]
Out[39]:
2017-09-30 06:00:00-05:00 0.380138
2017-09-30 23:00:00-05:00 0.774696
2017-09-30 23:00:00-05:00 0.774696
2017-10-01 00:00:00-05:00 0.728027
dtype: float64
# Added a label that does not exist in the Series
import pandas as pd
import pytz
idx = pd.date_range('2011-01-01', '2017-10-01 00:00:00', freq='h', tz='America/Chicago')
s = pd.Series(np.random.rand(len(idx)), index=idx)
i = pd.to_datetime(['2017-09-30 06:00:00', '2017-09-30 23:00:00', '2017-09-30 23:00:00', '2017-10-01 00:00:00', '2017-10-01 01:00:00',]).tz_localize('America/Chicago')
s.loc[i]
Out[40]:
2017-09-30 11:00:00-05:00 0.645350
2017-10-01 04:00:00-05:00 0.099323
2017-10-01 04:00:00-05:00 0.099323
2017-10-01 05:00:00-05:00 0.037136
2017-10-01 06:00:00-05:00 NaN
dtype: float64
```
#### Output of ``pd.show_versions()``
<details>
pd.show_versions()
2017-10-30 03:15:08 [pip.utils] [DEBUG] lzma module is not available
2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: git
2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: hg
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_DEBUG' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'WITH_PYMALLOC' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_UNICODE_SIZE' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_DEBUG' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'WITH_PYMALLOC' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.pep425tags] [DEBUG] Config variable 'Py_UNICODE_SIZE' is unset, Python ABI tag may be incorrect
2017-10-30 03:15:08 [pip.vcs] [DEBUG] Registered VCS backend: svn
2017-10-30 03:15:09 [pip.vcs] [DEBUG] Registered VCS backend: bzr
INSTALLED VERSIONS
------------------
commit: None
python: 2.7.12.final.0
python-bits: 64
OS: Windows
OS-release: 8.1
machine: AMD64
processor: Intel64 Family 6 Model 60 Stepping 3, GenuineIntel
byteorder: little
LC_ALL: None
LANG: None
LOCALE: None.None
pandas: 0.20.3
pytest: 3.0.7
pip: 9.0.1
setuptools: 36.5.0
Cython: None
numpy: 1.13.1
scipy: 0.18.1
xarray: 0.9.6
IPython: 5.5.0
sphinx: None
patsy: 0.4.1
dateutil: 2.6.1
pytz: 2017.2
blosc: None
bottleneck: None
tables: None
numexpr: 2.6.2
feather: None
matplotlib: 2.0.2
openpyxl: 2.4.8
xlrd: 1.0.0
xlwt: 1.2.0
xlsxwriter: None
lxml: 3.8.0
bs4: 4.5.1
html5lib: 0.999999999
sqlalchemy: None
pymysql: None
psycopg2: None
jinja2: 2.9.6
s3fs: None
pandas_gbq: None
pandas_datareader: 0.4.0
</details>
|
non_process
|
specific timestamps breaks time series indexing loc returns wrong results when try to access labels loc by using a specific list of timestamp objects or a datetimeindex object see attached csv file the resulting index is returned in utc offset but the original timezone is not removed this seems to happen only in very specific cases when the index passed to loc contains labels that do not exist in the dataframe and also contains duplicates yuval jether python import pandas as pd import pytz idx pd date range freq h tz america chicago s pd series np random rand len idx index idx i pd to datetime tz localize america chicago s loc out dtype added a label that does not exist in the series import pandas as pd import pytz idx pd date range freq h tz america chicago s pd series np random rand len idx index idx i pd to datetime tz localize america chicago s loc out nan dtype output of pd show versions pd show versions lzma module is not available registered vcs backend git registered vcs backend hg config variable py debug is unset python abi tag may be incorrect config variable with pymalloc is unset python abi tag may be incorrect config variable py unicode size is unset python abi tag may be incorrect config variable py debug is unset python abi tag may be incorrect config variable with pymalloc is unset python abi tag may be incorrect config variable py unicode size is unset python abi tag may be incorrect registered vcs backend svn registered vcs backend bzr installed versions commit none python final python bits os windows os release machine processor family model stepping genuineintel byteorder little lc all none lang none locale none none pandas pytest pip setuptools cython none numpy scipy xarray ipython sphinx none patsy dateutil pytz blosc none bottleneck none tables none numexpr feather none matplotlib openpyxl xlrd xlwt xlsxwriter none lxml sqlalchemy none pymysql none none none pandas gbq none pandas datareader
| 0
|
400,104
| 11,769,162,500
|
IssuesEvent
|
2020-03-15 13:45:07
|
microsoft/terraform-provider-azuredevops
|
https://api.github.com/repos/microsoft/terraform-provider-azuredevops
|
closed
|
Implement a Terraform data source to lookup Git repositories
|
good first issue priority-low
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
The provider currently only provides a way to provision Azure Git Repos, but does not allow us to reference a pre-existing repository. Having a data source for Git Repos will allow people to leverage other resources, like build definitions and repository permissions, without needing to create a new repository.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* data.azuredevops_git_repository
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azuredevops_project" "main_project" {
project_name = "Test Project"
description = "Test Project Description"
visibility = "private"
enable_tfvc = False
work_item_template = "Agile"
}
data azuredevops_git_repository "lookup-repo" {
project_id = azuredevops_project.main_project.id
name = "Test Lookup Repo"
}
```
### Acceptance Criteria
- [ ] A data source exists that allows for query of an Azure Git Repos based on the following fields:
- [ ] `project_id` - ID of project in which the repository exists
- [ ] `name` - name of repository
- [ ] The `computed` fields of the resource should match the resource implementation of [Azure Git Repos](https://github.com/microsoft/terraform-provider-azuredevops/blob/master/azuredevops/resource_azure_git_repository.go). There may be an opportunity to re-use the whole resource schema!
- [ ] Data source is validated with unit tests
- [ ] Data source is validated with acceptance tests
- [ ] Documentation is added in the client-facing docs
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
* #84
* [Repositories - List](https://docs.microsoft.com/en-us/rest/api/azure/devops/git/repositories/list?view=azure-devops-rest-5.1)
|
1.0
|
Implement a Terraform data source to lookup Git repositories - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
The provider currently only provides a way to provision Azure Git Repos, but does not allow us to reference a pre-existing repository. Having a data source for Git Repos will allow people to leverage other resources, like build definitions and repository permissions, without needing to create a new repository.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* data.azuredevops_git_repository
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azuredevops_project" "main_project" {
project_name = "Test Project"
description = "Test Project Description"
visibility = "private"
enable_tfvc = False
work_item_template = "Agile"
}
data azuredevops_git_repository "lookup-repo" {
project_id = azuredevops_project.main_project.id
name = "Test Lookup Repo"
}
```
### Acceptance Criteria
- [ ] A data source exists that allows for query of an Azure Git Repos based on the following fields:
- [ ] `project_id` - ID of project in which the repository exists
- [ ] `name` - name of repository
- [ ] The `computed` fields of the resource should match the resource implementation of [Azure Git Repos](https://github.com/microsoft/terraform-provider-azuredevops/blob/master/azuredevops/resource_azure_git_repository.go). There may be an opportunity to re-use the whole resource schema!
- [ ] Data source is validated with unit tests
- [ ] Data source is validated with acceptance tests
- [ ] Documentation is added in the client-facing docs
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
* #84
* [Repositories - List](https://docs.microsoft.com/en-us/rest/api/azure/devops/git/repositories/list?view=azure-devops-rest-5.1)
|
non_process
|
implement a terraform data source to lookup git repositories community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the provider currently only provides a way to provision azure git repos but does not allow us to reference a pre existing repository having a data source for git repos will allow people to leverage other resources like build definitions and repository permissions without needing to create a new repository new or affected resource s data azuredevops git repository potential terraform configuration hcl resource azuredevops project main project project name test project description test project description visibility private enable tfvc false work item template agile data azuredevops git repository lookup repo project id azuredevops project main project id name test lookup repo acceptance criteria a data source exists that allows for query of an azure git repos based on the following fields project id id of project in which the repository exists name name of repository the computed fields of the resource should match the resource implementation of there may be an opportunity to re use the whole resource schema data source is validated with unit tests data source is validated with acceptance tests documentation is added in the client facing docs references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation
| 0
|
13,762
| 16,521,836,538
|
IssuesEvent
|
2021-05-26 15:17:53
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
allowPartiallySucceededBuilds for deployment jobs
|
cba devops-cicd-process/tech devops/prod product-question
|
Hey Devops Gurus,
Is there to allowPartiallySucceededBuilds for deployment jobs?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 5aeeaace-1c5b-a51b-e41f-f25b806155b8
* Version Independent ID: fd7ff690-b2e4-41c7-a342-e528b911c6e1
* Content: [Deployment jobs - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/deployment-jobs?view=azure-devops)
* Content Source: [docs/pipelines/process/deployment-jobs.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/deployment-jobs.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
allowPartiallySucceededBuilds for deployment jobs - Hey Devops Gurus,
Is there to allowPartiallySucceededBuilds for deployment jobs?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 5aeeaace-1c5b-a51b-e41f-f25b806155b8
* Version Independent ID: fd7ff690-b2e4-41c7-a342-e528b911c6e1
* Content: [Deployment jobs - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/deployment-jobs?view=azure-devops)
* Content Source: [docs/pipelines/process/deployment-jobs.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/deployment-jobs.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
allowpartiallysucceededbuilds for deployment jobs hey devops gurus is there to allowpartiallysucceededbuilds for deployment jobs document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
22,625
| 31,847,770,452
|
IssuesEvent
|
2023-09-14 21:28:34
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] Port MLv1 `hasWritePermission` to MLv2 (only for native queries)
|
Querying/Native .Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
There're a few MLv1 methods the editor uses that'd be nice to port to MLv2 as well.
Port `hasWritePermission` (only native) to MLv2
https://github.com/metabase/metabase/blob/dbfca6c6d173294ddcf97b394750574b4ef10221/frontend/src/metabase-lib/queries/NativeQuery.ts#L200
|
1.0
|
[MLv2] Port MLv1 `hasWritePermission` to MLv2 (only for native queries) - There're a few MLv1 methods the editor uses that'd be nice to port to MLv2 as well.
Port `hasWritePermission` (only native) to MLv2
https://github.com/metabase/metabase/blob/dbfca6c6d173294ddcf97b394750574b4ef10221/frontend/src/metabase-lib/queries/NativeQuery.ts#L200
|
process
|
port haswritepermission to only for native queries there re a few methods the editor uses that d be nice to port to as well port haswritepermission only native to
| 1
|
75,803
| 3,476,132,228
|
IssuesEvent
|
2015-12-26 14:24:59
|
Stephane-D/SGDK
|
https://api.github.com/repos/Stephane-D/SGDK
|
closed
|
recomp sprite collision
|
Priority-Low
|
Hi, rescomp readme is wrong:
Collision should be set to box by default:
https://github.com/Stephane-D/SGDK/blob/master/tools/rescomp/rescomp.txt#L143
but:
https://github.com/Stephane-D/SGDK/blob/10f2039eacac7d72a6eb63b274af6b95c91a85b8/tools/rescomp/src/sprite.c#L72
it's set to none
|
1.0
|
recomp sprite collision - Hi, rescomp readme is wrong:
Collision should be set to box by default:
https://github.com/Stephane-D/SGDK/blob/master/tools/rescomp/rescomp.txt#L143
but:
https://github.com/Stephane-D/SGDK/blob/10f2039eacac7d72a6eb63b274af6b95c91a85b8/tools/rescomp/src/sprite.c#L72
it's set to none
|
non_process
|
recomp sprite collision hi rescomp readme is wrong collision should be set to box by default but it s set to none
| 0
|
13,548
| 16,090,323,496
|
IssuesEvent
|
2021-04-26 15:55:24
|
kubernetes/minikube
|
https://api.github.com/repos/kubernetes/minikube
|
closed
|
remove depricated jenkins jobs from set pending
|
kind/process lifecycle/stale priority/important-longterm
|
for example
windows virtualbox
podman ,,,
these still are being set as pending,even though we dont run them
|
1.0
|
remove depricated jenkins jobs from set pending - for example
windows virtualbox
podman ,,,
these still are being set as pending,even though we dont run them
|
process
|
remove depricated jenkins jobs from set pending for example windows virtualbox podman these still are being set as pending even though we dont run them
| 1
|
393,762
| 11,624,364,335
|
IssuesEvent
|
2020-02-27 10:38:04
|
dhowe/spectre
|
https://api.github.com/repos/dhowe/spectre
|
closed
|
Add circular cutout to profile image in templates/email.html
|
med-priority needs-verification
|
this is static html, just need the image to look like it does for avatars on the main site
you can test in templates/output.html (a generated file), then move the code to templates/email.html (the template file) when ready
if finished, check for new (unassigned) tickets from @billposters - he was testing last night
(post-experience is still on-hold)

|
1.0
|
Add circular cutout to profile image in templates/email.html - this is static html, just need the image to look like it does for avatars on the main site
you can test in templates/output.html (a generated file), then move the code to templates/email.html (the template file) when ready
if finished, check for new (unassigned) tickets from @billposters - he was testing last night
(post-experience is still on-hold)

|
non_process
|
add circular cutout to profile image in templates email html this is static html just need the image to look like it does for avatars on the main site you can test in templates output html a generated file then move the code to templates email html the template file when ready if finished check for new unassigned tickets from billposters he was testing last night post experience is still on hold
| 0
|
5,863
| 8,682,734,453
|
IssuesEvent
|
2018-12-02 11:41:34
|
bitshares/bitshares-community-ui
|
https://api.github.com/repos/bitshares/bitshares-community-ui
|
closed
|
Backup not functioning
|
Privatekey Backup bug process
|
when clicking on Backup (header) then on 'continue to backup' (the popup) then on 'I understand' button it wont take to the next step.
|
1.0
|
Backup not functioning - when clicking on Backup (header) then on 'continue to backup' (the popup) then on 'I understand' button it wont take to the next step.
|
process
|
backup not functioning when clicking on backup header then on continue to backup the popup then on i understand button it wont take to the next step
| 1
|
7,328
| 10,468,918,467
|
IssuesEvent
|
2019-09-22 17:02:37
|
produvia/ai-platform
|
https://api.github.com/repos/produvia/ai-platform
|
closed
|
Text Classification
|
natural-language-processing task wontfix
|
# Goal(s)
- Assign a sentence or document an appropriate category
# Input(s)
- Sentence or document
# Output(s)
- Category
# Objective Function(s)
- TBD
|
1.0
|
Text Classification - # Goal(s)
- Assign a sentence or document an appropriate category
# Input(s)
- Sentence or document
# Output(s)
- Category
# Objective Function(s)
- TBD
|
process
|
text classification goal s assign a sentence or document an appropriate category input s sentence or document output s category objective function s tbd
| 1
|
12,110
| 14,740,468,313
|
IssuesEvent
|
2021-01-07 09:08:16
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
RE: SA Billing - Save Credit Card to Account
|
anc-process anp-1 ant-bug
|
In GitLab by @kdjstudios on Nov 9, 2018, 11:52
**Submitted by:** "Arianna Screen" <arianna.screen@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-09-49358
**Server:** Internal
**Client/Site:** NA
**Account:** NA
**Issue:**
Today I am running credit cards and following the instructions to save the credit cards.
I have all the information for the payment in and there is no check box to save the credit card information.
|
1.0
|
RE: SA Billing - Save Credit Card to Account - In GitLab by @kdjstudios on Nov 9, 2018, 11:52
**Submitted by:** "Arianna Screen" <arianna.screen@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-09-49358
**Server:** Internal
**Client/Site:** NA
**Account:** NA
**Issue:**
Today I am running credit cards and following the instructions to save the credit cards.
I have all the information for the payment in and there is no check box to save the credit card information.
|
process
|
re sa billing save credit card to account in gitlab by kdjstudios on nov submitted by arianna screen helpdesk server internal client site na account na issue today i am running credit cards and following the instructions to save the credit cards i have all the information for the payment in and there is no check box to save the credit card information
| 1
|
12,854
| 21,012,322,589
|
IssuesEvent
|
2022-03-30 07:53:21
|
renovatebot/renovate
|
https://api.github.com/repos/renovatebot/renovate
|
opened
|
[maven]org.apache.maven.plugins:maven-compiler-plugin get configuration/source and configuration/target
|
type:feature status:requirements priority-5-triage
|
### What would you like Renovate to be able to do?
[springshell-rce-0-day-vulnerability](https://www.cyberkendra.com/2022/03/springshell-rce-0-day-vulnerability.html)
Some vulnerability in the java ecosystem are based on the JDK version, org.apache.maven.plugins:maven-compiler-plugin configuration/source and configuration/target is so important for upgraded dependencies versions
I hope the JDK version can be configured in the renovate configuration file with regexManagers
### If you have any ideas on how this should be implemented, please tell us here.
I hope the JDK version can be configured in the renovate configuration file with regexManagers
### Is this a feature you are interested in implementing yourself?
No
|
1.0
|
[maven]org.apache.maven.plugins:maven-compiler-plugin get configuration/source and configuration/target - ### What would you like Renovate to be able to do?
[springshell-rce-0-day-vulnerability](https://www.cyberkendra.com/2022/03/springshell-rce-0-day-vulnerability.html)
Some vulnerability in the java ecosystem are based on the JDK version, org.apache.maven.plugins:maven-compiler-plugin configuration/source and configuration/target is so important for upgraded dependencies versions
I hope the JDK version can be configured in the renovate configuration file with regexManagers
### If you have any ideas on how this should be implemented, please tell us here.
I hope the JDK version can be configured in the renovate configuration file with regexManagers
### Is this a feature you are interested in implementing yourself?
No
|
non_process
|
org apache maven plugins maven compiler plugin get configuration source and configuration target what would you like renovate to be able to do some vulnerability in the java ecosystem are based on the jdk version org apache maven plugins maven compiler plugin configuration source and configuration target is so important for upgraded dependencies versions i hope the jdk version can be configured in the renovate configuration file with regexmanagers if you have any ideas on how this should be implemented please tell us here i hope the jdk version can be configured in the renovate configuration file with regexmanagers is this a feature you are interested in implementing yourself no
| 0
|
13,343
| 15,801,515,409
|
IssuesEvent
|
2021-04-03 05:17:52
|
PyCQA/flake8
|
https://api.github.com/repos/PyCQA/flake8
|
closed
|
Flake8 3.0 does not work, throws AttributeError exception
|
bug:confirmed component:multiprocessing component:pyflakes priority:high
|
In GitLab by @akittas on Jul 25, 2016, 08:52
Python version: 3.5.2 64-bit
OS: Windows 10 Pro 64 bit
Installation: pip install flake8
pip version: 8.1.2
setuptools version: 25.0.1
flake 8 --version: 3.0.0 (pycodestyle: 2.0.0, mccabe: 0.5.0, pyflakes: 1.2.3) CPython 3.5.2 on Windows
running method: flake8 program.py or python -m flake8 program.py
Flake8 3.0 throws the following exception when executing on any file: **AttributeError: 'FlakesChecker' object has no attribute 'with_doctest'**
```
Traceback (most recent call last):
File "c:\programs\python35\lib\multiprocessing\process.py", line 249, in _bootstrap
self.run()
File "c:\programs\python35\lib\multiprocessing\process.py", line 93, in run
self._target(*self._args, **self._kwargs)
File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 222, in _run_checks_from_queue
checker.run_checks(self.results_queue, self.statistics_queue)
File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 563, in run_checks
self.run_ast_checks()
File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 468, in run_ast_checks
checker = self.run_check(plugin, tree=ast)
File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 448, in run_check
return plugin.execute(**arguments)
File "c:\programs\python35\lib\site-packages\flake8\plugins\manager.py", line 126, in execute
return self.plugin(*args, **kwargs) # pylint: disable=not-callable
File "c:\programs\python35\lib\site-packages\flake8\plugins\pyflakes.py", line 53, in __init__
with_doctest = self.with_doctest
AttributeError: 'FlakesChecker' object has no attribute 'with_doctest'
```
Flake8 2.6.2 works as expected.
|
1.0
|
Flake8 3.0 does not work, throws AttributeError exception - In GitLab by @akittas on Jul 25, 2016, 08:52
Python version: 3.5.2 64-bit
OS: Windows 10 Pro 64 bit
Installation: pip install flake8
pip version: 8.1.2
setuptools version: 25.0.1
flake 8 --version: 3.0.0 (pycodestyle: 2.0.0, mccabe: 0.5.0, pyflakes: 1.2.3) CPython 3.5.2 on Windows
running method: flake8 program.py or python -m flake8 program.py
Flake8 3.0 throws the following exception when executing on any file: **AttributeError: 'FlakesChecker' object has no attribute 'with_doctest'**
```
Traceback (most recent call last):
File "c:\programs\python35\lib\multiprocessing\process.py", line 249, in _bootstrap
self.run()
File "c:\programs\python35\lib\multiprocessing\process.py", line 93, in run
self._target(*self._args, **self._kwargs)
File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 222, in _run_checks_from_queue
checker.run_checks(self.results_queue, self.statistics_queue)
File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 563, in run_checks
self.run_ast_checks()
File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 468, in run_ast_checks
checker = self.run_check(plugin, tree=ast)
File "c:\programs\python35\lib\site-packages\flake8\checker.py", line 448, in run_check
return plugin.execute(**arguments)
File "c:\programs\python35\lib\site-packages\flake8\plugins\manager.py", line 126, in execute
return self.plugin(*args, **kwargs) # pylint: disable=not-callable
File "c:\programs\python35\lib\site-packages\flake8\plugins\pyflakes.py", line 53, in __init__
with_doctest = self.with_doctest
AttributeError: 'FlakesChecker' object has no attribute 'with_doctest'
```
Flake8 2.6.2 works as expected.
|
process
|
does not work throws attributeerror exception in gitlab by akittas on jul python version bit os windows pro bit installation pip install pip version setuptools version flake version pycodestyle mccabe pyflakes cpython on windows running method program py or python m program py throws the following exception when executing on any file attributeerror flakeschecker object has no attribute with doctest traceback most recent call last file c programs lib multiprocessing process py line in bootstrap self run file c programs lib multiprocessing process py line in run self target self args self kwargs file c programs lib site packages checker py line in run checks from queue checker run checks self results queue self statistics queue file c programs lib site packages checker py line in run checks self run ast checks file c programs lib site packages checker py line in run ast checks checker self run check plugin tree ast file c programs lib site packages checker py line in run check return plugin execute arguments file c programs lib site packages plugins manager py line in execute return self plugin args kwargs pylint disable not callable file c programs lib site packages plugins pyflakes py line in init with doctest self with doctest attributeerror flakeschecker object has no attribute with doctest works as expected
| 1
|
452,912
| 32,073,573,391
|
IssuesEvent
|
2023-09-25 09:32:47
|
dappnode/DAppNodeDocs
|
https://api.github.com/repos/dappnode/DAppNodeDocs
|
closed
|
New documentation tree
|
documentation 2023
|
The new documentation will follow this tree:
1. User docs
2. Dev docs
3. Smooth
4. DAO (show warning about deprecation or remove)
Inside user docs we should include:
1. Get started
a. Access Dappnode (via WiFi)
b. Set up
Register
Repository (IPFS + Ethereum)
Host password
WiFi password
c. Discover the features:
Become a validator
Install blockchain node packages
....
5. Access my Dappnode
a. WiFi
b. VPN (local and remote)
c. Local (recovery)
d. ssh (advanced)
6. Install Dappnode
a. ISO
b. script
7. Staking
a. Solo staking:
Network
Clients
Signer
MEV Boost
Smoothing Pool
b. Rocketpool
c. Stakehouse
d. Stakewise (when published)
8. Videos & Tutorials
9. FAQs (get the ones we already have from Discord)
10. Packages (highlight some like HOPR)
11. Ethical Metrics
We should also:
1. Embed videos everywhere they are needed
2. Show highlighted items in documentation index
3. Enable contributions and make it easy for anybody to add new information
4. Be very verbose in clients and speak about the diversity
|
1.0
|
New documentation tree - The new documentation will follow this tree:
1. User docs
2. Dev docs
3. Smooth
4. DAO (show warning about deprecation or remove)
Inside user docs we should include:
1. Get started
a. Access Dappnode (via WiFi)
b. Set up
Register
Repository (IPFS + Ethereum)
Host password
WiFi password
c. Discover the features:
Become a validator
Install blockchain node packages
....
5. Access my Dappnode
a. WiFi
b. VPN (local and remote)
c. Local (recovery)
d. ssh (advanced)
6. Install Dappnode
a. ISO
b. script
7. Staking
a. Solo staking:
Network
Clients
Signer
MEV Boost
Smoothing Pool
b. Rocketpool
c. Stakehouse
d. Stakewise (when published)
8. Videos & Tutorials
9. FAQs (get the ones we already have from Discord)
10. Packages (highlight some like HOPR)
11. Ethical Metrics
We should also:
1. Embed videos everywhere they are needed
2. Show highlighted items in documentation index
3. Enable contributions and make it easy for anybody to add new information
4. Be very verbose in clients and speak about the diversity
|
non_process
|
new documentation tree the new documentation will follow this tree user docs dev docs smooth dao show warning about deprecation or remove inside user docs we should include get started a access dappnode via wifi b set up register repository ipfs ethereum host password wifi password c discover the features become a validator install blockchain node packages access my dappnode a wifi b vpn local and remote c local recovery d ssh advanced install dappnode a iso b script staking a solo staking network clients signer mev boost smoothing pool b rocketpool c stakehouse d stakewise when published videos tutorials faqs get the ones we already have from discord packages highlight some like hopr ethical metrics we should also embed videos everywhere they are needed show highlighted items in documentation index enable contributions and make it easy for anybody to add new information be very verbose in clients and speak about the diversity
| 0
|
3,210
| 9,213,817,246
|
IssuesEvent
|
2019-03-10 15:09:07
|
jimkyndemeyer/js-graphql-intellij-plugin
|
https://api.github.com/repos/jimkyndemeyer/js-graphql-intellij-plugin
|
closed
|
Support for multiple schemas (on multiple endpoints)
|
enhancement v2-alpha v2-architecture
|
I need to access two different graphql endpoints from a project, where each endpoint has a different schema.
I'm currently keeping two different copies of graphql.config.json that I swap between as needed.
Is there a better way to handle this? If not, consider this an enhancement request.
|
1.0
|
Support for multiple schemas (on multiple endpoints) - I need to access two different graphql endpoints from a project, where each endpoint has a different schema.
I'm currently keeping two different copies of graphql.config.json that I swap between as needed.
Is there a better way to handle this? If not, consider this an enhancement request.
|
non_process
|
support for multiple schemas on multiple endpoints i need to access two different graphql endpoints from a project where each endpoint has a different schema i m currently keeping two different copies of graphql config json that i swap between as needed is there a better way to handle this if not consider this an enhancement request
| 0
|
49,212
| 6,157,458,330
|
IssuesEvent
|
2017-06-28 18:59:45
|
mapzen/android
|
https://api.github.com/repos/mapzen/android
|
opened
|
Sample App Major Section: More Info
|
Sample App Redesign
|
- [ ] Links to the places demo
- [ ] Release info
- [ ] Feature list
- [ ] Contact
- [ ] Participate
- [ ] Download
- [ ] API keys
|
1.0
|
Sample App Major Section: More Info - - [ ] Links to the places demo
- [ ] Release info
- [ ] Feature list
- [ ] Contact
- [ ] Participate
- [ ] Download
- [ ] API keys
|
non_process
|
sample app major section more info links to the places demo release info feature list contact participate download api keys
| 0
|
514,457
| 14,939,551,731
|
IssuesEvent
|
2021-01-25 17:05:05
|
idaholab/raven
|
https://api.github.com/repos/idaholab/raven
|
closed
|
[TASK] ROMCollection.Interpolated truncated lifetime
|
FutureRAVENv2.1 priority_minor task
|
--------
Issue Description
--------
For workflow debugging, an option to limit the number of cycles sampled in an Interpolated ROMCollection would be useful.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or task?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
1.0
|
[TASK] ROMCollection.Interpolated truncated lifetime - --------
Issue Description
--------
For workflow debugging, an option to limit the number of cycles sampled in an Interpolated ROMCollection would be useful.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or task?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
non_process
|
romcollection interpolated truncated lifetime issue description for workflow debugging an option to limit the number of cycles sampled in an interpolated romcollection would be useful for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
| 0
|
133,660
| 18,299,017,408
|
IssuesEvent
|
2021-10-05 23:53:38
|
bsbtd/Teste
|
https://api.github.com/repos/bsbtd/Teste
|
opened
|
CVE-2020-11112 (High) detected in jackson-databind-2.9.5.jar, jackson-databind-2.6.7.3.jar
|
security vulnerability
|
## CVE-2020-11112 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.6.7.3.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Teste/liferay-portal/modules/etl/talend/talend-runtime/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- components-api-0.25.3.jar (Root Library)
- daikon-0.27.0.jar
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.7.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Teste/opentelemetry-java/sdk_contrib/aws_v1_support/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.7.3/c8bad9f57a5d298280f8bc0efcb7bf8393a41534/jackson-databind-2.6.7.3.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-ec2-1.11.701.jar (Root Library)
- jmespath-java-1.11.701.jar
- :x: **jackson-databind-2.6.7.3.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.proxy.provider.remoting.RmiProvider (aka apache/commons-proxy).
<p>Publish Date: 2020-03-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11112>CVE-2020-11112</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112</a></p>
<p>Release Date: 2020-03-31</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11112 (High) detected in jackson-databind-2.9.5.jar, jackson-databind-2.6.7.3.jar - ## CVE-2020-11112 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.6.7.3.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Teste/liferay-portal/modules/etl/talend/talend-runtime/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- components-api-0.25.3.jar (Root Library)
- daikon-0.27.0.jar
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.7.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Teste/opentelemetry-java/sdk_contrib/aws_v1_support/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.7.3/c8bad9f57a5d298280f8bc0efcb7bf8393a41534/jackson-databind-2.6.7.3.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-ec2-1.11.701.jar (Root Library)
- jmespath-java-1.11.701.jar
- :x: **jackson-databind-2.6.7.3.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.proxy.provider.remoting.RmiProvider (aka apache/commons-proxy).
<p>Publish Date: 2020-03-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11112>CVE-2020-11112</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112</a></p>
<p>Release Date: 2020-03-31</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in jackson databind jar jackson databind jar cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file teste liferay portal modules etl talend talend runtime pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy components api jar root library daikon jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file teste opentelemetry java sdk contrib aws support build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy aws java sdk jar root library jmespath java jar x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons proxy provider remoting rmiprovider aka apache commons proxy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
| 0
|
21,442
| 29,478,655,563
|
IssuesEvent
|
2023-06-02 02:09:32
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
'webpack-preprocessor': project is missing dependency '@types/bluebird'
|
type: typings npm: @cypress/webpack-preprocessor stale
|
### Current behavior
Cypress project type-checking fails if we try to type-check `cypress/plugins/index.js` and it includes `require('@cypress/webpack-preprocessor')`

### Desired behavior
Type-checking succeeds
### Test code to reproduce
```js
// This function is called when a project is opened or re-opened (e.g. due to the project's config
// changing).
// read more: https://on.cypress.io/plugins-guide
const webpackPreprocessor = require('@cypress/webpack-preprocessor');
/** @type {Cypress.PluginConfig} */
module.exports = (on, config) => {
on(
'file:preprocessor',
webpackPreprocessor({
webpackOptions: {},
}),
);
return config;
};
```
### Versions
Introduced in https://github.com/cypress-io/cypress-webpack-preprocessor/pull/83/files#diff-dcdc3e0b3362edb8fec2a51d3fa51f8fb8af8f70247e06d9887fa934834c9122R2 (merged in v5.2.1)
From looking around, it seems that `@types/bluebird` is not provided as a dependency in any of the cypress packages. It seems weird to me that users need to figure out on their own that this dependency is needed. If usage of `@cypress/webpack-preprocessor` _requires_ a `@types/` module to type-check, I would expect it to be in `dependencies` and installed automatically.
|
1.0
|
'webpack-preprocessor': project is missing dependency '@types/bluebird' - ### Current behavior
Cypress project type-checking fails if we try to type-check `cypress/plugins/index.js` and it includes `require('@cypress/webpack-preprocessor')`

### Desired behavior
Type-checking succeeds
### Test code to reproduce
```js
// This function is called when a project is opened or re-opened (e.g. due to the project's config
// changing).
// read more: https://on.cypress.io/plugins-guide
const webpackPreprocessor = require('@cypress/webpack-preprocessor');
/** @type {Cypress.PluginConfig} */
module.exports = (on, config) => {
on(
'file:preprocessor',
webpackPreprocessor({
webpackOptions: {},
}),
);
return config;
};
```
### Versions
Introduced in https://github.com/cypress-io/cypress-webpack-preprocessor/pull/83/files#diff-dcdc3e0b3362edb8fec2a51d3fa51f8fb8af8f70247e06d9887fa934834c9122R2 (merged in v5.2.1)
From looking around, it seems that `@types/bluebird` is not provided as a dependency in any of the cypress packages. It seems weird to me that users need to figure out on their own that this dependency is needed. If usage of `@cypress/webpack-preprocessor` _requires_ a `@types/` module to type-check, I would expect it to be in `dependencies` and installed automatically.
|
process
|
webpack preprocessor project is missing dependency types bluebird current behavior cypress project type checking fails if we try to type check cypress plugins index js and it includes require cypress webpack preprocessor desired behavior type checking succeeds test code to reproduce js this function is called when a project is opened or re opened e g due to the project s config changing read more const webpackpreprocessor require cypress webpack preprocessor type cypress pluginconfig module exports on config on file preprocessor webpackpreprocessor webpackoptions return config versions introduced in merged in from looking around it seems that types bluebird is not provided as a dependency in any of the cypress packages it seems weird to me that users need to figure out on their own that this dependency is needed if usage of cypress webpack preprocessor requires a types module to type check i would expect it to be in dependencies and installed automatically
| 1
|
347,977
| 24,905,285,730
|
IssuesEvent
|
2022-10-29 06:42:41
|
wmo-im/wis2box
|
https://api.github.com/repos/wmo-im/wis2box
|
closed
|
add installation by VM instructions
|
documentation
|
Per discussion/presentation with @golfvert, add documentation for installing wis2box by way of VM (Proxmox).
|
1.0
|
add installation by VM instructions - Per discussion/presentation with @golfvert, add documentation for installing wis2box by way of VM (Proxmox).
|
non_process
|
add installation by vm instructions per discussion presentation with golfvert add documentation for installing by way of vm proxmox
| 0
|
21,144
| 28,121,790,020
|
IssuesEvent
|
2023-03-31 14:42:04
|
MPMG-DCC-UFMG/C01
|
https://api.github.com/repos/MPMG-DCC-UFMG/C01
|
closed
|
Post mortem DEBUG - Trace Viewer
|
[2] Alta Prioridade [1] Requisito [0] Desenvolvimento [3] Processamento Dinâmico
|
## Comportamento Esperado
Desejamos incluir o "Trace Viewer" do Playwright como ferramenta de debug para coletores dinâmicos. É uma funcionalidade com uma ótima interface que permite a execução passo a passo de um coletor dinâmico, permitindo a visualização das páginas, snapshots, e inspeção do código fonte de cada instante.
## Comportamento Atual
São necessárias ferramentas para debug de coletores, principalmente os que utilizam o mecanismo de passos. O Trace Viewer poderá ser utilizado pela equipe para investigar bugs do sistema. Os usuários de outras trilhas também podem se beneficiar, usando-o para depurar seus próprios coletores e para obterem mais informações sobre futuros erros a serem reportados para nossa equipe.
## Passos para reproduzir o erro
Não se aplica.
## Especificações da Coleta
Não se aplica
## Sistema (caso necessário)
Não se aplica
## Screenshots (caso necessário)
Uma amostra da interface do Trace Viewer:

Mais explicações sobre a ferramenta aqui: https://github.com/mpmg-dcc-ufmg/c01/issues/2333
|
1.0
|
Post mortem DEBUG - Trace Viewer - ## Comportamento Esperado
Desejamos incluir o "Trace Viewer" do Playwright como ferramenta de debug para coletores dinâmicos. É uma funcionalidade com uma ótima interface que permite a execução passo a passo de um coletor dinâmico, permitindo a visualização das páginas, snapshots, e inspeção do código fonte de cada instante.
## Comportamento Atual
São necessárias ferramentas para debug de coletores, principalmente os que utilizam o mecanismo de passos. O Trace Viewer poderá ser utilizado pela equipe para investigar bugs do sistema. Os usuários de outras trilhas também podem se beneficiar, usando-o para depurar seus próprios coletores e para obterem mais informações sobre futuros erros a serem reportados para nossa equipe.
## Passos para reproduzir o erro
Não se aplica.
## Especificações da Coleta
Não se aplica
## Sistema (caso necessário)
Não se aplica
## Screenshots (caso necessário)
Uma amostra da interface do Trace Viewer:

Mais explicações sobre a ferramenta aqui: https://github.com/mpmg-dcc-ufmg/c01/issues/2333
|
process
|
post mortem debug trace viewer comportamento esperado desejamos incluir o trace viewer do playwright como ferramenta de debug para coletores dinâmicos é uma funcionalidade com uma ótima interface que permite a execução passo a passo de um coletor dinâmico permitindo a visualização das páginas snapshots e inspeção do código fonte de cada instante comportamento atual são necessárias ferramentas para debug de coletores principalmente os que utilizam o mecanismo de passos o trace viewer poderá ser utilizado pela equipe para investigar bugs do sistema os usuários de outras trilhas também podem se beneficiar usando o para depurar seus próprios coletores e para obterem mais informações sobre futuros erros a serem reportados para nossa equipe passos para reproduzir o erro não se aplica especificações da coleta não se aplica sistema caso necessário não se aplica screenshots caso necessário uma amostra da interface do trace viewer mais explicações sobre a ferramenta aqui
| 1
|
313,442
| 26,931,362,950
|
IssuesEvent
|
2023-02-07 17:03:22
|
dotnet/source-build
|
https://api.github.com/repos/dotnet/source-build
|
closed
|
Remove security-partners-dotnet pipeline from main branch
|
area-ci-testing
|
It's my understanding that the [security-partners-dotnet.yml](https://github.com/dotnet/installer/blob/main/src/SourceBuild/content/eng/pipelines/security-partners-dotnet.yml) pipeline is not used in the main branch and only used for .NET 6/7 servicing. If so, then this file should be removed from the main branch.
|
1.0
|
Remove security-partners-dotnet pipeline from main branch - It's my understanding that the [security-partners-dotnet.yml](https://github.com/dotnet/installer/blob/main/src/SourceBuild/content/eng/pipelines/security-partners-dotnet.yml) pipeline is not used in the main branch and only used for .NET 6/7 servicing. If so, then this file should be removed from the main branch.
|
non_process
|
remove security partners dotnet pipeline from main branch it s my understanding that the pipeline is not used in the main branch and only used for net servicing if so then this file should be removed from the main branch
| 0
|
102,065
| 11,273,465,603
|
IssuesEvent
|
2020-01-14 16:36:40
|
GluuFederation/casa
|
https://api.github.com/repos/GluuFederation/casa
|
closed
|
Enable Plugins Isolated Way to Persist Configuration
|
Needs Documentation Needs QA enhancement
|
Plugin devs should have the possibility to store relevant configurations in the same database attribute the application uses for settings. It's good to have a single point of configuration
We have to offer a safe means so that developers cannot read or write other plugins configs or core configs
|
1.0
|
Enable Plugins Isolated Way to Persist Configuration - Plugin devs should have the possibility to store relevant configurations in the same database attribute the application uses for settings. It's good to have a single point of configuration
We have to offer a safe means so that developers cannot read or write other plugins configs or core configs
|
non_process
|
enable plugins isolated way to persist configuration plugin devs should have the possibility to store relevant configurations in the same database attribute the application uses for settings it s good to have a single point of configuration we have to offer a safe means so that developers cannot read or write other plugins configs or core configs
| 0
|
8,492
| 11,647,929,489
|
IssuesEvent
|
2020-03-01 17:49:56
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Assert failure currentProcessCpuCount == g_processAffinitySet.Count()
|
area-System.Diagnostics.Process untriaged
|
https://helix.dot.net/api/2019-06-17/jobs/5c24c550-4c42-459f-bbfd-21e69cbf8a27/workitems/System.Diagnostics.Process.Tests/console
`Libraries Test Run checked coreclr Linux x64 Debug `
```
Assert failure(PID 15210 [0x00003b6a], Thread: 15210 [0x3b6a]): currentProcessCpuCount == g_processAffinitySet.Count()
File: /__w/1/s/src/coreclr/src/vm/gcenv.os.cpp Line: 123
Image: /home/helixbot/work/AB3D09A7/p/dotnet
System.Diagnostics.Tests.ProcessTests.TestProcessorAffinity [FAIL]
System.InvalidOperationException : Cannot process request because the process (15210) has exited.
Stack Trace:
System.Diagnostics.Process/src/System/Diagnostics/Process.Unix.cs(343,0): at System.Diagnostics.Process.ThrowIfExited(Boolean refresh)
System.Diagnostics.Process/src/System/Diagnostics/Process.cs(932,0): at System.Diagnostics.Process.EnsureState(State state)
System.Diagnostics.Process/src/System/Diagnostics/Process.Linux.cs(191,0): at System.Diagnostics.Process.set_ProcessorAffinityCore(IntPtr value)
System.Diagnostics.Process/src/System/Diagnostics/Process.cs(532,0): at System.Diagnostics.Process.set_ProcessorAffinity(IntPtr value)
System.Diagnostics.Process/tests/ProcessTests.cs(861,0): at System.Diagnostics.Tests.ProcessTests.TestProcessorAffinity()
System.Diagnostics.Tests.ProcessTests.TestProcessRecycledPid [SKIP]
Condition(s) not met: "IsStressModeEnabled"
Finished: System.Diagnostics.Process.Tests
```
Seen in https://github.com/dotnet/runtime/pull/32538
|
1.0
|
Assert failure currentProcessCpuCount == g_processAffinitySet.Count() - https://helix.dot.net/api/2019-06-17/jobs/5c24c550-4c42-459f-bbfd-21e69cbf8a27/workitems/System.Diagnostics.Process.Tests/console
`Libraries Test Run checked coreclr Linux x64 Debug `
```
Assert failure(PID 15210 [0x00003b6a], Thread: 15210 [0x3b6a]): currentProcessCpuCount == g_processAffinitySet.Count()
File: /__w/1/s/src/coreclr/src/vm/gcenv.os.cpp Line: 123
Image: /home/helixbot/work/AB3D09A7/p/dotnet
System.Diagnostics.Tests.ProcessTests.TestProcessorAffinity [FAIL]
System.InvalidOperationException : Cannot process request because the process (15210) has exited.
Stack Trace:
System.Diagnostics.Process/src/System/Diagnostics/Process.Unix.cs(343,0): at System.Diagnostics.Process.ThrowIfExited(Boolean refresh)
System.Diagnostics.Process/src/System/Diagnostics/Process.cs(932,0): at System.Diagnostics.Process.EnsureState(State state)
System.Diagnostics.Process/src/System/Diagnostics/Process.Linux.cs(191,0): at System.Diagnostics.Process.set_ProcessorAffinityCore(IntPtr value)
System.Diagnostics.Process/src/System/Diagnostics/Process.cs(532,0): at System.Diagnostics.Process.set_ProcessorAffinity(IntPtr value)
System.Diagnostics.Process/tests/ProcessTests.cs(861,0): at System.Diagnostics.Tests.ProcessTests.TestProcessorAffinity()
System.Diagnostics.Tests.ProcessTests.TestProcessRecycledPid [SKIP]
Condition(s) not met: "IsStressModeEnabled"
Finished: System.Diagnostics.Process.Tests
```
Seen in https://github.com/dotnet/runtime/pull/32538
|
process
|
assert failure currentprocesscpucount g processaffinityset count libraries test run checked coreclr linux debug assert failure pid thread currentprocesscpucount g processaffinityset count file w s src coreclr src vm gcenv os cpp line image home helixbot work p dotnet system diagnostics tests processtests testprocessoraffinity system invalidoperationexception cannot process request because the process has exited stack trace system diagnostics process src system diagnostics process unix cs at system diagnostics process throwifexited boolean refresh system diagnostics process src system diagnostics process cs at system diagnostics process ensurestate state state system diagnostics process src system diagnostics process linux cs at system diagnostics process set processoraffinitycore intptr value system diagnostics process src system diagnostics process cs at system diagnostics process set processoraffinity intptr value system diagnostics process tests processtests cs at system diagnostics tests processtests testprocessoraffinity system diagnostics tests processtests testprocessrecycledpid condition s not met isstressmodeenabled finished system diagnostics process tests seen in
| 1
|
87,382
| 25,107,004,353
|
IssuesEvent
|
2022-11-08 17:19:58
|
spack/spack
|
https://api.github.com/repos/spack/spack
|
closed
|
Installation issue: Binary libffi fails checksum verification
|
build-error
|
### Steps to reproduce the issue
```
# On develop branch
spack mirror add binary_mirror https://binaries.spack.io/releases/v0.18
spack buildcache keys --install --trust
spack -vvv install /d6d3lh3
```
### Error message
==> Installing libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd
==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spec.json.sig
gpg: Signature made Tue 31 May 2022 07:27:40 AM UTC using RSA key ID 3DB0C723
gpg: Good signature from "Spack Project Official Binaries <maintainers@spack.io>"
==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4/gcc-7.3.1/libffi-3.4.2/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spack
==> Extracting libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd from binary cache
==> Error: Failed to install libffi due to NoChecksumException: Package tarball failed checksum verification.
It cannot be installed.
### Information on your system
[ec2-user@ip-10-7-228-156 ~]$ spack debug report
* **Spack:** 0.19.0.dev0 (b065d69136023693ace8e4d35d9fdad817d63d93)
* **Python:** 3.7.10
* **Platform:** linux-amzn2-skylake_avx512
* **Concretizer:** clingo
### Additional information
_No response_
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
1.0
|
Installation issue: Binary libffi fails checksum verification - ### Steps to reproduce the issue
```
# On develop branch
spack mirror add binary_mirror https://binaries.spack.io/releases/v0.18
spack buildcache keys --install --trust
spack -vvv install /d6d3lh3
```
### Error message
==> Installing libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd
==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spec.json.sig
gpg: Signature made Tue 31 May 2022 07:27:40 AM UTC using RSA key ID 3DB0C723
gpg: Good signature from "Spack Project Official Binaries <maintainers@spack.io>"
==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4/gcc-7.3.1/libffi-3.4.2/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spack
==> Extracting libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd from binary cache
==> Error: Failed to install libffi due to NoChecksumException: Package tarball failed checksum verification.
It cannot be installed.
### Information on your system
[ec2-user@ip-10-7-228-156 ~]$ spack debug report
* **Spack:** 0.19.0.dev0 (b065d69136023693ace8e4d35d9fdad817d63d93)
* **Python:** 3.7.10
* **Platform:** linux-amzn2-skylake_avx512
* **Concretizer:** clingo
### Additional information
_No response_
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
non_process
|
installation issue binary libffi fails checksum verification steps to reproduce the issue on develop branch spack mirror add binary mirror spack buildcache keys install trust spack vvv install error message installing libffi fetching gpg signature made tue may am utc using rsa key id gpg good signature from spack project official binaries fetching extracting libffi from binary cache error failed to install libffi due to nochecksumexception package tarball failed checksum verification it cannot be installed information on your system spack debug report spack python platform linux skylake concretizer clingo additional information no response general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
| 0
|
17,611
| 23,430,202,294
|
IssuesEvent
|
2022-08-15 00:46:00
|
sparc4-dev/astropop
|
https://api.github.com/repos/sparc4-dev/astropop
|
closed
|
Avoid breaking in image registration lists
|
bug image-processing
|
When the image registration fails for a list, it breaks. Add option to ignore and log the errors only. May be warn it on header too.
|
1.0
|
Avoid breaking in image registration lists - When the image registration fails for a list, it breaks. Add option to ignore and log the errors only. May be warn it on header too.
|
process
|
avoid breaking in image registration lists when the image registration fails for a list it breaks add option to ignore and log the errors only may be warn it on header too
| 1
|
17,554
| 23,368,077,419
|
IssuesEvent
|
2022-08-10 17:07:30
|
nghi-huynh/HPA_HuBMAP_Kaggle_competition
|
https://api.github.com/repos/nghi-huynh/HPA_HuBMAP_Kaggle_competition
|
closed
|
WSI Pre-processing: tiling + tissue segmentation
|
data preprocessing data preparation
|
WSI preprocessing: Tiling + Tissue Segmentation
**Goal**: using simple thresholding techniques like Otsu or Triangle binarization to identify tissue/non-tissue in tiles and save it for training purpose
Refer to [WSI processing: Tiling + Tissue Segmentation](https://www.kaggle.com/code/nghihuynh/wsi-preprocessing-tiling-tissue-segmentation) for more detail
- [x] Rescale
- [x] Thresholding
- [x] Tile + Selection
- [x] Save
|
1.0
|
WSI Pre-processing: tiling + tissue segmentation - WSI preprocessing: Tiling + Tissue Segmentation
**Goal**: using simple thresholding techniques like Otsu or Triangle binarization to identify tissue/non-tissue in tiles and save it for training purpose
Refer to [WSI processing: Tiling + Tissue Segmentation](https://www.kaggle.com/code/nghihuynh/wsi-preprocessing-tiling-tissue-segmentation) for more detail
- [x] Rescale
- [x] Thresholding
- [x] Tile + Selection
- [x] Save
|
process
|
wsi pre processing tiling tissue segmentation wsi preprocessing tiling tissue segmentation goal using simple thresholding techniques like otsu or triangle binarization to identify tissue non tissue in tiles and save it for training purpose refer to for more detail rescale thresholding tile selection save
| 1
|
8,007
| 7,106,350,087
|
IssuesEvent
|
2018-01-16 16:27:06
|
OpenLiberty/open-liberty
|
https://api.github.com/repos/OpenLiberty/open-liberty
|
opened
|
Generate keystore.password during server create
|
Epic in:Security story team:Zombie Apocalypse
|
Implements the resolution of design issue #1175
- During `server create` time [generate a random secure password](http://crunchify.com/java-generate-strong-random-password-securerandom/) into a `keystore.password` property in `${server.config.dir}/server.env`
- We will add a new flag to server create (e.g. `server create myServer --no-password`) which will disable the default random password generation. This will be helpful for situations where tools wish to generate their own value for `keystore.password`, or simply do not wish to generate one.
- Priorities to override keystore password:
1. Use what's configured in the `<keyStore password="..."/>` field in server.xml
2. Use the `keystore.password` value set in the user environment
3. Use the `keystore.password` defined in server.env as the password value for the default keystore element<br><br>
- The `<keyStore>` metatype will be updated as follows:
```xml
<AD id="password" ... ibm:variable="env.keystore.password"/>
```
|
True
|
Generate keystore.password during server create - Implements the resolution of design issue #1175
- During `server create` time [generate a random secure password](http://crunchify.com/java-generate-strong-random-password-securerandom/) into a `keystore.password` property in `${server.config.dir}/server.env`
- We will add a new flag to server create (e.g. `server create myServer --no-password`) which will disable the default random password generation. This will be helpful for situations where tools wish to generate their own value for `keystore.password`, or simply do not wish to generate one.
- Priorities to override keystore password:
1. Use what's configured in the `<keyStore password="..."/>` field in server.xml
2. Use the `keystore.password` value set in the user environment
3. Use the `keystore.password` defined in server.env as the password value for the default keystore element<br><br>
- The `<keyStore>` metatype will be updated as follows:
```xml
<AD id="password" ... ibm:variable="env.keystore.password"/>
```
|
non_process
|
generate keystore password during server create implements the resolution of design issue during server create time into a keystore password property in server config dir server env we will add a new flag to server create e g server create myserver no password which will disable the default random password generation this will be helpful for situations where tools wish to generate their own value for keystore password or simply do not wish to generate one priorities to override keystore password use what s configured in the field in server xml use the keystore password value set in the user environment use the keystore password defined in server env as the password value for the default keystore element the metatype will be updated as follows xml
| 0
|
6,001
| 8,808,922,197
|
IssuesEvent
|
2018-12-27 16:54:37
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
closed
|
office documents from tasks bug
|
2.0.6 Fixed Process bug critical
|
after creating a task, and then going to the documents tab, clicking on manage documents
create new item doesnt update the list, and after editing the document it isnt saved after refreshing the page

|
1.0
|
office documents from tasks bug - after creating a task, and then going to the documents tab, clicking on manage documents
create new item doesnt update the list, and after editing the document it isnt saved after refreshing the page

|
process
|
office documents from tasks bug after creating a task and then going to the documents tab clicking on manage documents create new item doesnt update the list and after editing the document it isnt saved after refreshing the page
| 1
|
13,254
| 15,725,718,794
|
IssuesEvent
|
2021-03-29 10:20:11
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
amazon-import post-processor handles multiple disks improperly
|
bug post-processor/amazon-import remote-plugin/amazon
|
#### Overview of the Issue
In my case, I'm using the hyperv-iso builder and have entries within `disk_additional_size`, when this is passed on to the amazon-import post-processor it ends up finding the wrong disk.
The code in question isn't particularly smart. It iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix (in my case vhdx). The issue with this, on Windows at least, is that `vm-0.vhdx` is found before the core image without the index suffix, i.e. `vm.vhdx` when iterating over the artefacts.
for _, path := range artifact.Files() {
if strings.HasSuffix(path, "."+p.config.Format) {
source = path
break
}
}
It would make sense to perhaps specify which image you would like to be imported, rather than attempting to guess which one. As an additional improvement, Amazon's VM Import supports multiple disk images so this could be implemented (it would still make sense to add an optional config to specify disk images, perhaps using an explicit list or regex/glob).
### Packer version
1.5.1 (master)
### Operating system and Environment details
Windows 10 Pro
amd64
|
1.0
|
amazon-import post-processor handles multiple disks improperly - #### Overview of the Issue
In my case, I'm using the hyperv-iso builder and have entries within `disk_additional_size`, when this is passed on to the amazon-import post-processor it ends up finding the wrong disk.
The code in question isn't particularly smart. It iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix (in my case vhdx). The issue with this, on Windows at least, is that `vm-0.vhdx` is found before the core image without the index suffix, i.e. `vm.vhdx` when iterating over the artefacts.
for _, path := range artifact.Files() {
if strings.HasSuffix(path, "."+p.config.Format) {
source = path
break
}
}
It would make sense to perhaps specify which image you would like to be imported, rather than attempting to guess which one. As an additional improvement, Amazon's VM Import supports multiple disk images so this could be implemented (it would still make sense to add an optional config to specify disk images, perhaps using an explicit list or regex/glob).
### Packer version
1.5.1 (master)
### Operating system and Environment details
Windows 10 Pro
amd64
|
process
|
amazon import post processor handles multiple disks improperly overview of the issue in my case i m using the hyperv iso builder and have entries within disk additional size when this is passed on to the amazon import post processor it ends up finding the wrong disk the code in question isn t particularly smart it iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix in my case vhdx the issue with this on windows at least is that vm vhdx is found before the core image without the index suffix i e vm vhdx when iterating over the artefacts for path range artifact files if strings hassuffix path p config format source path break it would make sense to perhaps specify which image you would like to be imported rather than attempting to guess which one as an additional improvement amazon s vm import supports multiple disk images so this could be implemented it would still make sense to add an optional config to specify disk images perhaps using an explicit list or regex glob packer version master operating system and environment details windows pro
| 1
|
104,515
| 13,095,183,229
|
IssuesEvent
|
2020-08-03 13:43:07
|
canonical-web-and-design/certification.ubuntu.com
|
https://api.github.com/repos/canonical-web-and-design/certification.ubuntu.com
|
closed
|
Hide "Certification Notes" when there are no notes to display.
|
Priority: Medium Redesign
|
Every certificate includes a "Certification Notes" statement that says by default "There are no notes for this release."
It would be cleaner, and make more sense, to hide the "Certification Notes" section rather than displaying it only to say there are none.
|
1.0
|
Hide "Certification Notes" when there are no notes to display. - Every certificate includes a "Certification Notes" statement that says by default "There are no notes for this release."
It would be cleaner, and make more sense, to hide the "Certification Notes" section rather than displaying it only to say there are none.
|
non_process
|
hide certification notes when there are no notes to display every certificate includes a certification notes statement that says by default there are no notes for this release it would be cleaner and make more sense to hide the certification notes section rather than displaying it only to say there are none
| 0
|
131,310
| 12,481,013,847
|
IssuesEvent
|
2020-05-29 21:26:42
|
Students-of-the-city-of-Kostroma/trpo_automation
|
https://api.github.com/repos/Students-of-the-city-of-Kostroma/trpo_automation
|
closed
|
Реализовать unit-тесты на функцию FormListWithLetters
|
Sprint 11 Story Testing documentation
|
Реализовать unit-тесты на функцию FormListWithLetters с основой на сценарии по [ссылке ](https://docs.google.com/spreadsheets/d/1IHaHZS0_vT2SRZCWRdLPDvhNNvPcKf5QjA6GztFQgF4/edit#gid=208599158)
|
1.0
|
Реализовать unit-тесты на функцию FormListWithLetters - Реализовать unit-тесты на функцию FormListWithLetters с основой на сценарии по [ссылке ](https://docs.google.com/spreadsheets/d/1IHaHZS0_vT2SRZCWRdLPDvhNNvPcKf5QjA6GztFQgF4/edit#gid=208599158)
|
non_process
|
реализовать unit тесты на функцию formlistwithletters реализовать unit тесты на функцию formlistwithletters с основой на сценарии по
| 0
|
326,735
| 24,099,051,895
|
IssuesEvent
|
2022-09-19 21:48:15
|
marcosdosea/AgendeMe
|
https://api.github.com/repos/marcosdosea/AgendeMe
|
closed
|
Atualizar collaborations na seção Design Model, "ManterProfissional", "ManterOrgaoPublico", "ManterServicoPublico"
|
documentation 2.0
|
# Objetivo
## Collaborations com a Persistência
Atualizar a persistência para o novo artefato AgendeMeContext. Por segurança, basta apenas usar o comando de Delete(não precisa usar o comando Delete From Model)
Resultado deve ser
Antes:

Depois:

Importante lembrar de utilizar os métodos oferecidos pela classe AgendeMeContext!
Então para adicionar, utilizar comando Add e na sequência o Save Changes.
O mesmo para as demais alterações, sejam de Remover(Delete), Add(Alterar/Adicionar), Select(consultar).
O método Save Changes sempre irá ser executado após qualquer um dos comandos acimas, exceto o "Select"
|
1.0
|
Atualizar collaborations na seção Design Model, "ManterProfissional", "ManterOrgaoPublico", "ManterServicoPublico" - # Objetivo
## Collaborations com a Persistência
Atualizar a persistência para o novo artefato AgendeMeContext. Por segurança, basta apenas usar o comando de Delete(não precisa usar o comando Delete From Model)
Resultado deve ser
Antes:

Depois:

Importante lembrar de utilizar os métodos oferecidos pela classe AgendeMeContext!
Então para adicionar, utilizar comando Add e na sequência o Save Changes.
O mesmo para as demais alterações, sejam de Remover(Delete), Add(Alterar/Adicionar), Select(consultar).
O método Save Changes sempre irá ser executado após qualquer um dos comandos acimas, exceto o "Select"
|
non_process
|
atualizar collaborations na seção design model manterprofissional manterorgaopublico manterservicopublico objetivo collaborations com a persistência atualizar a persistência para o novo artefato agendemecontext por segurança basta apenas usar o comando de delete não precisa usar o comando delete from model resultado deve ser antes depois importante lembrar de utilizar os métodos oferecidos pela classe agendemecontext então para adicionar utilizar comando add e na sequência o save changes o mesmo para as demais alterações sejam de remover delete add alterar adicionar select consultar o método save changes sempre irá ser executado após qualquer um dos comandos acimas exceto o select
| 0
|
325,893
| 24,065,186,588
|
IssuesEvent
|
2022-09-17 11:32:33
|
pyxem/orix
|
https://api.github.com/repos/pyxem/orix
|
closed
|
Restructure documentation into examples, tutorials and an improved API reference
|
documentation
|
I plan to restructure the orix documentation in a similar fashion as [I've done in kikuchipy](https://kikuchipy.org/en/latest/#learning-resources), based on the [Diátaxis](https://diataxis.fr/) framework.
The changes I plan:
* Separate the user guide into *Examples* and *Tutorials*. The examples are task oriented (no/minimal explanations), while the tutorials are learning oriented.
* Examples are short recipies of common tasks in orix (e.g. #370, #364, #332) represented in a [sphinx-gallery](https://sphinx-gallery.github.io/stable/index.html) from Python scripts (not notebooks), very similar to Matplotlib's or scikit-images'. Whenever we have nice plot we've made or a procedure which we think might be useful to others, we can add an example to the appropriate section (or make a new section). This is better than trying to shoehorn more stuff into the appropriate tutorial. See the [kikuchipy examples](https://kikuchipy.org/en/latest/examples/index.html) for an example.
* Citing Diátaxis, tutorials are *lessons that take the reader by the hand through a series of steps to complete a project of some kind.* These will be the existing user guides, built from notebooks with `nbsphinx`. See the [kikuchipy tutorials](https://kikuchipy.org/en/latest/tutorials/index.html) for an example.
* The API reference will get an overhaul, with one listing the top modules, then one page per module listing classes and functions, then one page per class listing methods and attributes, then one page per method and attribute. These pages are automatically generated when building the docs with `make html`, and using custom Sphinx templates and some smart use of Sphinx, we shouldn't have to remember to list every method and attribute manually in `reference.rst`. The result is a reference that is simpler to navigate and lists the example(s) where this particular functionality is used. See the [kikuchipy API](https://kikuchipy.org/en/latest/reference/generated/kikuchipy.data.nickel_ebsd_master_pattern_small.html#kikuchipy.data.nickel_ebsd_master_pattern_small) for an example.
I've done this once for kikuchipy, so shouldn't be too difficult to do for orix.
|
1.0
|
Restructure documentation into examples, tutorials and an improved API reference - I plan to restructure the orix documentation in a similar fashion as [I've done in kikuchipy](https://kikuchipy.org/en/latest/#learning-resources), based on the [Diátaxis](https://diataxis.fr/) framework.
The changes I plan:
* Separate the user guide into *Examples* and *Tutorials*. The examples are task oriented (no/minimal explanations), while the tutorials are learning oriented.
* Examples are short recipies of common tasks in orix (e.g. #370, #364, #332) represented in a [sphinx-gallery](https://sphinx-gallery.github.io/stable/index.html) from Python scripts (not notebooks), very similar to Matplotlib's or scikit-images'. Whenever we have nice plot we've made or a procedure which we think might be useful to others, we can add an example to the appropriate section (or make a new section). This is better than trying to shoehorn more stuff into the appropriate tutorial. See the [kikuchipy examples](https://kikuchipy.org/en/latest/examples/index.html) for an example.
* Citing Diátaxis, tutorials are *lessons that take the reader by the hand through a series of steps to complete a project of some kind.* These will be the existing user guides, built from notebooks with `nbsphinx`. See the [kikuchipy tutorials](https://kikuchipy.org/en/latest/tutorials/index.html) for an example.
* The API reference will get an overhaul, with one listing the top modules, then one page per module listing classes and functions, then one page per class listing methods and attributes, then one page per method and attribute. These pages are automatically generated when building the docs with `make html`, and using custom Sphinx templates and some smart use of Sphinx, we shouldn't have to remember to list every method and attribute manually in `reference.rst`. The result is a reference that is simpler to navigate and lists the example(s) where this particular functionality is used. See the [kikuchipy API](https://kikuchipy.org/en/latest/reference/generated/kikuchipy.data.nickel_ebsd_master_pattern_small.html#kikuchipy.data.nickel_ebsd_master_pattern_small) for an example.
I've done this once for kikuchipy, so shouldn't be too difficult to do for orix.
|
non_process
|
restructure documentation into examples tutorials and an improved api reference i plan to restructure the orix documentation in a similar fashion as based on the framework the changes i plan separate the user guide into examples and tutorials the examples are task oriented no minimal explanations while the tutorials are learning oriented examples are short recipies of common tasks in orix e g represented in a from python scripts not notebooks very similar to matplotlib s or scikit images whenever we have nice plot we ve made or a procedure which we think might be useful to others we can add an example to the appropriate section or make a new section this is better than trying to shoehorn more stuff into the appropriate tutorial see the for an example citing diátaxis tutorials are lessons that take the reader by the hand through a series of steps to complete a project of some kind these will be the existing user guides built from notebooks with nbsphinx see the for an example the api reference will get an overhaul with one listing the top modules then one page per module listing classes and functions then one page per class listing methods and attributes then one page per method and attribute these pages are automatically generated when building the docs with make html and using custom sphinx templates and some smart use of sphinx we shouldn t have to remember to list every method and attribute manually in reference rst the result is a reference that is simpler to navigate and lists the example s where this particular functionality is used see the for an example i ve done this once for kikuchipy so shouldn t be too difficult to do for orix
| 0
|
18,277
| 24,357,836,694
|
IssuesEvent
|
2022-10-03 09:03:54
|
aiidateam/aiida-core
|
https://api.github.com/repos/aiidateam/aiida-core
|
closed
|
Add command to `verdi` to "revive" dormant process
|
topic/verdi type/accepted feature priority/nice-to-have topic/processes
|
There is a bug where it is possible that the process task is lost with RabbitMQ before the process is completed. This will cause that the process is never picked up by the daemon and will be dormant. It will also not respond to RPC calls, such as `verdi process play` etc. There is a way to recreate the task that can revive the task
```
from aiida.manage import get_manager
get_manager().get_process_controller().continue_process(PID)
```
We have had to communicate this multiple times to users during debugging. It was never added as an official method to the CLI since it is treating a symptom, and when used incorrectly can cause bigger problems, such as the process and its children being executed by multiple workers. Still, everytime we have to use it, having users copy paste it, is not ideal.
|
1.0
|
Add command to `verdi` to "revive" dormant process - There is a bug where it is possible that the process task is lost with RabbitMQ before the process is completed. This will cause that the process is never picked up by the daemon and will be dormant. It will also not respond to RPC calls, such as `verdi process play` etc. There is a way to recreate the task that can revive the task
```
from aiida.manage import get_manager
get_manager().get_process_controller().continue_process(PID)
```
We have had to communicate this multiple times to users during debugging. It was never added as an official method to the CLI since it is treating a symptom, and when used incorrectly can cause bigger problems, such as the process and its children being executed by multiple workers. Still, everytime we have to use it, having users copy paste it, is not ideal.
|
process
|
add command to verdi to revive dormant process there is a bug where it is possible that the process task is lost with rabbitmq before the process is completed this will cause that the process is never picked up by the daemon and will be dormant it will also not respond to rpc calls such as verdi process play etc there is a way to recreate the task that can revive the task from aiida manage import get manager get manager get process controller continue process pid we have had to communicate this multiple times to users during debugging it was never added as an official method to the cli since it is treating a symptom and when used incorrectly can cause bigger problems such as the process and its children being executed by multiple workers still everytime we have to use it having users copy paste it is not ideal
| 1
|
11,537
| 14,410,655,161
|
IssuesEvent
|
2020-12-04 05:27:30
|
nion-software/nionswift
|
https://api.github.com/repos/nion-software/nionswift
|
closed
|
1D-Collection of 1D-Data does not update display when changing "Index" slider in Inspector
|
f - displays f - processing f - sequences stage - planning type - bug
|
To reproduce, open a console and run:
```python
show(numpy.random.rand(2, 32))
```
Then go to `Processing -> Redimension Data ->Redimension to Collection of 2 Spectra of Length 32`
Try changing the "Index" slider in the Inspector to see the second "slice" of data. The display does not update.
This is with current master branch of nionswift.
|
1.0
|
1D-Collection of 1D-Data does not update display when changing "Index" slider in Inspector - To reproduce, open a console and run:
```python
show(numpy.random.rand(2, 32))
```
Then go to `Processing -> Redimension Data ->Redimension to Collection of 2 Spectra of Length 32`
Try changing the "Index" slider in the Inspector to see the second "slice" of data. The display does not update.
This is with current master branch of nionswift.
|
process
|
collection of data does not update display when changing index slider in inspector to reproduce open a console and run python show numpy random rand then go to processing redimension data redimension to collection of spectra of length try changing the index slider in the inspector to see the second slice of data the display does not update this is with current master branch of nionswift
| 1
|
107,255
| 11,524,588,977
|
IssuesEvent
|
2020-02-15 01:26:36
|
10quality/wpmvc-website
|
https://api.github.com/repos/10quality/wpmvc-website
|
opened
|
Gulp watch documentation
|
documentation
|
Add documentation for gulp commands:
```bash
gulp watch
gulp watch-js
gulp watch-styles
```
|
1.0
|
Gulp watch documentation - Add documentation for gulp commands:
```bash
gulp watch
gulp watch-js
gulp watch-styles
```
|
non_process
|
gulp watch documentation add documentation for gulp commands bash gulp watch gulp watch js gulp watch styles
| 0
|
12,192
| 14,742,313,034
|
IssuesEvent
|
2021-01-07 12:04:37
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Payment Revert Issue
|
anc-process anp-1 ant-bug ant-child/secondary ant-enhancement
|
In GitLab by @pchaudhary on Apr 8, 2019, 07:24
The same billing cycle payment is not reverting on the invoice if there are more than 1 invoice for the same billing cycle.
* Let's say there are 2 invoices
* I1 = $950
* I2 = $950
* Acc Bal is `$1900`
* We made a payment of $ 950. (Manual Payment)
* Acc Bal is `$950`
* Revert the payment.
* Balance on I1 is $0
* Balance on I2 is $950
* Acc Bal is `$1900`
|
1.0
|
Payment Revert Issue - In GitLab by @pchaudhary on Apr 8, 2019, 07:24
The same billing cycle payment is not reverting on the invoice if there are more than 1 invoice for the same billing cycle.
* Let's say there are 2 invoices
* I1 = $950
* I2 = $950
* Acc Bal is `$1900`
* We made a payment of $ 950. (Manual Payment)
* Acc Bal is `$950`
* Revert the payment.
* Balance on I1 is $0
* Balance on I2 is $950
* Acc Bal is `$1900`
|
process
|
payment revert issue in gitlab by pchaudhary on apr the same billing cycle payment is not reverting on the invoice if there are more than invoice for the same billing cycle let s say there are invoices acc bal is we made a payment of manual payment acc bal is revert the payment balance on is balance on is acc bal is
| 1
|
760,387
| 26,638,329,282
|
IssuesEvent
|
2023-01-25 00:41:23
|
sczerwinski/wavefront-obj-intellij-plugin
|
https://api.github.com/repos/sczerwinski/wavefront-obj-intellij-plugin
|
closed
|
Face culling setting
|
type:feature resolution:done priority:low component:3d component:settings
|
Add face culling selection in 3D preview settings from `graphics.glimpse.FaceCullingMode` enum.
- `DISABLED` – disabled
- `FRONT` – front-facing facets
- `BACK` – back-facing facets
- `FRONT_AND_BACK` – both front- and back-facing facets (does it make sense?)
|
1.0
|
Face culling setting - Add face culling selection in 3D preview settings from `graphics.glimpse.FaceCullingMode` enum.
- `DISABLED` – disabled
- `FRONT` – front-facing facets
- `BACK` – back-facing facets
- `FRONT_AND_BACK` – both front- and back-facing facets (does it make sense?)
|
non_process
|
face culling setting add face culling selection in preview settings from graphics glimpse facecullingmode enum disabled – disabled front – front facing facets back – back facing facets front and back – both front and back facing facets does it make sense
| 0
|
199,312
| 22,693,285,263
|
IssuesEvent
|
2022-07-05 01:07:57
|
Baneeishaque/printing_press_erp-Maven
|
https://api.github.com/repos/Baneeishaque/printing_press_erp-Maven
|
opened
|
CVE-2020-36518 (High) detected in jackson-databind-2.10.0.jar
|
security vulnerability
|
## CVE-2020-36518 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.10.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.0/jackson-databind-2.10.0.jar</p>
<p>
Dependency Hierarchy:
- jasperreports-6.13.0.jar (Root Library)
- :x: **jackson-databind-2.10.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.
Mend Note: After conducting further research, Mend has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.
<p>Publish Date: 2022-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-11</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.12.6.1</p>
<p>Direct dependency fix Resolution (net.sf.jasperreports:jasperreports): 6.19.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-36518 (High) detected in jackson-databind-2.10.0.jar - ## CVE-2020-36518 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.10.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.0/jackson-databind-2.10.0.jar</p>
<p>
Dependency Hierarchy:
- jasperreports-6.13.0.jar (Root Library)
- :x: **jackson-databind-2.10.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.
Mend Note: After conducting further research, Mend has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.
<p>Publish Date: 2022-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-11</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.12.6.1</p>
<p>Direct dependency fix Resolution (net.sf.jasperreports:jasperreports): 6.19.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy jasperreports jar root library x jackson databind jar vulnerable library found in base branch master vulnerability details jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects mend note after conducting further research mend has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution net sf jasperreports jasperreports step up your open source security game with mend
| 0
|
352,094
| 10,531,779,068
|
IssuesEvent
|
2019-10-01 09:16:51
|
AY1920S1-CS2103T-T13-4/main
|
https://api.github.com/repos/AY1920S1-CS2103T-T13-4/main
|
opened
|
Implement full autotag command
|
priority.High type.Story
|
- [ ] Include not-conditions (`nu/`, `nt/`, `nf/`) in search
- [ ] Apply autotag whenever bookmarks are updated
Extension of #39
|
1.0
|
Implement full autotag command - - [ ] Include not-conditions (`nu/`, `nt/`, `nf/`) in search
- [ ] Apply autotag whenever bookmarks are updated
Extension of #39
|
non_process
|
implement full autotag command include not conditions nu nt nf in search apply autotag whenever bookmarks are updated extension of
| 0
|
3,079
| 6,096,823,903
|
IssuesEvent
|
2017-06-20 00:29:20
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
vagrant-cloud lacks the ability to specify a provider causing version collisions
|
enhancement post-processor/vagrant
|
I updated my packer configs so they would use the vagrant-cloud post processor instead of the atlas post-processor, and have run into a few problems.
1) The biggest issue seems to be alternate versions of a box, intended for different providers get rejected because the version numbers collide. After glancing at the code, it appears the plugin is using the input id to determine the provider, which doesn't work well with complicated json files. II can't set the output for the every artifact to virtualbox.box... that would create quite the mess. Can we add the ability to override the provider string, in the same way we could with atlas plugin? If we're going with 'box_tag' then perhaps 'box_provider' would make the most sense for a json label.
2) Packer doesn't seem to use the ATLAS_TOKEN with the vagrant cloud plugin. Is there an equivalent environment variable? I'm reading in the environment variable manually at the moment, and then setting the token attribute, but it's clunky.
|
1.0
|
vagrant-cloud lacks the ability to specify a provider causing version collisions - I updated my packer configs so they would use the vagrant-cloud post processor instead of the atlas post-processor, and have run into a few problems.
1) The biggest issue seems to be alternate versions of a box, intended for different providers get rejected because the version numbers collide. After glancing at the code, it appears the plugin is using the input id to determine the provider, which doesn't work well with complicated json files. II can't set the output for the every artifact to virtualbox.box... that would create quite the mess. Can we add the ability to override the provider string, in the same way we could with atlas plugin? If we're going with 'box_tag' then perhaps 'box_provider' would make the most sense for a json label.
2) Packer doesn't seem to use the ATLAS_TOKEN with the vagrant cloud plugin. Is there an equivalent environment variable? I'm reading in the environment variable manually at the moment, and then setting the token attribute, but it's clunky.
|
process
|
vagrant cloud lacks the ability to specify a provider causing version collisions i updated my packer configs so they would use the vagrant cloud post processor instead of the atlas post processor and have run into a few problems the biggest issue seems to be alternate versions of a box intended for different providers get rejected because the version numbers collide after glancing at the code it appears the plugin is using the input id to determine the provider which doesn t work well with complicated json files ii can t set the output for the every artifact to virtualbox box that would create quite the mess can we add the ability to override the provider string in the same way we could with atlas plugin if we re going with box tag then perhaps box provider would make the most sense for a json label packer doesn t seem to use the atlas token with the vagrant cloud plugin is there an equivalent environment variable i m reading in the environment variable manually at the moment and then setting the token attribute but it s clunky
| 1
|
19,136
| 25,192,838,214
|
IssuesEvent
|
2022-11-12 05:41:14
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
ARM disassembler for ARM:LE:32:v5t:default no longer works in Ghidra 10.2
|
Feature: Processor/ARM Status: Internal
|
**Describe the bug**
The ARM disassembler for language ARM:LE:32:v5t:default no longer works in Ghidra 10.2.
I assume the same issue might apply for other ARM versions, but I haven't checked.
**To Reproduce**
Steps to reproduce the behavior:
1. Create a new project.
2. Open the CodeBrowser.
3. Import an armv5t executable appropriate for the language setting ARM:LE:32:v5t:default, run analysis, and navigate to any function.
4. Observe that instructions are not disassembled. Explicitly disassembling at the offset gives an error with the message "Error [Bad Instruction]: Unable to resolve constructor at \<address\>".
**Expected behavior**
ARM disassembly should work. It used to work in Ghidra 10.1.2 (I've also tested Ghidra 10.1.5, and it still works there).
**Screenshots**
Disassembly worked in Ghidra 10.1.2: 
Disassembly no longer works in Ghidra 10.2: 
**Attachments**
Here's a simple, no-op C program cross-compiled for armv5te (but the issue remains if compiled for armv5t) with `arm-none-eabi-gcc`: [a.out.zip](https://github.com/NationalSecurityAgency/ghidra/files/9944909/a.out.zip). The above screenshots are taken using this binary (`main()` is at offset 0x8000).
**Environment (please complete the following information):**
- OS: macOS 11.7, but I've also tested on Windows 10 and the same issue occurs
- Java Version: 17.0.1
- Ghidra Version: 10.2
- Ghidra Origin: official GitHub distro
|
1.0
|
ARM disassembler for ARM:LE:32:v5t:default no longer works in Ghidra 10.2 - **Describe the bug**
The ARM disassembler for language ARM:LE:32:v5t:default no longer works in Ghidra 10.2.
I assume the same issue might apply for other ARM versions, but I haven't checked.
**To Reproduce**
Steps to reproduce the behavior:
1. Create a new project.
2. Open the CodeBrowser.
3. Import an armv5t executable appropriate for the language setting ARM:LE:32:v5t:default, run analysis, and navigate to any function.
4. Observe that instructions are not disassembled. Explicitly disassembling at the offset gives an error with the message "Error [Bad Instruction]: Unable to resolve constructor at \<address\>".
**Expected behavior**
ARM disassembly should work. It used to work in Ghidra 10.1.2 (I've also tested Ghidra 10.1.5, and it still works there).
**Screenshots**
Disassembly worked in Ghidra 10.1.2: 
Disassembly no longer works in Ghidra 10.2: 
**Attachments**
Here's a simple, no-op C program cross-compiled for armv5te (but the issue remains if compiled for armv5t) with `arm-none-eabi-gcc`: [a.out.zip](https://github.com/NationalSecurityAgency/ghidra/files/9944909/a.out.zip). The above screenshots are taken using this binary (`main()` is at offset 0x8000).
**Environment (please complete the following information):**
- OS: macOS 11.7, but I've also tested on Windows 10 and the same issue occurs
- Java Version: 17.0.1
- Ghidra Version: 10.2
- Ghidra Origin: official GitHub distro
|
process
|
arm disassembler for arm le default no longer works in ghidra describe the bug the arm disassembler for language arm le default no longer works in ghidra i assume the same issue might apply for other arm versions but i haven t checked to reproduce steps to reproduce the behavior create a new project open the codebrowser import an executable appropriate for the language setting arm le default run analysis and navigate to any function observe that instructions are not disassembled explicitly disassembling at the offset gives an error with the message error unable to resolve constructor at expected behavior arm disassembly should work it used to work in ghidra i ve also tested ghidra and it still works there screenshots disassembly worked in ghidra disassembly no longer works in ghidra attachments here s a simple no op c program cross compiled for but the issue remains if compiled for with arm none eabi gcc the above screenshots are taken using this binary main is at offset environment please complete the following information os macos but i ve also tested on windows and the same issue occurs java version ghidra version ghidra origin official github distro
| 1
|
46,378
| 7,254,803,182
|
IssuesEvent
|
2018-02-16 12:33:10
|
RubenVerborgh/N3.js
|
https://api.github.com/repos/RubenVerborgh/N3.js
|
closed
|
Store doesn't support adding triples using prefixed IRIs
|
documentation question
|
In the documentation it says that the prefixes would support the addition **and** lookup of triples, but when adding a triple with a previously added prefix the prefixed URI is not being expanded and the triple is being store with the prefixed URI. I would have expected the triple to be stored with the expanded absolute URIs.
The lookup with prefixed URIs works fine.
Is this the intended behaviour?
|
1.0
|
Store doesn't support adding triples using prefixed IRIs - In the documentation it says that the prefixes would support the addition **and** lookup of triples, but when adding a triple with a previously added prefix the prefixed URI is not being expanded and the triple is being store with the prefixed URI. I would have expected the triple to be stored with the expanded absolute URIs.
The lookup with prefixed URIs works fine.
Is this the intended behaviour?
|
non_process
|
store doesn t support adding triples using prefixed iris in the documentation it says that the prefixes would support the addition and lookup of triples but when adding a triple with a previously added prefix the prefixed uri is not being expanded and the triple is being store with the prefixed uri i would have expected the triple to be stored with the expanded absolute uris the lookup with prefixed uris works fine is this the intended behaviour
| 0
|
74,233
| 3,436,560,317
|
IssuesEvent
|
2015-12-12 13:59:05
|
pywinauto/pywinauto
|
https://api.github.com/repos/pywinauto/pywinauto
|
opened
|
ActiveX controls support for desktop apps
|
enhancement Priority-Low
|
Native apps can embed ActiveX controls which text or other data cannot be recognized by pywinauto. There are few MFC samples including ActiveX controls: https://msdn.microsoft.com/en-us/library/aa728874(v=vs.71).aspx
Related question on StackOverflow: http://stackoverflow.com/q/34181583/3648361
|
1.0
|
ActiveX controls support for desktop apps - Native apps can embed ActiveX controls which text or other data cannot be recognized by pywinauto. There are few MFC samples including ActiveX controls: https://msdn.microsoft.com/en-us/library/aa728874(v=vs.71).aspx
Related question on StackOverflow: http://stackoverflow.com/q/34181583/3648361
|
non_process
|
activex controls support for desktop apps native apps can embed activex controls which text or other data cannot be recognized by pywinauto there are few mfc samples including activex controls related question on stackoverflow
| 0
|
15,356
| 19,529,633,938
|
IssuesEvent
|
2021-12-30 14:26:12
|
MikeKSmith/The_Lazy_Producer
|
https://api.github.com/repos/MikeKSmith/The_Lazy_Producer
|
opened
|
Discuss alternative / non-linear pattern sequencers like Less Concepts, Euclidean, Turing machines
|
process
|
As a process, should discuss alternative sequencer types than the standard, "linear" type with fixed length and fixed divisions.
Less Concepts is random and unpredictable. Might be interesting to include this in here.
|
1.0
|
Discuss alternative / non-linear pattern sequencers like Less Concepts, Euclidean, Turing machines - As a process, should discuss alternative sequencer types than the standard, "linear" type with fixed length and fixed divisions.
Less Concepts is random and unpredictable. Might be interesting to include this in here.
|
process
|
discuss alternative non linear pattern sequencers like less concepts euclidean turing machines as a process should discuss alternative sequencer types than the standard linear type with fixed length and fixed divisions less concepts is random and unpredictable might be interesting to include this in here
| 1
|
15,734
| 19,910,044,899
|
IssuesEvent
|
2022-01-25 16:18:53
|
opensearch-project/data-prepper
|
https://api.github.com/repos/opensearch-project/data-prepper
|
closed
|
Support parsing messages with key-value strings such as queries and properties
|
enhancement plugin - processor KeyValueProcessor
|
Data Prepper should have a processor which can parse key-value strings from one field and save the results into another field as a map.
For example, on the input event:
```
{
"query": "key1:value1&key2:value2&key3:value3"
}
```
Could update the event to include a field:
```
{
"query_params" : {
"key1" : "value1",
"key2" : "value2",
"key3" : "value3"
}
}
```
This processor must allow configuration of:
* The source field (`query` in the example above)
* The destination field (`query_params` in the example above)
* The delimiter between fields (`&` in the example above)
* The delimiter between key and value (`:` in the example above)
* non_match value - default `null`, the value to assign a key when it has no match
* prefix - A prefix to all keys. default `""`.
It should also support regex expressions for both types of delimiters.
|
2.0
|
Support parsing messages with key-value strings such as queries and properties - Data Prepper should have a processor which can parse key-value strings from one field and save the results into another field as a map.
For example, on the input event:
```
{
"query": "key1:value1&key2:value2&key3:value3"
}
```
Could update the event to include a field:
```
{
"query_params" : {
"key1" : "value1",
"key2" : "value2",
"key3" : "value3"
}
}
```
This processor must allow configuration of:
* The source field (`query` in the example above)
* The destination field (`query_params` in the example above)
* The delimiter between fields (`&` in the example above)
* The delimiter between key and value (`:` in the example above)
* non_match value - default `null`, the value to assign a key when it has no match
* prefix - A prefix to all keys. default `""`.
It should also support regex expressions for both types of delimiters.
|
process
|
support parsing messages with key value strings such as queries and properties data prepper should have a processor which can parse key value strings from one field and save the results into another field as a map for example on the input event query could update the event to include a field query params this processor must allow configuration of the source field query in the example above the destination field query params in the example above the delimiter between fields in the example above the delimiter between key and value in the example above non match value default null the value to assign a key when it has no match prefix a prefix to all keys default it should also support regex expressions for both types of delimiters
| 1
|
49,949
| 26,403,450,982
|
IssuesEvent
|
2023-01-13 04:53:26
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
opened
|
Linq enumeration of a string's characters is twice as slow as enumerating them using string.ToCharArray()
|
tenet-performance
|
### Description
Enumerating characters of a string using Linq is slower than doing it on its copy created by `ToCharArray()`. Because of this behavior, people might start seeing `ToCharArray()` as a performance magic (there are already some stackoverflow entries about it), and it might get used everywhere, causing memory overhead unnecessarily.
Given the benchmark below:
```csharp
[MemoryDiagnoser]
public partial class StringVsToCharArrayBenchmark
{
private static string longText = new string('a', 1000000);
[Benchmark]
public bool Linq_ToCharArray()
{
return longText.ToCharArray().Where(c => c == 'a').Count() > 100;
}
[Benchmark]
public bool Linq_String()
{
return longText.Where(c => c == 'a').Count() > 100;
}
static void Main()
{
_ = BenchmarkRunner.Run<StringVsToCharArrayBenchmark>();
}
}
```
The benchmark `Linq_String` is twice as slow as `Linq_ToCharArray` despite that ToCharArray() allocates two strings worth of extra memory. The relative 2x performance difference doesn't change based on the string size. Here is the BenchmarkDotNet output for the given code above:
| Method | Mean | Error | StdDev | Gen0 | Gen1 | Gen2 | Allocated |
|----------------- |---------:|----------:|----------:|---------:|---------:|---------:|----------:|
| Linq_ToCharArray | 1.858 ms | 0.0289 ms | 0.0271 ms | 253.9063 | 253.9063 | 253.9063 | 2000935 B |
| Linq_String | 4.839 ms | 0.0954 ms | 0.0980 ms | - | - | - | 92 B |
(`foreach (var s in string)` is still the fastest of the bunch of course)
### Configuration
```
BenchmarkDotNet=v0.13.3, OS=Windows 11 (10.0.22623.1095)
AMD Ryzen 9 5950X, 1 CPU, 32 logical and 16 physical cores
.NET SDK=7.0.101
[Host] : .NET 7.0.1 (7.0.122.56804), X64 RyuJIT AVX2 [AttachedDebugger]
DefaultJob : .NET 7.0.1 (7.0.122.56804), X64 RyuJIT AVX2
```
### Regression?
The difference is still there on .NET 4.8.1 but less stark because both implementations are very slow compared to .NET 7.0. :)
| Method | Job | Runtime | Mean | Error | StdDev |
|----------------- |--------------------- |--------------------- |---------:|----------:|----------:|
| Linq_ToCharArray | .NET 7.0 | .NET 7.0 | 2.117 ms | 0.0276 ms | 0.0258 ms |
| Linq_String | .NET 7.0 | .NET 7.0 | 5.284 ms | 0.1053 ms | 0.1576 ms |
| Linq_ToCharArray | .NET Framework 4.8.1 | .NET Framework 4.8.1 | 4.095 ms | 0.0239 ms | 0.0224 ms |
| Linq_String | .NET Framework 4.8.1 | .NET Framework 4.8.1 | 6.799 ms | 0.0166 ms | 0.0147 ms |
### Analysis
Linq's string enumeration uses `WhereEnumerableIterator<TSource>`, while `ToCharArray()` version uses `WhereArrayIterator<TSource>` which is optimized for arrays despite that string is perfectly capable of indexed element access and precalculated length. I believe the overhead comes from slow Enumerable-based iteration.
### Solution proposal
I propose a `WhereStringIterator` and an accompanying `WhereSelectStringIterator` classes for strings. It might look like this (based on `WhereArrayIterator`):
```csharp
/// <summary>
/// An iterator that filters each character of a string.
/// </summary>
internal sealed partial class WhereStringIterator : Iterator<char>
{
private readonly string _source;
private readonly Func<char, bool> _predicate;
public WhereStringIterator(string source, Func<char, bool> predicate)
{
Debug.Assert(source != null && source.Length > 0);
Debug.Assert(predicate != null);
_source = source;
_predicate = predicate;
}
public override Iterator<char> Clone() =>
new WhereStringIterator(_source, _predicate);
public override bool MoveNext()
{
int index = _state - 1;
string source = _source;
while (unchecked((uint)index < (uint)source.Length))
{
char item = source[index];
index = _state++;
if (_predicate(item))
{
_current = item;
return true;
}
}
Dispose();
return false;
}
public override IEnumerable<TResult> Select<TResult>(Func<TSource, TResult> selector) =>
new WhereSelectStringIterator<TResult>(_source, _predicate, selector);
public override IEnumerable<TSource> Where(Func<TSource, bool> predicate) =>
new WhereStringIterator(_source, CombinePredicates(_predicate, predicate));
}
```
The catch is that `Enumerable.Where<TSource>()` implementation in `Where.cs` will have an additional type check like this:
```csharp
public static IEnumerable<TSource> Where<TSource>(this IEnumerable<TSource> source, Func<TSource, bool> predicate)
{
// ... irrelevant code removed
// @ssg: these consecutive ifs below can also benefit from being converted
// to a switch expression, not sure if there's any performance difference though.
if (source is Iterator<TSource> iterator)
{
return iterator.Where(predicate);
}
if (source is TSource[] array)
{
return array.Length == 0 ?
Empty<TSource>() :
new WhereArrayIterator<TSource>(array, predicate);
}
if (source is List<TSource> list)
{
return new WhereListIterator<TSource>(list, predicate);
}
// @ssg: This is what we're adding.
if (source is string str)
{
return str.Length == 0 ?
Empty<TSource>() :
new WhereStringIterator(str, predicate);
}
return new WhereEnumerableIterator<TSource>(source, predicate);
}
```
If you can confirm that this is way to go, I can go ahead and try creating a PR for this on System.Linq. It would make string enumeration as fast as array enumeration and would remove the superstition around using `ToCharArray()` unnecessarily.
It might also be worthwhile to investigate why `WhereEnumerableIterator<TSource>` is twice as slow as the array iterator. My guess would be because of the overhead of an extra `MoveNext()` call every iteration instead of just an indexed memory access.
|
True
|
Linq enumeration of a string's characters is twice as slow as enumerating them using string.ToCharArray() - ### Description
Enumerating characters of a string using Linq is slower than doing it on its copy created by `ToCharArray()`. Because of this behavior, people might start seeing `ToCharArray()` as a performance magic (there are already some stackoverflow entries about it), and it might get used everywhere, causing memory overhead unnecessarily.
Given the benchmark below:
```csharp
[MemoryDiagnoser]
public partial class StringVsToCharArrayBenchmark
{
private static string longText = new string('a', 1000000);
[Benchmark]
public bool Linq_ToCharArray()
{
return longText.ToCharArray().Where(c => c == 'a').Count() > 100;
}
[Benchmark]
public bool Linq_String()
{
return longText.Where(c => c == 'a').Count() > 100;
}
static void Main()
{
_ = BenchmarkRunner.Run<StringVsToCharArrayBenchmark>();
}
}
```
The benchmark `Linq_String` is twice as slow as `Linq_ToCharArray` despite that ToCharArray() allocates two strings worth of extra memory. The relative 2x performance difference doesn't change based on the string size. Here is the BenchmarkDotNet output for the given code above:
| Method | Mean | Error | StdDev | Gen0 | Gen1 | Gen2 | Allocated |
|----------------- |---------:|----------:|----------:|---------:|---------:|---------:|----------:|
| Linq_ToCharArray | 1.858 ms | 0.0289 ms | 0.0271 ms | 253.9063 | 253.9063 | 253.9063 | 2000935 B |
| Linq_String | 4.839 ms | 0.0954 ms | 0.0980 ms | - | - | - | 92 B |
(`foreach (var s in string)` is still the fastest of the bunch of course)
### Configuration
```
BenchmarkDotNet=v0.13.3, OS=Windows 11 (10.0.22623.1095)
AMD Ryzen 9 5950X, 1 CPU, 32 logical and 16 physical cores
.NET SDK=7.0.101
[Host] : .NET 7.0.1 (7.0.122.56804), X64 RyuJIT AVX2 [AttachedDebugger]
DefaultJob : .NET 7.0.1 (7.0.122.56804), X64 RyuJIT AVX2
```
### Regression?
The difference is still there on .NET 4.8.1 but less stark because both implementations are very slow compared to .NET 7.0. :)
| Method | Job | Runtime | Mean | Error | StdDev |
|----------------- |--------------------- |--------------------- |---------:|----------:|----------:|
| Linq_ToCharArray | .NET 7.0 | .NET 7.0 | 2.117 ms | 0.0276 ms | 0.0258 ms |
| Linq_String | .NET 7.0 | .NET 7.0 | 5.284 ms | 0.1053 ms | 0.1576 ms |
| Linq_ToCharArray | .NET Framework 4.8.1 | .NET Framework 4.8.1 | 4.095 ms | 0.0239 ms | 0.0224 ms |
| Linq_String | .NET Framework 4.8.1 | .NET Framework 4.8.1 | 6.799 ms | 0.0166 ms | 0.0147 ms |
### Analysis
Linq's string enumeration uses `WhereEnumerableIterator<TSource>`, while `ToCharArray()` version uses `WhereArrayIterator<TSource>` which is optimized for arrays despite that string is perfectly capable of indexed element access and precalculated length. I believe the overhead comes from slow Enumerable-based iteration.
### Solution proposal
I propose a `WhereStringIterator` and an accompanying `WhereSelectStringIterator` classes for strings. It might look like this (based on `WhereArrayIterator`):
```csharp
/// <summary>
/// An iterator that filters each character of a string.
/// </summary>
internal sealed partial class WhereStringIterator : Iterator<char>
{
private readonly string _source;
private readonly Func<char, bool> _predicate;
public WhereStringIterator(string source, Func<char, bool> predicate)
{
Debug.Assert(source != null && source.Length > 0);
Debug.Assert(predicate != null);
_source = source;
_predicate = predicate;
}
public override Iterator<char> Clone() =>
new WhereStringIterator(_source, _predicate);
public override bool MoveNext()
{
int index = _state - 1;
string source = _source;
while (unchecked((uint)index < (uint)source.Length))
{
char item = source[index];
index = _state++;
if (_predicate(item))
{
_current = item;
return true;
}
}
Dispose();
return false;
}
public override IEnumerable<TResult> Select<TResult>(Func<TSource, TResult> selector) =>
new WhereSelectStringIterator<TResult>(_source, _predicate, selector);
public override IEnumerable<TSource> Where(Func<TSource, bool> predicate) =>
new WhereStringIterator(_source, CombinePredicates(_predicate, predicate));
}
```
The catch is that `Enumerable.Where<TSource>()` implementation in `Where.cs` will have an additional type check like this:
```csharp
public static IEnumerable<TSource> Where<TSource>(this IEnumerable<TSource> source, Func<TSource, bool> predicate)
{
// ... irrelevant code removed
// @ssg: these consecutive ifs below can also benefit from being converted
// to a switch expression, not sure if there's any performance difference though.
if (source is Iterator<TSource> iterator)
{
return iterator.Where(predicate);
}
if (source is TSource[] array)
{
return array.Length == 0 ?
Empty<TSource>() :
new WhereArrayIterator<TSource>(array, predicate);
}
if (source is List<TSource> list)
{
return new WhereListIterator<TSource>(list, predicate);
}
// @ssg: This is what we're adding.
if (source is string str)
{
return str.Length == 0 ?
Empty<TSource>() :
new WhereStringIterator(str, predicate);
}
return new WhereEnumerableIterator<TSource>(source, predicate);
}
```
If you can confirm that this is way to go, I can go ahead and try creating a PR for this on System.Linq. It would make string enumeration as fast as array enumeration and would remove the superstition around using `ToCharArray()` unnecessarily.
It might also be worthwhile to investigate why `WhereEnumerableIterator<TSource>` is twice as slow as the array iterator. My guess would be because of the overhead of an extra `MoveNext()` call every iteration instead of just an indexed memory access.
|
non_process
|
linq enumeration of a string s characters is twice as slow as enumerating them using string tochararray description enumerating characters of a string using linq is slower than doing it on its copy created by tochararray because of this behavior people might start seeing tochararray as a performance magic there are already some stackoverflow entries about it and it might get used everywhere causing memory overhead unnecessarily given the benchmark below csharp public partial class stringvstochararraybenchmark private static string longtext new string a public bool linq tochararray return longtext tochararray where c c a count public bool linq string return longtext where c c a count static void main benchmarkrunner run the benchmark linq string is twice as slow as linq tochararray despite that tochararray allocates two strings worth of extra memory the relative performance difference doesn t change based on the string size here is the benchmarkdotnet output for the given code above method mean error stddev allocated linq tochararray ms ms ms b linq string ms ms ms b foreach var s in string is still the fastest of the bunch of course configuration benchmarkdotnet os windows amd ryzen cpu logical and physical cores net sdk net ryujit defaultjob net ryujit regression the difference is still there on net but less stark because both implementations are very slow compared to net method job runtime mean error stddev linq tochararray net net ms ms ms linq string net net ms ms ms linq tochararray net framework net framework ms ms ms linq string net framework net framework ms ms ms analysis linq s string enumeration uses whereenumerableiterator while tochararray version uses wherearrayiterator which is optimized for arrays despite that string is perfectly capable of indexed element access and precalculated length i believe the overhead comes from slow enumerable based iteration solution proposal i propose a wherestringiterator and an accompanying whereselectstringiterator classes for strings it might look like this based on wherearrayiterator csharp an iterator that filters each character of a string internal sealed partial class wherestringiterator iterator private readonly string source private readonly func predicate public wherestringiterator string source func predicate debug assert source null source length debug assert predicate null source source predicate predicate public override iterator clone new wherestringiterator source predicate public override bool movenext int index state string source source while unchecked uint index uint source length char item source index state if predicate item current item return true dispose return false public override ienumerable select func selector new whereselectstringiterator source predicate selector public override ienumerable where func predicate new wherestringiterator source combinepredicates predicate predicate the catch is that enumerable where implementation in where cs will have an additional type check like this csharp public static ienumerable where this ienumerable source func predicate irrelevant code removed ssg these consecutive ifs below can also benefit from being converted to a switch expression not sure if there s any performance difference though if source is iterator iterator return iterator where predicate if source is tsource array return array length empty new wherearrayiterator array predicate if source is list list return new wherelistiterator list predicate ssg this is what we re adding if source is string str return str length empty new wherestringiterator str predicate return new whereenumerableiterator source predicate if you can confirm that this is way to go i can go ahead and try creating a pr for this on system linq it would make string enumeration as fast as array enumeration and would remove the superstition around using tochararray unnecessarily it might also be worthwhile to investigate why whereenumerableiterator is twice as slow as the array iterator my guess would be because of the overhead of an extra movenext call every iteration instead of just an indexed memory access
| 0
|
25,266
| 11,164,769,086
|
IssuesEvent
|
2019-12-27 06:38:31
|
soumya132/java-code
|
https://api.github.com/repos/soumya132/java-code
|
opened
|
CVE-2018-14720 (High) detected in jackson-databind-2.8.1.jar
|
security vulnerability
|
## CVE-2018-14720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/java-code/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/soumya132/java-code/commit/323d775d03827a3c9662d0ba50b640fe7addea5a">323d775d03827a3c9662d0ba50b640fe7addea5a</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow attackers to conduct external XML entity (XXE) attacks by leveraging failure to block unspecified JDK classes from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14720>CVE-2018-14720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14720">https://nvd.nist.gov/vuln/detail/CVE-2018-14720</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-14720 (High) detected in jackson-databind-2.8.1.jar - ## CVE-2018-14720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/java-code/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/soumya132/java-code/commit/323d775d03827a3c9662d0ba50b640fe7addea5a">323d775d03827a3c9662d0ba50b640fe7addea5a</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow attackers to conduct external XML entity (XXE) attacks by leveraging failure to block unspecified JDK classes from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14720>CVE-2018-14720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14720">https://nvd.nist.gov/vuln/detail/CVE-2018-14720</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm java code pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter jersey release jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow attackers to conduct external xml entity xxe attacks by leveraging failure to block unspecified jdk classes from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
19,714
| 26,054,756,728
|
IssuesEvent
|
2022-12-22 23:19:36
|
bitfocus/companion-module-requests
|
https://api.github.com/repos/bitfocus/companion-module-requests
|
closed
|
Ma Lighting gMA3
|
NOT YET PROCESSED
|
On Bitfocus companion I only find ma lighting grandma 2, but does someone have companion module for grandMA 3?
Cheers
|
1.0
|
Ma Lighting gMA3 - On Bitfocus companion I only find ma lighting grandma 2, but does someone have companion module for grandMA 3?
Cheers
|
process
|
ma lighting on bitfocus companion i only find ma lighting grandma but does someone have companion module for grandma cheers
| 1
|
452
| 2,892,965,138
|
IssuesEvent
|
2015-06-15 15:34:55
|
cfpb/hmda-viz-prototype
|
https://api.github.com/repos/cfpb/hmda-viz-prototype
|
closed
|
Jekyll support for disclosure reports
|
Processing
|
Write scripts to power Jekyll
@awolfe76 Can you create a list of the requirements?
|
1.0
|
Jekyll support for disclosure reports - Write scripts to power Jekyll
@awolfe76 Can you create a list of the requirements?
|
process
|
jekyll support for disclosure reports write scripts to power jekyll can you create a list of the requirements
| 1
|
2,724
| 5,612,332,704
|
IssuesEvent
|
2017-04-03 04:27:40
|
alexrj/Slic3r
|
https://api.github.com/repos/alexrj/Slic3r
|
closed
|
Feature request- belt stretch/frame wobble compensation for tall prints
|
Feature request Fixable with post-process script Won't Change
|
My most recent build is a coreXY printer with belt-lifted Z axis. When the bed is loaded, the belts stretch. I have measured the amount of stretch in my machine at about 42 um/kg load. That means that as the print gets heavier, the print layers are going to thicken while getting narrower. If there was a way to enter the belt stretch value, say 42 um/kg in my case, and have slic3r compensate for the belt stretch by adjusting the layer thickness while maintaining constant extruder flow, even large, heavy prints would come out with uniform layer thickness/extrusion, and accurate height.
In tall printers with beds moving in the Z axis, when the bed is high it is closely coupled to the XY stage that tends to throw the printer around as the parts move. As the print gets taller, the bed drops and becomes more decoupled from the wobbling top of the printer, producing artifacts in the print surface. If the speed/acceleration could be adjusted as a function of print height the print artifacts could be minimized. The same could apply to i3 type machines lifting the X axis in Z.
|
1.0
|
Feature request- belt stretch/frame wobble compensation for tall prints - My most recent build is a coreXY printer with belt-lifted Z axis. When the bed is loaded, the belts stretch. I have measured the amount of stretch in my machine at about 42 um/kg load. That means that as the print gets heavier, the print layers are going to thicken while getting narrower. If there was a way to enter the belt stretch value, say 42 um/kg in my case, and have slic3r compensate for the belt stretch by adjusting the layer thickness while maintaining constant extruder flow, even large, heavy prints would come out with uniform layer thickness/extrusion, and accurate height.
In tall printers with beds moving in the Z axis, when the bed is high it is closely coupled to the XY stage that tends to throw the printer around as the parts move. As the print gets taller, the bed drops and becomes more decoupled from the wobbling top of the printer, producing artifacts in the print surface. If the speed/acceleration could be adjusted as a function of print height the print artifacts could be minimized. The same could apply to i3 type machines lifting the X axis in Z.
|
process
|
feature request belt stretch frame wobble compensation for tall prints my most recent build is a corexy printer with belt lifted z axis when the bed is loaded the belts stretch i have measured the amount of stretch in my machine at about um kg load that means that as the print gets heavier the print layers are going to thicken while getting narrower if there was a way to enter the belt stretch value say um kg in my case and have compensate for the belt stretch by adjusting the layer thickness while maintaining constant extruder flow even large heavy prints would come out with uniform layer thickness extrusion and accurate height in tall printers with beds moving in the z axis when the bed is high it is closely coupled to the xy stage that tends to throw the printer around as the parts move as the print gets taller the bed drops and becomes more decoupled from the wobbling top of the printer producing artifacts in the print surface if the speed acceleration could be adjusted as a function of print height the print artifacts could be minimized the same could apply to type machines lifting the x axis in z
| 1
|
127,971
| 10,509,099,144
|
IssuesEvent
|
2019-09-27 10:08:33
|
ampproject/amp-wp
|
https://api.github.com/repos/ampproject/amp-wp
|
closed
|
Error message: "this block contains unexpected or invalid content"
|
AMP Stories Feedback Needs Testing Status: Duplicate
|
Repro steps:
1. I created a story last week
1. My role was changed from author to editor (not sure if this step is required to repro)
1. I opened the story today, and I see the following error messages.

|
1.0
|
Error message: "this block contains unexpected or invalid content" - Repro steps:
1. I created a story last week
1. My role was changed from author to editor (not sure if this step is required to repro)
1. I opened the story today, and I see the following error messages.

|
non_process
|
error message this block contains unexpected or invalid content repro steps i created a story last week my role was changed from author to editor not sure if this step is required to repro i opened the story today and i see the following error messages
| 0
|
267,905
| 20,250,374,068
|
IssuesEvent
|
2022-02-14 17:17:23
|
openfeatureflags/spec
|
https://api.github.com/repos/openfeatureflags/spec
|
opened
|
Add architecture summary
|
documentation
|
There are some architecture sketches which could be added to the repository. It would be nice to do so
|
1.0
|
Add architecture summary - There are some architecture sketches which could be added to the repository. It would be nice to do so
|
non_process
|
add architecture summary there are some architecture sketches which could be added to the repository it would be nice to do so
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.