Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
679,655 | 23,241,035,001 | IssuesEvent | 2022-08-03 15:35:54 | ramp4-pcar4/ramp4-pcar4 | https://api.github.com/repos/ramp4-pcar4/ramp4-pcar4 | closed | Initial Filter gets erased by other filters | effort: small flavour: bug priority: must type: corrective | If a filter has been provided via the config `initialFilteredQuery` property, it will be lost when another filter gets applied (e.g. symbol checkboxes, applying grid filters to the layer).
Change the logic to make this persist unless a specific filter API call is made to remove it. | 1.0 | Initial Filter gets erased by other filters - If a filter has been provided via the config `initialFilteredQuery` property, it will be lost when another filter gets applied (e.g. symbol checkboxes, applying grid filters to the layer).
Change the logic to make this persist unless a specific filter API call is made to remove it. | non_test | initial filter gets erased by other filters if a filter has been provided via the config initialfilteredquery property it will be lost when another filter gets applied e g symbol checkboxes applying grid filters to the layer change the logic to make this persist unless a specific filter api call is made to remove it | 0 |
96,433 | 3,968,356,305 | IssuesEvent | 2016-05-03 19:21:58 | ualbertalib/discovery | https://api.github.com/repos/ualbertalib/discovery | closed | Suppress incomplete temp records | enhancement Low priority question | Not an issue with Blacklight, but possibly something it could help to filter out. Is there any way for Blacklight to suppress incomplete records created by NEOS libraries for purposes internal to the holding library, such as [this one (http://search.library.ualberta.ca/catalog/7027362)?
Since it's meant for UofA use, and we don't create records as brief as this this, I'm thinking that it might be possible to ignore records missing specific key MARC fields. In a large discovery system, they're unwanted noise. | 1.0 | Suppress incomplete temp records - Not an issue with Blacklight, but possibly something it could help to filter out. Is there any way for Blacklight to suppress incomplete records created by NEOS libraries for purposes internal to the holding library, such as [this one (http://search.library.ualberta.ca/catalog/7027362)?
Since it's meant for UofA use, and we don't create records as brief as this this, I'm thinking that it might be possible to ignore records missing specific key MARC fields. In a large discovery system, they're unwanted noise. | non_test | suppress incomplete temp records not an issue with blacklight but possibly something it could help to filter out is there any way for blacklight to suppress incomplete records created by neos libraries for purposes internal to the holding library such as this one since it s meant for uofa use and we don t create records as brief as this this i m thinking that it might be possible to ignore records missing specific key marc fields in a large discovery system they re unwanted noise | 0 |
101,601 | 8,791,282,940 | IssuesEvent | 2018-12-21 12:05:28 | SME-Issues/issues | https://api.github.com/repos/SME-Issues/issues | closed | Compound Query Tests Balance Partial - 21/12/18 11:01 - 5004 | NLP Api PETEDEV pulse_tests | **Compound Query Tests Balance Partial**
- Total: 19
- Passed: 12
- **Pass: 14 (74%)**
- Not Understood: 0
- Error (not understood): 0
- Failed but Understood: 5 (26%)
| 1.0 | Compound Query Tests Balance Partial - 21/12/18 11:01 - 5004 - **Compound Query Tests Balance Partial**
- Total: 19
- Passed: 12
- **Pass: 14 (74%)**
- Not Understood: 0
- Error (not understood): 0
- Failed but Understood: 5 (26%)
| test | compound query tests balance partial compound query tests balance partial total passed pass not understood error not understood failed but understood | 1 |
25,057 | 11,139,023,968 | IssuesEvent | 2019-12-21 01:18:24 | tilda/justlaunch | https://api.github.com/repos/tilda/justlaunch | closed | CVE-2019-12900 (High) detected in bzip2-sys-0.1.7.crate | security vulnerability | ## CVE-2019-12900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bzip2-sys-0.1.7.crate</b></p></summary>
<p>Bindings to libbzip2 for bzip2 compression and decompression exposed as
Reader/Writer streams.
</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/bzip2-sys/0.1.7/download">https://crates.io/api/v1/crates/bzip2-sys/0.1.7/download</a></p>
<p>
Dependency Hierarchy:
- zip-0.5.3.crate (Root Library)
- bzip2-0.3.3.crate
- :x: **bzip2-sys-0.1.7.crate** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tilda/justlaunch/commit/1832f386df5cef40a1e3434353179360867a101c">1832f386df5cef40a1e3434353179360867a101c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
BZ2_decompress in decompress.c in bzip2 through 1.0.6 has an out-of-bounds write when there are many selectors.
<p>Publish Date: 2019-06-19
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12900>CVE-2019-12900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-12900 (High) detected in bzip2-sys-0.1.7.crate - ## CVE-2019-12900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bzip2-sys-0.1.7.crate</b></p></summary>
<p>Bindings to libbzip2 for bzip2 compression and decompression exposed as
Reader/Writer streams.
</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/bzip2-sys/0.1.7/download">https://crates.io/api/v1/crates/bzip2-sys/0.1.7/download</a></p>
<p>
Dependency Hierarchy:
- zip-0.5.3.crate (Root Library)
- bzip2-0.3.3.crate
- :x: **bzip2-sys-0.1.7.crate** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tilda/justlaunch/commit/1832f386df5cef40a1e3434353179360867a101c">1832f386df5cef40a1e3434353179360867a101c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
BZ2_decompress in decompress.c in bzip2 through 1.0.6 has an out-of-bounds write when there are many selectors.
<p>Publish Date: 2019-06-19
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12900>CVE-2019-12900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in sys crate cve high severity vulnerability vulnerable library sys crate bindings to for compression and decompression exposed as reader writer streams library home page a href dependency hierarchy zip crate root library crate x sys crate vulnerable library found in head commit a href vulnerability details decompress in decompress c in through has an out of bounds write when there are many selectors publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
2,119 | 4,846,635,557 | IssuesEvent | 2016-11-10 12:29:36 | Yoast/wordpress-seo | https://api.github.com/repos/Yoast/wordpress-seo | closed | Primary category set to one value, actual slug picks lowest category ID anyway | bug compatibility primary category woocommerce | ### What did you expect to happen?
To set my primary product category to "roses" and see the backed generated slug being set to "roses".

### What happened instead?
Primary category is set to "roses" but the generated slug matches the lowest category ID

### How can we reproduce this behavior?
Install latest Yoast + WPML + WooCommerce on latest Wordpress
I have disabled WPML just to take possible conflicts out of picture, but the behavior stays exactly the same, even switching categories and saving the product after every change.
This issue looks identical to issues #4098 and #4095, but this site is in plain English.
Can you provide a link to a page which shows this issue?
https://www.sendflowersonlinefleurworld.com/s/fathers-day/red-roses-aspidistra-bouquet/
### Technical info
* WordPress version: 4.6.1
* Yoast SEO version: 3.7.1
* WooCommerce version: 2.6.7
* WPML version: 3.5.3.1 | True | Primary category set to one value, actual slug picks lowest category ID anyway - ### What did you expect to happen?
To set my primary product category to "roses" and see the backed generated slug being set to "roses".

### What happened instead?
Primary category is set to "roses" but the generated slug matches the lowest category ID

### How can we reproduce this behavior?
Install latest Yoast + WPML + WooCommerce on latest Wordpress
I have disabled WPML just to take possible conflicts out of picture, but the behavior stays exactly the same, even switching categories and saving the product after every change.
This issue looks identical to issues #4098 and #4095, but this site is in plain English.
Can you provide a link to a page which shows this issue?
https://www.sendflowersonlinefleurworld.com/s/fathers-day/red-roses-aspidistra-bouquet/
### Technical info
* WordPress version: 4.6.1
* Yoast SEO version: 3.7.1
* WooCommerce version: 2.6.7
* WPML version: 3.5.3.1 | non_test | primary category set to one value actual slug picks lowest category id anyway what did you expect to happen to set my primary product category to roses and see the backed generated slug being set to roses what happened instead primary category is set to roses but the generated slug matches the lowest category id how can we reproduce this behavior install latest yoast wpml woocommerce on latest wordpress i have disabled wpml just to take possible conflicts out of picture but the behavior stays exactly the same even switching categories and saving the product after every change this issue looks identical to issues and but this site is in plain english can you provide a link to a page which shows this issue technical info wordpress version yoast seo version woocommerce version wpml version | 0 |
284,218 | 24,583,561,130 | IssuesEvent | 2022-10-13 17:36:45 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | VPN context-menu items should use sentence casing | bug needs-text-change QA/Yes QA/Test-Plan-Specified OS/Desktop feature/vpn | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
VPN context-menu items should use sentence casing
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. install `1.39.87`
2. launch Brave
3. set up VPN
4. context-click on the VPN icon
5. look at the text capitalization
## Actual result:
<!--Please add screenshots if needed-->
<img width="349" alt="Screen Shot 2022-05-02 at 10 12 07 AM" src="https://user-images.githubusercontent.com/387249/166293211-38c3cdbd-a041-457c-9fd9-d2a3e8dfbb9e.png">
## Expected result:
<img width="829" alt="Screen Shot 2022-05-02 at 10 17 41 AM" src="https://user-images.githubusercontent.com/387249/166293939-2230438f-3c0d-46ae-a0af-b49a62ee0abc.png">
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
100%
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.39.87 Chromium: 101.0.4951.41ย (Official Build)ย betaย (x86_64)
-- | --
Revision | 93c720db8323b3ec10d056025ab95c23a31997c9-refs/branch-heads/4951@{#904}
OS | macOS Version 11.6.5 (Build 20G527)
cc @rmcfadden3 @bsclifton @Tonev | 1.0 | VPN context-menu items should use sentence casing - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
VPN context-menu items should use sentence casing
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. install `1.39.87`
2. launch Brave
3. set up VPN
4. context-click on the VPN icon
5. look at the text capitalization
## Actual result:
<!--Please add screenshots if needed-->
<img width="349" alt="Screen Shot 2022-05-02 at 10 12 07 AM" src="https://user-images.githubusercontent.com/387249/166293211-38c3cdbd-a041-457c-9fd9-d2a3e8dfbb9e.png">
## Expected result:
<img width="829" alt="Screen Shot 2022-05-02 at 10 17 41 AM" src="https://user-images.githubusercontent.com/387249/166293939-2230438f-3c0d-46ae-a0af-b49a62ee0abc.png">
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
100%
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.39.87 Chromium: 101.0.4951.41ย (Official Build)ย betaย (x86_64)
-- | --
Revision | 93c720db8323b3ec10d056025ab95c23a31997c9-refs/branch-heads/4951@{#904}
OS | macOS Version 11.6.5 (Build 20G527)
cc @rmcfadden3 @bsclifton @Tonev | test | vpn context menu items should use sentence casing have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description vpn context menu items should use sentence casing steps to reproduce install launch brave set up vpn context click on the vpn icon look at the text capitalization actual result img width alt screen shot at am src expected result img width alt screen shot at am src reproduces how often brave version brave version info brave chromium ย official build ย betaย revision refs branch heads os macos version build cc bsclifton tonev | 1 |
159,359 | 12,474,504,754 | IssuesEvent | 2020-05-29 09:46:25 | aliasrobotics/RVD | https://api.github.com/repos/aliasrobotics/RVD | opened | Use of possibly insecure function - consider using safer ast., /opt/ros_noetic_ws/src/rqt_service_caller/src/rqt_service_caller/service_caller_widget.py:237 | bandit bug static analysis testing triage | ```yaml
{
"id": 1,
"title": "Use of possibly insecure function - consider using safer ast., /opt/ros_noetic_ws/src/rqt_service_caller/src/rqt_service_caller/service_caller_widget.py:237",
"type": "bug",
"description": "HIGH confidence of MEDIUM severity bug. Use of possibly insecure function - consider using safer ast.literal_eval. at /opt/ros_noetic_ws/src/rqt_service_caller/src/rqt_service_caller/service_caller_widget.py:237 See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-29 (09:46)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-29 (09:46)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "/opt/ros_noetic_ws/src/rqt_service_caller/src/rqt_service_caller/service_caller_widget.py:237",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
``` | 1.0 | Use of possibly insecure function - consider using safer ast., /opt/ros_noetic_ws/src/rqt_service_caller/src/rqt_service_caller/service_caller_widget.py:237 - ```yaml
{
"id": 1,
"title": "Use of possibly insecure function - consider using safer ast., /opt/ros_noetic_ws/src/rqt_service_caller/src/rqt_service_caller/service_caller_widget.py:237",
"type": "bug",
"description": "HIGH confidence of MEDIUM severity bug. Use of possibly insecure function - consider using safer ast.literal_eval. at /opt/ros_noetic_ws/src/rqt_service_caller/src/rqt_service_caller/service_caller_widget.py:237 See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-29 (09:46)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-29 (09:46)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "/opt/ros_noetic_ws/src/rqt_service_caller/src/rqt_service_caller/service_caller_widget.py:237",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
``` | test | use of possibly insecure function consider using safer ast opt ros noetic ws src rqt service caller src rqt service caller service caller widget py yaml id title use of possibly insecure function consider using safer ast opt ros noetic ws src rqt service caller src rqt service caller service caller widget py type bug description high confidence of medium severity bug use of possibly insecure function consider using safer ast literal eval at opt ros noetic ws src rqt service caller src rqt service caller service caller widget py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace opt ros noetic ws src rqt service caller src rqt service caller service caller widget py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation | 1 |
26,499 | 4,227,435,160 | IssuesEvent | 2016-07-03 06:30:00 | datanucleus/datanucleus-api-jpa | https://api.github.com/repos/datanucleus/datanucleus-api-jpa | closed | TransactionalVersion in StateManagerImpl is null for versioned entites whitch are fetched eagerly by using relational annotation | testcase_required | Hello.
We use datanucleus 4.2.5 jpa with jcache (on appengine).
Recently we face with a problem of updating relational object in our entities with enabled 2nd level cache (jcache in our case).
We have the relation OneToOne for our entities: User - Address. Both of them have versioned field of Long type. During update we retrieve the User entity which has OneToOne Address entity which is fetched eagerly (default value for OneToOne mapping).
After updating fields of these entities on saving User we get the following exception:
"Caused by: org.datanucleus.exceptions.NucleusException: Object with id "1" in table address has no version set on the object in memory and you want to update it!! Please report this bug to the developers of DataNucleus with a way of reproducing it"
(The version value is set in database.)
After some investigation we found that the problem relates to the class StateManagerImpl the "transactionalVersion" field is not set for eagerly fetched entities.
For User entity "transactionalVersion" is set as expected, but for Address entity - no. So we got the error at the line: 384 of UpdateRequest class in this code:
"if (currentVersion == null)
{
// Somehow the version is not set on this object (not read in ?) so report the bug
String msg = Localiser.msg("052201", op.getInternalObjectId(), table);
NucleusLogger.PERSISTENCE.error(msg);
throw new NucleusException(msg);
}
"
We also found that if we change the fetching type = lazy - the problem disappears.
| 1.0 | TransactionalVersion in StateManagerImpl is null for versioned entites whitch are fetched eagerly by using relational annotation - Hello.
We use datanucleus 4.2.5 jpa with jcache (on appengine).
Recently we face with a problem of updating relational object in our entities with enabled 2nd level cache (jcache in our case).
We have the relation OneToOne for our entities: User - Address. Both of them have versioned field of Long type. During update we retrieve the User entity which has OneToOne Address entity which is fetched eagerly (default value for OneToOne mapping).
After updating fields of these entities on saving User we get the following exception:
"Caused by: org.datanucleus.exceptions.NucleusException: Object with id "1" in table address has no version set on the object in memory and you want to update it!! Please report this bug to the developers of DataNucleus with a way of reproducing it"
(The version value is set in database.)
After some investigation we found that the problem relates to the class StateManagerImpl the "transactionalVersion" field is not set for eagerly fetched entities.
For User entity "transactionalVersion" is set as expected, but for Address entity - no. So we got the error at the line: 384 of UpdateRequest class in this code:
"if (currentVersion == null)
{
// Somehow the version is not set on this object (not read in ?) so report the bug
String msg = Localiser.msg("052201", op.getInternalObjectId(), table);
NucleusLogger.PERSISTENCE.error(msg);
throw new NucleusException(msg);
}
"
We also found that if we change the fetching type = lazy - the problem disappears.
| test | transactionalversion in statemanagerimpl is null for versioned entites whitch are fetched eagerly by using relational annotation hello we use datanucleus jpa with jcache on appengine recently we face with a problem of updating relational object in our entities with enabled level cache jcache in our case we have the relation onetoone for our entities user address both of them have versioned field of long type during update we retrieve the user entity which has onetoone address entity which is fetched eagerly default value for onetoone mapping after updating fields of these entities on saving user we get the following exception caused by org datanucleus exceptions nucleusexception object with id in table address has no version set on the object in memory and you want to update it please report this bug to the developers of datanucleus with a way of reproducing it the version value is set in database after some investigation we found that the problem relates to the class statemanagerimpl the transactionalversion field is not set for eagerly fetched entities for user entity transactionalversion is set as expected but for address entity no so we got the error at the line of updaterequest class in this code if currentversion null somehow the version is not set on this object not read in so report the bug string msg localiser msg op getinternalobjectid table nucleuslogger persistence error msg throw new nucleusexception msg we also found that if we change the fetching type lazy the problem disappears | 1 |
32,126 | 8,797,821,706 | IssuesEvent | 2018-12-24 01:16:19 | dealii/dealii | https://api.github.com/repos/dealii/dealii | closed | Build error on cori/nersc [via spack] | Build system | I'm seeing errors building deallii via spack today.
However the same build worked before [a week back] - suggesting something on the machine - so this is a bit wierd.
Any idea what could trigger this error?
```
[ 77%] Built target obj_fe_release
lac/CMakeFiles/obj_lac_debug.dir/la_parallel_block_vector.cc.o:la_parallel_block_vector.cc:vtable for dealii::LinearAlgebra::distributed::BlockVector<double>: error: relocation refers to global symbol
"non-virtual thunk to dealii::LinearAlgebra::distributed::BlockVector<double>::~BlockVector()", which is defined in a discarded section
section group signature: "_ZN6dealii13LinearAlgebra11distributed11BlockVectorIdED1Ev"
prevailing definition is from numerics/CMakeFiles/obj_numerics_debug.dir/vector_tools_interpolate.cc.o
lac/CMakeFiles/obj_lac_debug.dir/la_parallel_block_vector.cc.o:la_parallel_block_vector.cc:vtable for dealii::LinearAlgebra::distributed::BlockVector<float>: error: relocation refers to global symbol
"non-virtual thunk to dealii::LinearAlgebra::distributed::BlockVector<float>::~BlockVector()", which is defined in a discarded section
section group signature: "_ZN6dealii13LinearAlgebra11distributed11BlockVectorIfED1Ev"
prevailing definition is from numerics/CMakeFiles/obj_numerics_debug.dir/vector_tools_interpolate.cc.o
lac/CMakeFiles/obj_lac_debug.dir/la_parallel_block_vector.cc.o:la_parallel_block_vector.cc:vtable for dealii::LinearAlgebra::distributed::BlockVector<std::complex<double> >: error: relocation refers to
global symbol "non-virtual thunk to dealii::LinearAlgebra::distributed::BlockVector<std::complex<double> >::~BlockVector()", which is defined in a discarded section
section group signature: "_ZN6dealii13LinearAlgebra11distributed11BlockVectorISt7complexIdEED1Ev"
prevailing definition is from numerics/CMakeFiles/obj_numerics_debug.dir/vector_tools_interpolate.cc.o
lac/CMakeFiles/obj_lac_debug.dir/la_parallel_block_vector.cc.o:la_parallel_block_vector.cc:vtable for dealii::LinearAlgebra::distributed::BlockVector<std::complex<float> >: error: relocation refers to global
symbol "non-virtual thunk to dealii::LinearAlgebra::distributed::BlockVector<std::complex<float> >::~BlockVector()", which is defined in a discarded section
section group signature: "_ZN6dealii13LinearAlgebra11distributed11BlockVectorISt7complexIfEED1Ev"
prevailing definition is from numerics/CMakeFiles/obj_numerics_debug.dir/vector_tools_interpolate.cc.o
source/CMakeFiles/deal_II.g.dir/build.make:818: recipe for target 'lib/libdeal_II.g.so.9.0.1' failed
```
[cori-dealii-spack-build-old.out.txt](https://github.com/dealii/dealii/files/2703284/cori-dealii-spack-build-old.out.txt)
[spack-build.env.txt](https://github.com/dealii/dealii/files/2703285/spack-build.env.txt)
| 1.0 | Build error on cori/nersc [via spack] - I'm seeing errors building deallii via spack today.
However the same build worked before [a week back] - suggesting something on the machine - so this is a bit wierd.
Any idea what could trigger this error?
```
[ 77%] Built target obj_fe_release
lac/CMakeFiles/obj_lac_debug.dir/la_parallel_block_vector.cc.o:la_parallel_block_vector.cc:vtable for dealii::LinearAlgebra::distributed::BlockVector<double>: error: relocation refers to global symbol
"non-virtual thunk to dealii::LinearAlgebra::distributed::BlockVector<double>::~BlockVector()", which is defined in a discarded section
section group signature: "_ZN6dealii13LinearAlgebra11distributed11BlockVectorIdED1Ev"
prevailing definition is from numerics/CMakeFiles/obj_numerics_debug.dir/vector_tools_interpolate.cc.o
lac/CMakeFiles/obj_lac_debug.dir/la_parallel_block_vector.cc.o:la_parallel_block_vector.cc:vtable for dealii::LinearAlgebra::distributed::BlockVector<float>: error: relocation refers to global symbol
"non-virtual thunk to dealii::LinearAlgebra::distributed::BlockVector<float>::~BlockVector()", which is defined in a discarded section
section group signature: "_ZN6dealii13LinearAlgebra11distributed11BlockVectorIfED1Ev"
prevailing definition is from numerics/CMakeFiles/obj_numerics_debug.dir/vector_tools_interpolate.cc.o
lac/CMakeFiles/obj_lac_debug.dir/la_parallel_block_vector.cc.o:la_parallel_block_vector.cc:vtable for dealii::LinearAlgebra::distributed::BlockVector<std::complex<double> >: error: relocation refers to
global symbol "non-virtual thunk to dealii::LinearAlgebra::distributed::BlockVector<std::complex<double> >::~BlockVector()", which is defined in a discarded section
section group signature: "_ZN6dealii13LinearAlgebra11distributed11BlockVectorISt7complexIdEED1Ev"
prevailing definition is from numerics/CMakeFiles/obj_numerics_debug.dir/vector_tools_interpolate.cc.o
lac/CMakeFiles/obj_lac_debug.dir/la_parallel_block_vector.cc.o:la_parallel_block_vector.cc:vtable for dealii::LinearAlgebra::distributed::BlockVector<std::complex<float> >: error: relocation refers to global
symbol "non-virtual thunk to dealii::LinearAlgebra::distributed::BlockVector<std::complex<float> >::~BlockVector()", which is defined in a discarded section
section group signature: "_ZN6dealii13LinearAlgebra11distributed11BlockVectorISt7complexIfEED1Ev"
prevailing definition is from numerics/CMakeFiles/obj_numerics_debug.dir/vector_tools_interpolate.cc.o
source/CMakeFiles/deal_II.g.dir/build.make:818: recipe for target 'lib/libdeal_II.g.so.9.0.1' failed
```
[cori-dealii-spack-build-old.out.txt](https://github.com/dealii/dealii/files/2703284/cori-dealii-spack-build-old.out.txt)
[spack-build.env.txt](https://github.com/dealii/dealii/files/2703285/spack-build.env.txt)
| non_test | build error on cori nersc i m seeing errors building deallii via spack today however the same build worked before suggesting something on the machine so this is a bit wierd any idea what could trigger this error built target obj fe release lac cmakefiles obj lac debug dir la parallel block vector cc o la parallel block vector cc vtable for dealii linearalgebra distributed blockvector error relocation refers to global symbol non virtual thunk to dealii linearalgebra distributed blockvector blockvector which is defined in a discarded section section group signature prevailing definition is from numerics cmakefiles obj numerics debug dir vector tools interpolate cc o lac cmakefiles obj lac debug dir la parallel block vector cc o la parallel block vector cc vtable for dealii linearalgebra distributed blockvector error relocation refers to global symbol non virtual thunk to dealii linearalgebra distributed blockvector blockvector which is defined in a discarded section section group signature prevailing definition is from numerics cmakefiles obj numerics debug dir vector tools interpolate cc o lac cmakefiles obj lac debug dir la parallel block vector cc o la parallel block vector cc vtable for dealii linearalgebra distributed blockvector error relocation refers to global symbol non virtual thunk to dealii linearalgebra distributed blockvector blockvector which is defined in a discarded section section group signature prevailing definition is from numerics cmakefiles obj numerics debug dir vector tools interpolate cc o lac cmakefiles obj lac debug dir la parallel block vector cc o la parallel block vector cc vtable for dealii linearalgebra distributed blockvector error relocation refers to global symbol non virtual thunk to dealii linearalgebra distributed blockvector blockvector which is defined in a discarded section section group signature prevailing definition is from numerics cmakefiles obj numerics debug dir vector tools interpolate cc o source cmakefiles deal ii g dir build make recipe for target lib libdeal ii g so failed | 0 |
41,453 | 21,694,547,821 | IssuesEvent | 2022-05-09 18:41:20 | mozilla-mobile/fenix | https://api.github.com/repos/mozilla-mobile/fenix | closed | My site is not performing well | performance needs:triage | Hi, I have done on-page SEO and some off-page SEO of my [nursery](https://gardeninglimited.com/) website but it is not performing well. Suggest me techniques to improve performance of my website. | True | My site is not performing well - Hi, I have done on-page SEO and some off-page SEO of my [nursery](https://gardeninglimited.com/) website but it is not performing well. Suggest me techniques to improve performance of my website. | non_test | my site is not performing well hi i have done on page seo and some off page seo of my website but it is not performing well suggest me techniques to improve performance of my website | 0 |
68,738 | 17,391,450,060 | IssuesEvent | 2021-08-02 07:57:32 | google/mediapipe | https://api.github.com/repos/google/mediapipe | closed | How to "include ./" ? | platform:python stalled stat:awaiting response type:build/install |
```console
โ mediapipe git:(master) โ python setup.py build
running build
running build_binary_graphs
generating binarypb: mediapipe/modules/face_detection/face_detection_short_range_cpu
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'com_google_absl' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'com_google_protobuf' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'com_google_googletest' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'com_github_gflags_gflags' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'bazel_skylib' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'pybind11' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/tf_runtime/third_party/cuda/dependencies.bzl:51:10: The following command will download NVIDIA proprietary software. By using the software you agree to comply with the terms of the license agreement that accompanies the software. If you do not agree to the terms of the license agreement, do not use the software.
INFO: Build options --action_env, --compilation_mode, --copt, and 1 more have changed, discarding analysis cache.
INFO: Analyzed target //mediapipe/modules/face_detection:face_detection_short_range_cpu (124 packages loaded, 6167 targets configured).
INFO: Found 1 target...
ERROR: ....../mediapipe/mediapipe/framework/formats/BUILD:158:11: C++ compilation of rule '//mediapipe/framework/formats:location' failed (Exit 1): gcc failed: error executing command /usr/local/bin/gcc -U_FORTIFY_SOURCE -fstack-protector -Wall -Wunused-but-set-parameter -Wno-free-nonheap-object -fno-omit-frame-pointer -g0 -O2 '-D_FORTIFY_SOURCE=1' -DNDEBUG -ffunction-sections ... (remaining 43 argument(s) skipped)
Use --sandbox_debug to see verbose messages from the sandbox gcc failed: error executing command /usr/local/bin/gcc -U_FORTIFY_SOURCE -fstack-protector -Wall -Wunused-but-set-parameter -Wno-free-nonheap-object -fno-omit-frame-pointer -g0 -O2 '-D_FORTIFY_SOURCE=1' -DNDEBUG -ffunction-sections ... (remaining 43 argument(s) skipped)
Use --sandbox_debug to see verbose messages from the sandbox
In file included from mediapipe/framework/formats/location.cc:15:
./mediapipe/framework/formats/location.h:44:10: fatal error: mediapipe/framework/port/opencv_core_inc.h: No such file or directory
44 | #include "mediapipe/framework/port/opencv_core_inc.h"
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
compilation terminated.
Target //mediapipe/modules/face_detection:face_detection_short_range_cpu failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 231.757s, Critical Path: 220.33s
INFO: 1479 processes: 130 internal, 1349 linux-sandbox.
FAILED: Build did NOT complete successfully
```
| 1.0 | How to "include ./" ? -
```console
โ mediapipe git:(master) โ python setup.py build
running build
running build_binary_graphs
generating binarypb: mediapipe/modules/face_detection/face_detection_short_range_cpu
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'com_google_absl' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'com_google_protobuf' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'com_google_googletest' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'com_github_gflags_gflags' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'bazel_skylib' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/org_tensorflow/third_party/repo.bzl:108:14:
Warning: skipping import of repository 'pybind11' because it already exists.
DEBUG: ~/.cache/bazel/_bazel_lvision/dfcc17eb3fa3a6c44920fcc8986c0699/external/tf_runtime/third_party/cuda/dependencies.bzl:51:10: The following command will download NVIDIA proprietary software. By using the software you agree to comply with the terms of the license agreement that accompanies the software. If you do not agree to the terms of the license agreement, do not use the software.
INFO: Build options --action_env, --compilation_mode, --copt, and 1 more have changed, discarding analysis cache.
INFO: Analyzed target //mediapipe/modules/face_detection:face_detection_short_range_cpu (124 packages loaded, 6167 targets configured).
INFO: Found 1 target...
ERROR: ....../mediapipe/mediapipe/framework/formats/BUILD:158:11: C++ compilation of rule '//mediapipe/framework/formats:location' failed (Exit 1): gcc failed: error executing command /usr/local/bin/gcc -U_FORTIFY_SOURCE -fstack-protector -Wall -Wunused-but-set-parameter -Wno-free-nonheap-object -fno-omit-frame-pointer -g0 -O2 '-D_FORTIFY_SOURCE=1' -DNDEBUG -ffunction-sections ... (remaining 43 argument(s) skipped)
Use --sandbox_debug to see verbose messages from the sandbox gcc failed: error executing command /usr/local/bin/gcc -U_FORTIFY_SOURCE -fstack-protector -Wall -Wunused-but-set-parameter -Wno-free-nonheap-object -fno-omit-frame-pointer -g0 -O2 '-D_FORTIFY_SOURCE=1' -DNDEBUG -ffunction-sections ... (remaining 43 argument(s) skipped)
Use --sandbox_debug to see verbose messages from the sandbox
In file included from mediapipe/framework/formats/location.cc:15:
./mediapipe/framework/formats/location.h:44:10: fatal error: mediapipe/framework/port/opencv_core_inc.h: No such file or directory
44 | #include "mediapipe/framework/port/opencv_core_inc.h"
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
compilation terminated.
Target //mediapipe/modules/face_detection:face_detection_short_range_cpu failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 231.757s, Critical Path: 220.33s
INFO: 1479 processes: 130 internal, 1349 linux-sandbox.
FAILED: Build did NOT complete successfully
```
| non_test | how to include console โ mediapipe git master โ python setup py build running build running build binary graphs generating binarypb mediapipe modules face detection face detection short range cpu debug cache bazel bazel lvision external org tensorflow third party repo bzl warning skipping import of repository com google absl because it already exists debug cache bazel bazel lvision external org tensorflow third party repo bzl warning skipping import of repository com google protobuf because it already exists debug cache bazel bazel lvision external org tensorflow third party repo bzl warning skipping import of repository com google googletest because it already exists debug cache bazel bazel lvision external org tensorflow third party repo bzl warning skipping import of repository com github gflags gflags because it already exists debug cache bazel bazel lvision external org tensorflow third party repo bzl warning skipping import of repository bazel skylib because it already exists debug cache bazel bazel lvision external org tensorflow third party repo bzl warning skipping import of repository because it already exists debug cache bazel bazel lvision external tf runtime third party cuda dependencies bzl the following command will download nvidia proprietary software by using the software you agree to comply with the terms of the license agreement that accompanies the software if you do not agree to the terms of the license agreement do not use the software info build options action env compilation mode copt and more have changed discarding analysis cache info analyzed target mediapipe modules face detection face detection short range cpu packages loaded targets configured info found target error mediapipe mediapipe framework formats build c compilation of rule mediapipe framework formats location failed exit gcc failed error executing command usr local bin gcc u fortify source fstack protector wall wunused but set parameter wno free nonheap object fno omit frame pointer d fortify source dndebug ffunction sections remaining argument s skipped use sandbox debug to see verbose messages from the sandbox gcc failed error executing command usr local bin gcc u fortify source fstack protector wall wunused but set parameter wno free nonheap object fno omit frame pointer d fortify source dndebug ffunction sections remaining argument s skipped use sandbox debug to see verbose messages from the sandbox in file included from mediapipe framework formats location cc mediapipe framework formats location h fatal error mediapipe framework port opencv core inc h no such file or directory include mediapipe framework port opencv core inc h compilation terminated target mediapipe modules face detection face detection short range cpu failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes internal linux sandbox failed build did not complete successfully | 0 |
190,093 | 14,532,984,927 | IssuesEvent | 2020-12-14 23:30:57 | QuantConnect/Lean | https://api.github.com/repos/QuantConnect/Lean | opened | Reduce Travis build logs | bug testing | <!--- This template provides sections for bugs and features. Please delete any irrelevant sections before submitting -->
#### Expected Behavior
<!--- Required. Describe the behavior you expect to see for your case. -->
- Travis build finishes successfully
#### Actual Behavior
<!--- Required. Describe the actual behavior for your case. -->
- Travis build is failing because the log count it too long see https://travis-ci.org/github/QuantConnect/Lean/builds/749648890 **master**
#### Potential Solution
<!--- Optional. Describe any potential solutions and/or thoughts as to what may be causing the difference between expected and actual behavior. -->
- Reviews travis logs and remove/ignore the ones unrequited. For Lean logs this could be achieve using a custom log handler which does not log at all.
#### Reproducing the Problem
<!--- Required for Bugs. Describe how to reproduce the problem. This can be via a failing unit test or a simplified algorithm that reliably demonstrates this issue. -->
N/A
#### System Information
<!--- Required for Bugs. Include any system specific information, such as OS. -->
N/A
#### Checklist
<!--- Confirm that you've provided all the required information. -->
<!--- Required fields --->
- [x] I have completely filled out this template
- [x] I have confirmed that this issue exists on the current `master` branch
- [x] I have confirmed that this is not a duplicate issue by searching [issues](https://github.com/QuantConnect/Lean/issues)
<!--- Required for Bugs, feature request can delete the line below. -->
- [x] I have provided detailed steps to reproduce the issue
<!--- Template inspired by https://github.com/stevemao/github-issue-templates --> | 1.0 | Reduce Travis build logs - <!--- This template provides sections for bugs and features. Please delete any irrelevant sections before submitting -->
#### Expected Behavior
<!--- Required. Describe the behavior you expect to see for your case. -->
- Travis build finishes successfully
#### Actual Behavior
<!--- Required. Describe the actual behavior for your case. -->
- Travis build is failing because the log count it too long see https://travis-ci.org/github/QuantConnect/Lean/builds/749648890 **master**
#### Potential Solution
<!--- Optional. Describe any potential solutions and/or thoughts as to what may be causing the difference between expected and actual behavior. -->
- Reviews travis logs and remove/ignore the ones unrequited. For Lean logs this could be achieve using a custom log handler which does not log at all.
#### Reproducing the Problem
<!--- Required for Bugs. Describe how to reproduce the problem. This can be via a failing unit test or a simplified algorithm that reliably demonstrates this issue. -->
N/A
#### System Information
<!--- Required for Bugs. Include any system specific information, such as OS. -->
N/A
#### Checklist
<!--- Confirm that you've provided all the required information. -->
<!--- Required fields --->
- [x] I have completely filled out this template
- [x] I have confirmed that this issue exists on the current `master` branch
- [x] I have confirmed that this is not a duplicate issue by searching [issues](https://github.com/QuantConnect/Lean/issues)
<!--- Required for Bugs, feature request can delete the line below. -->
- [x] I have provided detailed steps to reproduce the issue
<!--- Template inspired by https://github.com/stevemao/github-issue-templates --> | test | reduce travis build logs expected behavior travis build finishes successfully actual behavior travis build is failing because the log count it too long see master potential solution reviews travis logs and remove ignore the ones unrequited for lean logs this could be achieve using a custom log handler which does not log at all reproducing the problem n a system information n a checklist i have completely filled out this template i have confirmed that this issue exists on the current master branch i have confirmed that this is not a duplicate issue by searching i have provided detailed steps to reproduce the issue | 1 |
9,161 | 2,607,928,699 | IssuesEvent | 2015-02-26 00:25:46 | chrsmithdemos/minify | https://api.github.com/repos/chrsmithdemos/minify | opened | Move to autoloading instead of require_once | auto-migrated Maintainability Priority-Medium Type-Enhancement Usability | ```
Good reasons are given in the PEAR2 standards wiki:
http://wiki.pear.php.net/index.php/PEAR2_Standards#Introduction
http://wiki.pear.php.net/index.php/
PEAR2_Standards#use_of_include.2Frequire.2Frequire_once.2Finclude_once_not_allow
ed
Steps:
* copy trunk to branch "autoload"
* remove all require* used to load classes
* create Minify/autoload.php
* register function in min/index.php
* ab test comparison to trunk
* fix unit tests/extras
* merge to trunk
```
-----
Original issue reported on code.google.com by `mrclay....@gmail.com` on 7 Oct 2008 at 6:52 | 1.0 | Move to autoloading instead of require_once - ```
Good reasons are given in the PEAR2 standards wiki:
http://wiki.pear.php.net/index.php/PEAR2_Standards#Introduction
http://wiki.pear.php.net/index.php/
PEAR2_Standards#use_of_include.2Frequire.2Frequire_once.2Finclude_once_not_allow
ed
Steps:
* copy trunk to branch "autoload"
* remove all require* used to load classes
* create Minify/autoload.php
* register function in min/index.php
* ab test comparison to trunk
* fix unit tests/extras
* merge to trunk
```
-----
Original issue reported on code.google.com by `mrclay....@gmail.com` on 7 Oct 2008 at 6:52 | non_test | move to autoloading instead of require once good reasons are given in the standards wiki standards use of include once once not allow ed steps copy trunk to branch autoload remove all require used to load classes create minify autoload php register function in min index php ab test comparison to trunk fix unit tests extras merge to trunk original issue reported on code google com by mrclay gmail com on oct at | 0 |
80,880 | 10,064,901,855 | IssuesEvent | 2019-07-23 09:39:10 | horizontalsystems/unstoppable-wallet-ios | https://api.github.com/repos/horizontalsystems/unstoppable-wallet-ios | closed | Finalize Manage Keys | design enhancement | - [x] add text block between page title and keys. Text "Anyone with access to your Keys can steal your wallet funds. Store them safely and never show them to anyone. [2 line breaks] Backup each key to be able to restore funds if you loose or break your phone, or simply the wallet app gets deleted." (@Dianaaiym the text should be localized.)
- [x] make sure that text has two line breaks between the sentences.
- [x] remove word 'Key' from key headers. Should look like "Binance Chain", "EOS Blockchain" etc
- [x] rename Import to Restore
- [x] rename New to Create
- [ ] @Dianaaiym when doing localizations make sure that the wording on this controller is translated as close as possible to the original. this is the most important controller in whole app.
- [x] create button on EOS. The "create" action for EOS is not yet implemented. Therefore, when clicked the user should see following alert "At this point, the wallet is not able to create new EOS wallet accounts. You're only able to restore the existing one." @Dianaaiym the text needs to be localized
| 1.0 | Finalize Manage Keys - - [x] add text block between page title and keys. Text "Anyone with access to your Keys can steal your wallet funds. Store them safely and never show them to anyone. [2 line breaks] Backup each key to be able to restore funds if you loose or break your phone, or simply the wallet app gets deleted." (@Dianaaiym the text should be localized.)
- [x] make sure that text has two line breaks between the sentences.
- [x] remove word 'Key' from key headers. Should look like "Binance Chain", "EOS Blockchain" etc
- [x] rename Import to Restore
- [x] rename New to Create
- [ ] @Dianaaiym when doing localizations make sure that the wording on this controller is translated as close as possible to the original. this is the most important controller in whole app.
- [x] create button on EOS. The "create" action for EOS is not yet implemented. Therefore, when clicked the user should see following alert "At this point, the wallet is not able to create new EOS wallet accounts. You're only able to restore the existing one." @Dianaaiym the text needs to be localized
| non_test | finalize manage keys add text block between page title and keys text anyone with access to your keys can steal your wallet funds store them safely and never show them to anyone backup each key to be able to restore funds if you loose or break your phone or simply the wallet app gets deleted dianaaiym the text should be localized make sure that text has two line breaks between the sentences remove word key from key headers should look like binance chain eos blockchain etc rename import to restore rename new to create dianaaiym when doing localizations make sure that the wording on this controller is translated as close as possible to the original this is the most important controller in whole app create button on eos the create action for eos is not yet implemented therefore when clicked the user should see following alert at this point the wallet is not able to create new eos wallet accounts you re only able to restore the existing one dianaaiym the text needs to be localized | 0 |
340,870 | 30,550,448,983 | IssuesEvent | 2023-07-20 08:09:44 | IDgis/PlanoView2-bevindingen | https://api.github.com/repos/IDgis/PlanoView2-bevindingen | closed | Waarschuwing geven dat er niets is gevonden als er buiten het regelingsgebied in de kaart is geklikt | verbetering klaar om te testen | Het is verwarrend dat er niets wordt gevonden als er gezocht wordt op woorden in de tekst als er ook in de kaart buiten het regelingsgebied is geklikt. Daarom een waarschuwing geven als er buiten het/de regelingsgebied(en) is geklikt. | 1.0 | Waarschuwing geven dat er niets is gevonden als er buiten het regelingsgebied in de kaart is geklikt - Het is verwarrend dat er niets wordt gevonden als er gezocht wordt op woorden in de tekst als er ook in de kaart buiten het regelingsgebied is geklikt. Daarom een waarschuwing geven als er buiten het/de regelingsgebied(en) is geklikt. | test | waarschuwing geven dat er niets is gevonden als er buiten het regelingsgebied in de kaart is geklikt het is verwarrend dat er niets wordt gevonden als er gezocht wordt op woorden in de tekst als er ook in de kaart buiten het regelingsgebied is geklikt daarom een waarschuwing geven als er buiten het de regelingsgebied en is geklikt | 1 |
286,127 | 21,564,281,389 | IssuesEvent | 2022-05-01 16:29:34 | LucPrestin/Hidden-Modularity | https://api.github.com/repos/LucPrestin/Hidden-Modularity | closed | Think about entry points | documentation | As someone looking into the hidden modularity, I'd like to know a few entry points from which I could start my research.
---
**Describe the solution you'd like**
For each of the following rough parts of squeak/smalltalk, there should be at least one entry point documented:
- User Input
- Graphics Output
- Network Access
---
## Conditions of satisfaction
- [x] Each of the rough parts has at least one entry point
- [x] There is a wiki page documenting the entry points
- [x] For each entry point there is an extra issue that is at least a rough idea on how to tackle it
| 1.0 | Think about entry points - As someone looking into the hidden modularity, I'd like to know a few entry points from which I could start my research.
---
**Describe the solution you'd like**
For each of the following rough parts of squeak/smalltalk, there should be at least one entry point documented:
- User Input
- Graphics Output
- Network Access
---
## Conditions of satisfaction
- [x] Each of the rough parts has at least one entry point
- [x] There is a wiki page documenting the entry points
- [x] For each entry point there is an extra issue that is at least a rough idea on how to tackle it
| non_test | think about entry points as someone looking into the hidden modularity i d like to know a few entry points from which i could start my research describe the solution you d like for each of the following rough parts of squeak smalltalk there should be at least one entry point documented user input graphics output network access conditions of satisfaction each of the rough parts has at least one entry point there is a wiki page documenting the entry points for each entry point there is an extra issue that is at least a rough idea on how to tackle it | 0 |
21,271 | 6,132,582,625 | IssuesEvent | 2017-06-25 04:08:04 | ganeti/ganeti | https://api.github.com/repos/ganeti/ganeti | closed | htest/test fails to build when confd is disabled | imported_from_google_code Priority-Low SmallTask Status:Obsolete Type-Defect | Originally reported of Google Code with ID 332.
```
Building htest/test with โ./configure --disable-confdโ fails in a very non-obvious way. Either the target inMakefile should already fail with an error message, or at least the error generated by ghc should be more descriptive.
$ git describe
v2.6.1-899-g78fcfd4
$ ./autogen.sh && ./configure --prefix=/usr/local --sysconfdir=/etc --localstatedir=/var --enable-restricted-commands --disable-split-query --disable-confd && make -j2 htest/test
[โฆ]
htools/Ganeti/Query/Filter.hs:128:9:
No instance for (PCRE.RegexLike () [Char])
arising from a use of `PCRE.match'
Possible fix:
add an instance declaration for (PCRE.RegexLike () [Char])
In the second argument of `($!)', namely
`PCRE.match (compiledRegex re) (fromJSString val)'
In the expression:
Ok $! PCRE.match (compiledRegex re) (fromJSString val)
In an equation for `regexpFilter':
regexpFilter re (JSString val)
= Ok $! PCRE.match (compiledRegex re) (fromJSString val)
```
Originally added on 2012-12-10 15:04:37 +0000 UTC. | 1.0 | htest/test fails to build when confd is disabled - Originally reported of Google Code with ID 332.
```
Building htest/test with โ./configure --disable-confdโ fails in a very non-obvious way. Either the target inMakefile should already fail with an error message, or at least the error generated by ghc should be more descriptive.
$ git describe
v2.6.1-899-g78fcfd4
$ ./autogen.sh && ./configure --prefix=/usr/local --sysconfdir=/etc --localstatedir=/var --enable-restricted-commands --disable-split-query --disable-confd && make -j2 htest/test
[โฆ]
htools/Ganeti/Query/Filter.hs:128:9:
No instance for (PCRE.RegexLike () [Char])
arising from a use of `PCRE.match'
Possible fix:
add an instance declaration for (PCRE.RegexLike () [Char])
In the second argument of `($!)', namely
`PCRE.match (compiledRegex re) (fromJSString val)'
In the expression:
Ok $! PCRE.match (compiledRegex re) (fromJSString val)
In an equation for `regexpFilter':
regexpFilter re (JSString val)
= Ok $! PCRE.match (compiledRegex re) (fromJSString val)
```
Originally added on 2012-12-10 15:04:37 +0000 UTC. | non_test | htest test fails to build when confd is disabled originally reported of google code with id building htest test with โ configure disable confdโ fails in a very non obvious way either the target inmakefile should already fail with an error message or at least the error generated by ghc should be more descriptive git describe autogen sh configure prefix usr local sysconfdir etc localstatedir var enable restricted commands disable split query disable confd make htest test htools ganeti query filter hs no instance for pcre regexlike arising from a use of pcre match possible fix add an instance declaration for pcre regexlike in the second argument of namely pcre match compiledregex re fromjsstring val in the expression ok pcre match compiledregex re fromjsstring val in an equation for regexpfilter regexpfilter re jsstring val ok pcre match compiledregex re fromjsstring val originally added on utc | 0 |
258,753 | 22,345,491,116 | IssuesEvent | 2022-06-15 07:24:43 | gbowne1/taotnpwaft | https://api.github.com/repos/gbowne1/taotnpwaft | closed | [TODO] Website - Chat | bug enhancement help wanted question discussion dependencies feature request good second issue good intermediate issue good third issue Good advanced issue needs testing website bug | TODO - Website
Complete the website/chat functionality.
Not a high priority item, but at least get it ready for testing by website 1.0.0 milestone
Stub code is in /website/chat of this repository.
My thoughts are not to have chat outside of Discord till we have added 100 acrltive members.
| 1.0 | [TODO] Website - Chat - TODO - Website
Complete the website/chat functionality.
Not a high priority item, but at least get it ready for testing by website 1.0.0 milestone
Stub code is in /website/chat of this repository.
My thoughts are not to have chat outside of Discord till we have added 100 acrltive members.
| test | website chat todo website complete the website chat functionality not a high priority item but at least get it ready for testing by website milestone stub code is in website chat of this repository my thoughts are not to have chat outside of discord till we have added acrltive members | 1 |
101,082 | 8,773,934,105 | IssuesEvent | 2018-12-18 18:17:09 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Allow contributions to non-verified sites setting is not being respected | QA/Test-Plan-Specified QA/Yes bug feature/rewards priority/P3 | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
If I uncheck the 'Allow contributions to non-verified sites' setting, non-verified sites are added to the table with 0% attention until a verified site is added. Then all percentage values get recalculated.
If 'Allow contributions to non-verified sites' is unchecked, non-verified sites should not be added to Auto-Contribute table.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile, enable Rewards.
2. Open Auto-Contribute Settings.
3. Uncheck 'Allow contribution to non-verified sites'
4. Close Auto-Contribute Settings.
5. Open a new tab and visit a non-verified site (such as ecosia.org). Wait 10s or so.
6. Go back to Rewards page, site is added with 0% attention.
7. Open a new tab and visit a verified site (such as duckduckgo.com). Wait 10s or so.
8. Go back to Rewards page, site is added and all percentage values for Attention have been recalculated.
9. If you continue to add non-verified sites at this point, they will get added with 0% attention until a verified site is visited, at which point Attention will be recalculated for all sites in the list.
## Actual result:

## Expected result:
If 'Allow contributions to non-verified sites' is unchecked, non-verified sites should not be added to the Auto-Contribute table. This is how this setting worked in Brave Payments.
## Reproduces how often:
easily
## Brave version (brave://version info)
Brave | 0.56.12 Chromium: 70.0.3538.77ย (Official Build)ย (64-bit)
-- | --
Revision | 0f6ce0b0cd63a12cb4eccea3637b1bc9a29148d9-refs/branch-heads/3538@{#1039}
OS | Mac OS X
Brave | 0.57.6 Chromium: 71.0.3578.31ย (Official Build)ย beta(64-bit)
-- | --
Revision | c88fdf2a4ce19a713615ca4fbde7a0d0b5fe2363-refs/branch-heads/3578@{#427}
OS | Mac OS X
### Reproducible on current release:
- Does it reproduce on brave-browser dev/beta builds? yes reproduced on beta
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
### Additional Information
cc @brave/legacy_qa to check on other platforms
| 1.0 | Allow contributions to non-verified sites setting is not being respected - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
If I uncheck the 'Allow contributions to non-verified sites' setting, non-verified sites are added to the table with 0% attention until a verified site is added. Then all percentage values get recalculated.
If 'Allow contributions to non-verified sites' is unchecked, non-verified sites should not be added to Auto-Contribute table.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile, enable Rewards.
2. Open Auto-Contribute Settings.
3. Uncheck 'Allow contribution to non-verified sites'
4. Close Auto-Contribute Settings.
5. Open a new tab and visit a non-verified site (such as ecosia.org). Wait 10s or so.
6. Go back to Rewards page, site is added with 0% attention.
7. Open a new tab and visit a verified site (such as duckduckgo.com). Wait 10s or so.
8. Go back to Rewards page, site is added and all percentage values for Attention have been recalculated.
9. If you continue to add non-verified sites at this point, they will get added with 0% attention until a verified site is visited, at which point Attention will be recalculated for all sites in the list.
## Actual result:

## Expected result:
If 'Allow contributions to non-verified sites' is unchecked, non-verified sites should not be added to the Auto-Contribute table. This is how this setting worked in Brave Payments.
## Reproduces how often:
easily
## Brave version (brave://version info)
Brave | 0.56.12 Chromium: 70.0.3538.77ย (Official Build)ย (64-bit)
-- | --
Revision | 0f6ce0b0cd63a12cb4eccea3637b1bc9a29148d9-refs/branch-heads/3538@{#1039}
OS | Mac OS X
Brave | 0.57.6 Chromium: 71.0.3578.31ย (Official Build)ย beta(64-bit)
-- | --
Revision | c88fdf2a4ce19a713615ca4fbde7a0d0b5fe2363-refs/branch-heads/3578@{#427}
OS | Mac OS X
### Reproducible on current release:
- Does it reproduce on brave-browser dev/beta builds? yes reproduced on beta
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
### Additional Information
cc @brave/legacy_qa to check on other platforms
| test | allow contributions to non verified sites setting is not being respected have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description if i uncheck the allow contributions to non verified sites setting non verified sites are added to the table with attention until a verified site is added then all percentage values get recalculated if allow contributions to non verified sites is unchecked non verified sites should not be added to auto contribute table steps to reproduce clean profile enable rewards open auto contribute settings uncheck allow contribution to non verified sites close auto contribute settings open a new tab and visit a non verified site such as ecosia org wait or so go back to rewards page site is added with attention open a new tab and visit a verified site such as duckduckgo com wait or so go back to rewards page site is added and all percentage values for attention have been recalculated if you continue to add non verified sites at this point they will get added with attention until a verified site is visited at which point attention will be recalculated for all sites in the list actual result expected result if allow contributions to non verified sites is unchecked non verified sites should not be added to the auto contribute table this is how this setting worked in brave payments reproduces how often easily brave version brave version info brave chromium ย official build ย bit revision refs branch heads os mac os x brave chromium ย official build ย beta bit revision refs branch heads os mac os x reproducible on current release does it reproduce on brave browser dev beta builds yes reproduced on beta website problems only does the issue resolve itself when disabling brave shields n a is the issue reproducible on the latest version of chrome n a additional information cc brave legacy qa to check on other platforms | 1 |
749,195 | 26,153,224,009 | IssuesEvent | 2022-12-30 17:05:40 | w3c/epubcheck | https://api.github.com/repos/w3c/epubcheck | closed | Message OPF-054 lists duplicate source content | type: improvement status: wontfix priority: low | ```
ERROR | OPF-054 | /OEBPS/content.opf (line 10, col 35)
Date value 'May 21, 2011' is not valid as per http://www.w3.org/TR/NOTE-datetime:[For input string: "May 21, 2011"] is not an integer.
```
Source message is:
```
Date value '%1$s' is not valid as per http://www.w3.org/TR/NOTE-datetime:%2$s.
```
The following would be a better message:
```
Date value 'May 21, 2011' is not valid as per http://www.w3.org/TR/NOTE-datetime: is not an integer.
```
Or even better if it's doable to strip the parts in brackets:
```
Date value 'May 21, 2011' is not an integer. Invalid date as per http://www.w3.org/TR/NOTE-datetime
Date value '%1$s' %2$s. Invalid as per http://www.w3.org/TR/NOTE-datetime
```
Bot for that, other errors for OPF-054 must be checked first... | 1.0 | Message OPF-054 lists duplicate source content - ```
ERROR | OPF-054 | /OEBPS/content.opf (line 10, col 35)
Date value 'May 21, 2011' is not valid as per http://www.w3.org/TR/NOTE-datetime:[For input string: "May 21, 2011"] is not an integer.
```
Source message is:
```
Date value '%1$s' is not valid as per http://www.w3.org/TR/NOTE-datetime:%2$s.
```
The following would be a better message:
```
Date value 'May 21, 2011' is not valid as per http://www.w3.org/TR/NOTE-datetime: is not an integer.
```
Or even better if it's doable to strip the parts in brackets:
```
Date value 'May 21, 2011' is not an integer. Invalid date as per http://www.w3.org/TR/NOTE-datetime
Date value '%1$s' %2$s. Invalid as per http://www.w3.org/TR/NOTE-datetime
```
Bot for that, other errors for OPF-054 must be checked first... | non_test | message opf lists duplicate source content error opf oebps content opf line col date value may is not valid as per is not an integer source message is date value s is not valid as per the following would be a better message date value may is not valid as per is not an integer or even better if it s doable to strip the parts in brackets date value may is not an integer invalid date as per date value s s invalid as per bot for that other errors for opf must be checked first | 0 |
237,682 | 19,665,842,484 | IssuesEvent | 2022-01-10 22:23:39 | boltlabs-inc/libzkchannels-crypto | https://api.github.com/repos/boltlabs-inc/libzkchannels-crypto | closed | Compare FindoraNetwork implementation to ours | testing | Is it well written?
Can we use it as a reference implementation?
Do they have any testing techniques we can use? | 1.0 | Compare FindoraNetwork implementation to ours - Is it well written?
Can we use it as a reference implementation?
Do they have any testing techniques we can use? | test | compare findoranetwork implementation to ours is it well written can we use it as a reference implementation do they have any testing techniques we can use | 1 |
494,621 | 14,261,767,744 | IssuesEvent | 2020-11-20 11:52:31 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | PHP Notice Undefined index SERVER_NAME in BuddyBoss products (Platform, Platform Pro, Theme) | bug priority: medium | **Describe the bug**
With DEBUG = true, I do receive some PHP Notice related to SERVER_NAME. Because I use Virtual Host, I haven't set the SERVER_NAME in the apache configuration file. Maybe I'm wrong, but that config is not required and I suppose others could have the same issue.
`[20-Nov-2020 11:35:02 UTC] PHP Notice: Undefined index: SERVER_NAME in <MyServerPath>/htdocs/wp-content/themes/buddyboss-theme/inc/admin/admin-init.php on line 18`
When I do search in the code of BuddyBoss Products (Platform, Platform Pro and Theme), I see that you use $_SERVER['SERVER_NAME'] at some places in the code.
For a security standpoint, I can read on the page https://www.php.net/manual/en/reserved.variables.server.php :
> Note: Under Apache 2, you must set UseCanonicalName = On and ServerName. Otherwise, this value reflects the hostname supplied by the client, which can be spoofed. It is not safe to rely on this value in security-dependent contexts.
Maybe you should use another variable instead ?
**To Reproduce**
1. Server which uses virtual hosts
**Expected behavior**
Check that SERVER_NAME is defined. Use another variable instead.
**Screenshots**
none
**Support ticket links**
none
| 1.0 | PHP Notice Undefined index SERVER_NAME in BuddyBoss products (Platform, Platform Pro, Theme) - **Describe the bug**
With DEBUG = true, I do receive some PHP Notice related to SERVER_NAME. Because I use Virtual Host, I haven't set the SERVER_NAME in the apache configuration file. Maybe I'm wrong, but that config is not required and I suppose others could have the same issue.
`[20-Nov-2020 11:35:02 UTC] PHP Notice: Undefined index: SERVER_NAME in <MyServerPath>/htdocs/wp-content/themes/buddyboss-theme/inc/admin/admin-init.php on line 18`
When I do search in the code of BuddyBoss Products (Platform, Platform Pro and Theme), I see that you use $_SERVER['SERVER_NAME'] at some places in the code.
For a security standpoint, I can read on the page https://www.php.net/manual/en/reserved.variables.server.php :
> Note: Under Apache 2, you must set UseCanonicalName = On and ServerName. Otherwise, this value reflects the hostname supplied by the client, which can be spoofed. It is not safe to rely on this value in security-dependent contexts.
Maybe you should use another variable instead ?
**To Reproduce**
1. Server which uses virtual hosts
**Expected behavior**
Check that SERVER_NAME is defined. Use another variable instead.
**Screenshots**
none
**Support ticket links**
none
| non_test | php notice undefined index server name in buddyboss products platform platform pro theme describe the bug with debug true i do receive some php notice related to server name because i use virtual host i haven t set the server name in the apache configuration file maybe i m wrong but that config is not required and i suppose others could have the same issue php notice undefined index server name in htdocs wp content themes buddyboss theme inc admin admin init php on line when i do search in the code of buddyboss products platform platform pro and theme i see that you use server at some places in the code for a security standpoint i can read on the page note under apache you must set usecanonicalname on and servername otherwise this value reflects the hostname supplied by the client which can be spoofed it is not safe to rely on this value in security dependent contexts maybe you should use another variable instead to reproduce server which uses virtual hosts expected behavior check that server name is defined use another variable instead screenshots none support ticket links none | 0 |
74,117 | 7,375,272,658 | IssuesEvent | 2018-03-13 23:33:19 | SpongePowered/SpongeForge | https://api.github.com/repos/SpongePowered/SpongeForge | closed | Item duping with Integrated Dynamics | status: needs testing system: event system: inventory system: phase tracker version: 1.12 | *Note: this issue was originally reported by @xILikepuddingx in CyclopsMC/IntegratedDynamics#560, so if additional information is required, @xILikepuddingx will be able to help*
**I am currently running**
- SpongeForge version: v1.12.2-2611-7.1.0-BETA-2931
- Forge version: 14.23.2.2619
- Java version: 1.8 (?)
- Operating System: ?
- Plugins/Mods:
Unknown
**Issue Description**
When SpongeForge is installed, some Integrated Dynamics blocks are duped. It looks like that when ID parts are broken, they are dropped multiple times.
Video: https://files.pudding.tk/wgh6cl2lvq.mp4
Original issue: https://github.com/CyclopsMC/IntegratedDynamics/issues/560
_Note that ID uses a custom multipart system. I'll investigate this issue on my end, but reporting this here in any case, as you may be able to help tracking down the cause._
| 1.0 | Item duping with Integrated Dynamics - *Note: this issue was originally reported by @xILikepuddingx in CyclopsMC/IntegratedDynamics#560, so if additional information is required, @xILikepuddingx will be able to help*
**I am currently running**
- SpongeForge version: v1.12.2-2611-7.1.0-BETA-2931
- Forge version: 14.23.2.2619
- Java version: 1.8 (?)
- Operating System: ?
- Plugins/Mods:
Unknown
**Issue Description**
When SpongeForge is installed, some Integrated Dynamics blocks are duped. It looks like that when ID parts are broken, they are dropped multiple times.
Video: https://files.pudding.tk/wgh6cl2lvq.mp4
Original issue: https://github.com/CyclopsMC/IntegratedDynamics/issues/560
_Note that ID uses a custom multipart system. I'll investigate this issue on my end, but reporting this here in any case, as you may be able to help tracking down the cause._
| test | item duping with integrated dynamics note this issue was originally reported by xilikepuddingx in cyclopsmc integrateddynamics so if additional information is required xilikepuddingx will be able to help i am currently running spongeforge version beta forge version java version operating system plugins mods unknown issue description when spongeforge is installed some integrated dynamics blocks are duped it looks like that when id parts are broken they are dropped multiple times video original issue note that id uses a custom multipart system i ll investigate this issue on my end but reporting this here in any case as you may be able to help tracking down the cause | 1 |
264,756 | 8,319,278,439 | IssuesEvent | 2018-09-25 16:46:03 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.esic.in - desktop site instead of mobile site | browser-firefox priority-normal | <!-- @browser: Firefox 63.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:63.0) Gecko/20100101 Firefox/63.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.esic.in/ESICInsurance1/RevenueOne/Monthly%20Contribution/OnlineMonthlyContribution.aspx?ContributionPeriod=aj/07AZFgv8Xbovpy64c4w==&EmployerCode=rK2ozUC9vMNsVF6p1/6/jCX82zGIzJua&ContributionType=JUGm1Rbj4gM=
**Browser / Version**: Firefox 63.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: no. of days of wages is not fill
**Steps to Reproduce**:
monthl;y contribution
[](https://webcompat.com/uploads/2018/9/7a2705ba-4179-4aa1-8c6c-3fde83705a7e.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>buildID: 20180920135444</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | 1.0 | www.esic.in - desktop site instead of mobile site - <!-- @browser: Firefox 63.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:63.0) Gecko/20100101 Firefox/63.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.esic.in/ESICInsurance1/RevenueOne/Monthly%20Contribution/OnlineMonthlyContribution.aspx?ContributionPeriod=aj/07AZFgv8Xbovpy64c4w==&EmployerCode=rK2ozUC9vMNsVF6p1/6/jCX82zGIzJua&ContributionType=JUGm1Rbj4gM=
**Browser / Version**: Firefox 63.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: no. of days of wages is not fill
**Steps to Reproduce**:
monthl;y contribution
[](https://webcompat.com/uploads/2018/9/7a2705ba-4179-4aa1-8c6c-3fde83705a7e.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>buildID: 20180920135444</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | non_test | desktop site instead of mobile site url browser version firefox operating system windows tested another browser yes problem type desktop site instead of mobile site description no of days of wages is not fill steps to reproduce monthl y contribution browser configuration mixed active content blocked false buildid tracking content blocked false gfx webrender blob images true gfx webrender all false mixed passive content blocked false gfx webrender enabled false image mem shared true channel beta from with โค๏ธ | 0 |
86,563 | 15,755,680,864 | IssuesEvent | 2021-03-31 02:12:31 | SmartBear/ready-aws-plugin | https://api.github.com/repos/SmartBear/ready-aws-plugin | opened | CVE-2018-14719 (High) detected in jackson-databind-2.3.0.jar | security vulnerability | ## CVE-2018-14719 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.3.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: ready-aws-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.3.0/jackson-databind-2.3.0.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.3.0.jar (Root Library)
- keen-client-api-java-2.0.2.jar
- :x: **jackson-databind-2.3.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze-ds-opt and blaze-ds-core classes from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14719>CVE-2018-14719</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14719">https://nvd.nist.gov/vuln/detail/CVE-2018-14719</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.3.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.3.0;io.keen:keen-client-api-java:2.0.2;com.fasterxml.jackson.core:jackson-databind:2.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-14719","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze-ds-opt and blaze-ds-core classes from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14719","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-14719 (High) detected in jackson-databind-2.3.0.jar - ## CVE-2018-14719 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.3.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: ready-aws-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.3.0/jackson-databind-2.3.0.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.3.0.jar (Root Library)
- keen-client-api-java-2.0.2.jar
- :x: **jackson-databind-2.3.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze-ds-opt and blaze-ds-core classes from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14719>CVE-2018-14719</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14719">https://nvd.nist.gov/vuln/detail/CVE-2018-14719</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.3.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.3.0;io.keen:keen-client-api-java:2.0.2;com.fasterxml.jackson.core:jackson-databind:2.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-14719","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze-ds-opt and blaze-ds-core classes from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14719","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file ready aws plugin pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy ready api soapui pro jar root library keen client api java jar x jackson databind jar vulnerable library found in base branch master vulnerability details fasterxml jackson databind x before might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze ds opt and blaze ds core classes from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com smartbear ready api soapui pro io keen keen client api java com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze ds opt and blaze ds core classes from polymorphic deserialization vulnerabilityurl | 0 |
22,332 | 3,953,629,443 | IssuesEvent | 2016-04-29 14:11:09 | metafizzy/isotope | https://api.github.com/repos/metafizzy/isotope | closed | Weird white space on initial lode in mobile (one col) view | test case required | I've attached some screen shots here. There is a HUGE amount of white space after each item only when you first lode in mobile mode, which is one column. if you filter, or resize your screen or load more everything goes back to normal.
What's causing this issue and how do we fix it?
http://snag.gy/xhrK9.jpg - see the styles in the inspector and the white space
http://snag.gy/mv77H.jpg - correct display achieved after i resized the viewport.
You can see it in action here - http://rush-health-wellness.centreteksolutions.net/health-wellness/discover-health | 1.0 | Weird white space on initial lode in mobile (one col) view - I've attached some screen shots here. There is a HUGE amount of white space after each item only when you first lode in mobile mode, which is one column. if you filter, or resize your screen or load more everything goes back to normal.
What's causing this issue and how do we fix it?
http://snag.gy/xhrK9.jpg - see the styles in the inspector and the white space
http://snag.gy/mv77H.jpg - correct display achieved after i resized the viewport.
You can see it in action here - http://rush-health-wellness.centreteksolutions.net/health-wellness/discover-health | test | weird white space on initial lode in mobile one col view i ve attached some screen shots here there is a huge amount of white space after each item only when you first lode in mobile mode which is one column if you filter or resize your screen or load more everything goes back to normal what s causing this issue and how do we fix it see the styles in the inspector and the white space correct display achieved after i resized the viewport you can see it in action here | 1 |
271,948 | 8,494,121,269 | IssuesEvent | 2018-10-28 18:35:35 | angular-buddies/angular-buddies | https://api.github.com/repos/angular-buddies/angular-buddies | opened | Automatically format files generated by other schematics | comp: prettier effort2: medium (days) priority: 2 (required) type: feature | ### Bug Report or Feature Request (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x] feature request
```
### Package (mark with an `x`)
```
- [x] @angular-buddies/prettier
```
### Versions
@angular-buddies/prettier:1.0.0-alpha.0
### Desired functionality
When running `ng generate` commands (component, service, etc...), the resulting files should be automatically formatted using prettier.
### Mention any other details that might be useful
To do so, we need to:
* [ ] create a prettify rule w/ associated task (see https://github.com/angular/angular-cli/blob/master/packages/schematics/angular/utility/lint-fix.ts)
* [ ] create all the required schematics, which would simply call the ones from [schematics/angular](https://github.com/angular/angular-cli/tree/master/packages/schematics/angular) and run prettier on touched files
* [ ] add the @angular-buddies/prettier schematics collection as the defaultCollection in angular.json - this feature has been developed, but was reverted in af8e92a37eeee970d921143b2475b6285ea853a5 for v1.0.0-alpha.0, with the related unit tests keeped as xit
:warning: we don't want to break this buddy when using it without angular
| 1.0 | Automatically format files generated by other schematics - ### Bug Report or Feature Request (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x] feature request
```
### Package (mark with an `x`)
```
- [x] @angular-buddies/prettier
```
### Versions
@angular-buddies/prettier:1.0.0-alpha.0
### Desired functionality
When running `ng generate` commands (component, service, etc...), the resulting files should be automatically formatted using prettier.
### Mention any other details that might be useful
To do so, we need to:
* [ ] create a prettify rule w/ associated task (see https://github.com/angular/angular-cli/blob/master/packages/schematics/angular/utility/lint-fix.ts)
* [ ] create all the required schematics, which would simply call the ones from [schematics/angular](https://github.com/angular/angular-cli/tree/master/packages/schematics/angular) and run prettier on touched files
* [ ] add the @angular-buddies/prettier schematics collection as the defaultCollection in angular.json - this feature has been developed, but was reverted in af8e92a37eeee970d921143b2475b6285ea853a5 for v1.0.0-alpha.0, with the related unit tests keeped as xit
:warning: we don't want to break this buddy when using it without angular
| non_test | automatically format files generated by other schematics bug report or feature request mark with an x bug report please search issues before submitting feature request package mark with an x angular buddies prettier versions angular buddies prettier alpha desired functionality when running ng generate commands component service etc the resulting files should be automatically formatted using prettier mention any other details that might be useful to do so we need to create a prettify rule w associated task see create all the required schematics which would simply call the ones from and run prettier on touched files add the angular buddies prettier schematics collection as the defaultcollection in angular json this feature has been developed but was reverted in for alpha with the related unit tests keeped as xit warning we don t want to break this buddy when using it without angular | 0 |
432,002 | 30,262,240,011 | IssuesEvent | 2023-07-07 09:05:25 | systemd/systemd | https://api.github.com/repos/systemd/systemd | closed | document how hibernate operates without resume= switch on kernel cmdline | documentation hibernate-resume | **Is your feature request related to a problem? Please describe.**
I mistakenly thought that `resume`/`resume_offset` parameters were only needed on resume; not for hibernating in the first place. This led me on a bit of a wild goose chase wondering why resuming wasn't working properly, when it's simply because I never hibernated in the first place - `systemctl hibernate` (or from `HandleLidSwitch=hibernate`) was simply causing a shutdown.
**Describe the solution you'd like**
I'm a bit naive as to the implications, but ideally I'd think it could refuse to hibernate (exit with an error) if `/sys/power/resume` is `0:0`?
**Describe alternatives you've considered**
Alternatively, if not possible or the existing behaviour is desirable for some reason, I'd suggest putting a warning in the man page. Neither `systemctl(1)` nor `systemd.special(7)` mention this behaviour, as far as I can tell.
**The systemd version you checked that didn't have the feature you are asking for**
<!-- If this is not the most recently released upstream version, then please check first if it has that feature already. -->
```
systemd 250 (250.4-2-arch)
+PAM +AUDIT -SELINUX -APPARMOR -IMA +SMACK +SECCOMP +GCRYPT +GNUTLS +OPENSSL +ACL +BLKID +CURL +ELFUTILS +FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY +P11KIT -QRENCODE +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK +XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified
``` | 1.0 | document how hibernate operates without resume= switch on kernel cmdline - **Is your feature request related to a problem? Please describe.**
I mistakenly thought that `resume`/`resume_offset` parameters were only needed on resume; not for hibernating in the first place. This led me on a bit of a wild goose chase wondering why resuming wasn't working properly, when it's simply because I never hibernated in the first place - `systemctl hibernate` (or from `HandleLidSwitch=hibernate`) was simply causing a shutdown.
**Describe the solution you'd like**
I'm a bit naive as to the implications, but ideally I'd think it could refuse to hibernate (exit with an error) if `/sys/power/resume` is `0:0`?
**Describe alternatives you've considered**
Alternatively, if not possible or the existing behaviour is desirable for some reason, I'd suggest putting a warning in the man page. Neither `systemctl(1)` nor `systemd.special(7)` mention this behaviour, as far as I can tell.
**The systemd version you checked that didn't have the feature you are asking for**
<!-- If this is not the most recently released upstream version, then please check first if it has that feature already. -->
```
systemd 250 (250.4-2-arch)
+PAM +AUDIT -SELINUX -APPARMOR -IMA +SMACK +SECCOMP +GCRYPT +GNUTLS +OPENSSL +ACL +BLKID +CURL +ELFUTILS +FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY +P11KIT -QRENCODE +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK +XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified
``` | non_test | document how hibernate operates without resume switch on kernel cmdline is your feature request related to a problem please describe i mistakenly thought that resume resume offset parameters were only needed on resume not for hibernating in the first place this led me on a bit of a wild goose chase wondering why resuming wasn t working properly when it s simply because i never hibernated in the first place systemctl hibernate or from handlelidswitch hibernate was simply causing a shutdown describe the solution you d like i m a bit naive as to the implications but ideally i d think it could refuse to hibernate exit with an error if sys power resume is describe alternatives you ve considered alternatively if not possible or the existing behaviour is desirable for some reason i d suggest putting a warning in the man page neither systemctl nor systemd special mention this behaviour as far as i can tell the systemd version you checked that didn t have the feature you are asking for systemd arch pam audit selinux apparmor ima smack seccomp gcrypt gnutls openssl acl blkid curl elfutils idn iptc kmod libcryptsetup libfdisk pwquality qrencode xz zlib zstd bpf framework xkbcommon utmp sysvinit default hierarchy unified | 0 |
41,543 | 16,778,111,552 | IssuesEvent | 2021-06-15 01:48:26 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | Pod & Service CIDR must be /18 or larger | ARO Service Attention question | Pod & Service CIDR must be /18 or larger, this should be documented in the CLI docs !
---
#### Dรฉtails du document
โ *Ne pas modifier cette section. Cโest obligatoire pour docs.microsoft.com โ Liaison des problรจmes GitHub.*
* ID: 08eb181a-b377-c854-69cc-3b9307f59f72
* Version Independent ID: 2142cb43-b027-86a1-6d3c-2f77bf86d177
* Content: [az aro](https://docs.microsoft.com/fr-fr/cli/azure/aro?view=azure-cli-latest#az-aro-create)
* Content Source: [src/azure-cli/azure/cli/command_modules/aro/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/aro/_help.py)
* GitHub Login: @rloutlaw
* Microsoft Alias: **routlaw** | 1.0 | Pod & Service CIDR must be /18 or larger - Pod & Service CIDR must be /18 or larger, this should be documented in the CLI docs !
---
#### Dรฉtails du document
โ *Ne pas modifier cette section. Cโest obligatoire pour docs.microsoft.com โ Liaison des problรจmes GitHub.*
* ID: 08eb181a-b377-c854-69cc-3b9307f59f72
* Version Independent ID: 2142cb43-b027-86a1-6d3c-2f77bf86d177
* Content: [az aro](https://docs.microsoft.com/fr-fr/cli/azure/aro?view=azure-cli-latest#az-aro-create)
* Content Source: [src/azure-cli/azure/cli/command_modules/aro/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/aro/_help.py)
* GitHub Login: @rloutlaw
* Microsoft Alias: **routlaw** | non_test | pod service cidr must be or larger pod service cidr must be or larger this should be documented in the cli docs dรฉtails du document โ ne pas modifier cette section cโest obligatoire pour docs microsoft com โ liaison des problรจmes github id version independent id content content source github login rloutlaw microsoft alias routlaw | 0 |
48,429 | 20,150,541,970 | IssuesEvent | 2022-02-09 11:56:27 | gradido/gradido | https://api.github.com/repos/gradido/gradido | closed | ๐ง [Refactor] Apollo Server has URL /graphql - why? | question refactor service: backend | ## ๐ง Refactor ticket
<!-- Describe your issue in detail. Include screenshots if needed. Give us as much information as possible. Use a clear and concise description of what the problem is.-->
Apollo Server has URL `/graphql` - why?
Please serve the graphql under `/` - or why would we need other parts that are not graphql?
-> This complicates nginx configuration (even tho solveable its not nice - I serve the server in nginx under `host/backend` and the server itself serves und `/graphql` making an URL `host/backend/graphql`) | 1.0 | ๐ง [Refactor] Apollo Server has URL /graphql - why? - ## ๐ง Refactor ticket
<!-- Describe your issue in detail. Include screenshots if needed. Give us as much information as possible. Use a clear and concise description of what the problem is.-->
Apollo Server has URL `/graphql` - why?
Please serve the graphql under `/` - or why would we need other parts that are not graphql?
-> This complicates nginx configuration (even tho solveable its not nice - I serve the server in nginx under `host/backend` and the server itself serves und `/graphql` making an URL `host/backend/graphql`) | non_test | ๐ง apollo server has url graphql why ๐ง refactor ticket apollo server has url graphql why please serve the graphql under or why would we need other parts that are not graphql this complicates nginx configuration even tho solveable its not nice i serve the server in nginx under host backend and the server itself serves und graphql making an url host backend graphql | 0 |
296,048 | 25,524,334,325 | IssuesEvent | 2022-11-29 00:02:56 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | closed | [Flaky Test] cancels dragging blocks from the global inserter by pressing Escape | [Status] Stale [Type] Flaky Test | <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
cancels dragging blocks from the global inserter by pressing Escape
## Test path
`/test/e2e/specs/editor/various/inserting-blocks.spec.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2022-10-27T11:09:38.009Z"><code>[2022-10-27T11:09:38.009Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/3336633956"><code>rnmobile/release_1.85.0</code></a>.
</summary>
```
locator.boundingBox: Timeout 10000ms exceeded.
=========================== logs ===========================
waiting for selector "[data-type="core/paragraph"] >> text=Dummy text"
selector resolved to visible <p tabindex="0" role="document" data-empty="false" โฆ>Dummy text</p>
============================================================
at /home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/various/inserting-blocks.spec.js:130:53
at /home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/workerRunner.js:426:9
at TestInfoImpl._runFn (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/testInfo.js:166:7)
at /home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/workerRunner.js:376:26
at TimeoutManager.runWithTimeout (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/timeoutManager.js:73:7)
at TestInfoImpl._runWithTimeout (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/testInfo.js:154:26)
at WorkerRunner._runTest (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/workerRunner.js:356:5)
at WorkerRunner.runTestGroup (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/workerRunner.js:218:11)
at process.<anonymous> (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/worker.js:88:5)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| 1.0 | [Flaky Test] cancels dragging blocks from the global inserter by pressing Escape - <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
cancels dragging blocks from the global inserter by pressing Escape
## Test path
`/test/e2e/specs/editor/various/inserting-blocks.spec.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2022-10-27T11:09:38.009Z"><code>[2022-10-27T11:09:38.009Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/3336633956"><code>rnmobile/release_1.85.0</code></a>.
</summary>
```
locator.boundingBox: Timeout 10000ms exceeded.
=========================== logs ===========================
waiting for selector "[data-type="core/paragraph"] >> text=Dummy text"
selector resolved to visible <p tabindex="0" role="document" data-empty="false" โฆ>Dummy text</p>
============================================================
at /home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/various/inserting-blocks.spec.js:130:53
at /home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/workerRunner.js:426:9
at TestInfoImpl._runFn (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/testInfo.js:166:7)
at /home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/workerRunner.js:376:26
at TimeoutManager.runWithTimeout (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/timeoutManager.js:73:7)
at TestInfoImpl._runWithTimeout (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/testInfo.js:154:26)
at WorkerRunner._runTest (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/workerRunner.js:356:5)
at WorkerRunner.runTestGroup (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/workerRunner.js:218:11)
at process.<anonymous> (/home/runner/work/gutenberg/gutenberg/node_modules/@playwright/test/lib/worker.js:88:5)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| test | cancels dragging blocks from the global inserter by pressing escape flaky test detected this is an auto generated issue by github actions please do not edit this manually test title cancels dragging blocks from the global inserter by pressing escape test path test specs editor various inserting blocks spec js errors test passed after failed attempt on a href locator boundingbox timeout exceeded logs waiting for selector text dummy text selector resolved to visible dummy text at home runner work gutenberg gutenberg test specs editor various inserting blocks spec js at home runner work gutenberg gutenberg node modules playwright test lib workerrunner js at testinfoimpl runfn home runner work gutenberg gutenberg node modules playwright test lib testinfo js at home runner work gutenberg gutenberg node modules playwright test lib workerrunner js at timeoutmanager runwithtimeout home runner work gutenberg gutenberg node modules playwright test lib timeoutmanager js at testinfoimpl runwithtimeout home runner work gutenberg gutenberg node modules playwright test lib testinfo js at workerrunner runtest home runner work gutenberg gutenberg node modules playwright test lib workerrunner js at workerrunner runtestgroup home runner work gutenberg gutenberg node modules playwright test lib workerrunner js at process home runner work gutenberg gutenberg node modules playwright test lib worker js | 1 |
80,415 | 10,172,690,399 | IssuesEvent | 2019-08-08 11:20:18 | phpMussel/phpMussel | https://api.github.com/repos/phpMussel/phpMussel | closed | Preview the alpha version of the english book! | Documentation | Hey,
I'm Max.
Teamwork by @Asrez :

```
Copyright 2019 Max Base
Last Updated: 26 May 2019 (2019.05.26).
Writing by Caleb Mazalevskis, Max Base
The first release, Jun 2019
```
Sorry, But I keep your name here. (Caleb Mazalevskis)
Not worry, because it's the first version.
and we will apply some new change...
https://github.com/BaseMax/phpMusselBook (You can make a star it if you love it)
You can preview the e-book file : [phpMussel.pdf](https://github.com/BaseMax/phpMusselBook/blob/master/phpMussel.pdf)
@nhlm @Maikuolan @petk @SuriyaaKudoIsc @DanielRuf
Regards,
Max | 1.0 | Preview the alpha version of the english book! - Hey,
I'm Max.
Teamwork by @Asrez :

```
Copyright 2019 Max Base
Last Updated: 26 May 2019 (2019.05.26).
Writing by Caleb Mazalevskis, Max Base
The first release, Jun 2019
```
Sorry, But I keep your name here. (Caleb Mazalevskis)
Not worry, because it's the first version.
and we will apply some new change...
https://github.com/BaseMax/phpMusselBook (You can make a star it if you love it)
You can preview the e-book file : [phpMussel.pdf](https://github.com/BaseMax/phpMusselBook/blob/master/phpMussel.pdf)
@nhlm @Maikuolan @petk @SuriyaaKudoIsc @DanielRuf
Regards,
Max | non_test | preview the alpha version of the english book hey i m max teamwork by asrez copyright max base last updated may writing by caleb mazalevskis max base the first release jun sorry but i keep your name here caleb mazalevskis not worry because it s the first version and we will apply some new change you can make a star it if you love it you can preview the e book file nhlm maikuolan petk suriyaakudoisc danielruf regards max | 0 |
246,790 | 26,622,352,482 | IssuesEvent | 2023-01-24 12:13:16 | elastic/integrations | https://api.github.com/repos/elastic/integrations | closed | cisco_ise tacacs pipeline missing error handling | Team:Security-External Integrations release-pending Integration:CiscoISE | I am investigating a problem of processing cisco_ise logs. Some of the logs are rejected with the following message:
`"error"=>{"type"=>"mapper_parsing_exception", "reason"=>"failed to parse field [AVPair] of type [keyword] in document with id 'YWm404IBFzFBE6qKgb3x'. Preview of field's value: '{elapsed_time=0, task_id=4f4ad6}'", "caused_by"=>{"type"=>"illegal_state_exception", "reason"=>"Can't get text on a START_OBJECT at 1:3163"}}}}}`
From what I see, the AVPair field should be removed at the end of the pipeline (remove processor). One way this could happen is when a previous processor fails. This may be the case, since not all processors have the igonre_failure flag set and there are no on_failure processors set in the pipeline. I have fixed this temporarily on my cluster by adding a on_failure processor to remove the temporary stuff and it seems to help. Another solution would be to set ignore_failure flags on all of the previous processors, but I don't know the integration well enough to evaluate the side effects. Overall, rejecting logs in case of parsing errors should be avoided if ever possible, since in case of UDP they will just get lost in this case.
| True | cisco_ise tacacs pipeline missing error handling - I am investigating a problem of processing cisco_ise logs. Some of the logs are rejected with the following message:
`"error"=>{"type"=>"mapper_parsing_exception", "reason"=>"failed to parse field [AVPair] of type [keyword] in document with id 'YWm404IBFzFBE6qKgb3x'. Preview of field's value: '{elapsed_time=0, task_id=4f4ad6}'", "caused_by"=>{"type"=>"illegal_state_exception", "reason"=>"Can't get text on a START_OBJECT at 1:3163"}}}}}`
From what I see, the AVPair field should be removed at the end of the pipeline (remove processor). One way this could happen is when a previous processor fails. This may be the case, since not all processors have the igonre_failure flag set and there are no on_failure processors set in the pipeline. I have fixed this temporarily on my cluster by adding a on_failure processor to remove the temporary stuff and it seems to help. Another solution would be to set ignore_failure flags on all of the previous processors, but I don't know the integration well enough to evaluate the side effects. Overall, rejecting logs in case of parsing errors should be avoided if ever possible, since in case of UDP they will just get lost in this case.
| non_test | cisco ise tacacs pipeline missing error handling i am investigating a problem of processing cisco ise logs some of the logs are rejected with the following message error type mapper parsing exception reason failed to parse field of type in document with id preview of field s value elapsed time task id caused by type illegal state exception reason can t get text on a start object at from what i see the avpair field should be removed at the end of the pipeline remove processor one way this could happen is when a previous processor fails this may be the case since not all processors have the igonre failure flag set and there are no on failure processors set in the pipeline i have fixed this temporarily on my cluster by adding a on failure processor to remove the temporary stuff and it seems to help another solution would be to set ignore failure flags on all of the previous processors but i don t know the integration well enough to evaluate the side effects overall rejecting logs in case of parsing errors should be avoided if ever possible since in case of udp they will just get lost in this case | 0 |
72,929 | 7,315,700,852 | IssuesEvent | 2018-03-01 12:02:58 | Kademi/kademi-dev | https://api.github.com/repos/Kademi/kademi-dev | closed | Adding pie component query cause error on the page frontend | Ready to Test QA bug | 
1. Add pie component
2. Save
3. View page on the frontend
4. Check error
http://vladtest34a.admin.kademi-ci.co/pages/vladtest34aweb/version1/ | 1.0 | Adding pie component query cause error on the page frontend - 
1. Add pie component
2. Save
3. View page on the frontend
4. Check error
http://vladtest34a.admin.kademi-ci.co/pages/vladtest34aweb/version1/ | test | adding pie component query cause error on the page frontend add pie component save view page on the frontend check error | 1 |
41,301 | 5,348,294,284 | IssuesEvent | 2017-02-18 03:21:00 | red/red | https://api.github.com/repos/red/red | closed | Field can show hint text when it's blank. | status.built status.tested type.wish | In old-style GUI design, a field is always next to a label used to tell user what the filed is about. In new-style GUI design, a field can have a prompt string shown such as "Search", "Your name here" when it's blank. I would like Red to support this feature in GUI engine and VID. Thanks. | 1.0 | Field can show hint text when it's blank. - In old-style GUI design, a field is always next to a label used to tell user what the filed is about. In new-style GUI design, a field can have a prompt string shown such as "Search", "Your name here" when it's blank. I would like Red to support this feature in GUI engine and VID. Thanks. | test | field can show hint text when it s blank in old style gui design a field is always next to a label used to tell user what the filed is about in new style gui design a field can have a prompt string shown such as search your name here when it s blank i would like red to support this feature in gui engine and vid thanks | 1 |
263,931 | 23,091,348,126 | IssuesEvent | 2022-07-26 15:27:15 | ImagingDataCommons/IDC-WebApp | https://api.github.com/repos/ImagingDataCommons/IDC-WebApp | closed | Domain/origin issue with pushState and portal.imaging vs. imaging URIs | bug merged:dev testing needed in production production | First, this collection is labeled as "Limited", but when it is selected, it shows CC-BY licenses for all cases in that collecton.
Second, in the tooltip, this collection points to https://wiki.cancerimagingarchive.net/display/Public/RIDER+Collections, while it should point to the collection-specific page https://wiki.cancerimagingarchive.net/display/Public/RIDER+NEURO+MRI.

| 1.0 | Domain/origin issue with pushState and portal.imaging vs. imaging URIs - First, this collection is labeled as "Limited", but when it is selected, it shows CC-BY licenses for all cases in that collecton.
Second, in the tooltip, this collection points to https://wiki.cancerimagingarchive.net/display/Public/RIDER+Collections, while it should point to the collection-specific page https://wiki.cancerimagingarchive.net/display/Public/RIDER+NEURO+MRI.

| test | domain origin issue with pushstate and portal imaging vs imaging uris first this collection is labeled as limited but when it is selected it shows cc by licenses for all cases in that collecton second in the tooltip this collection points to while it should point to the collection specific page | 1 |
297,785 | 9,181,659,663 | IssuesEvent | 2019-03-05 10:45:33 | PaxelK/ThesisWorkCybercom | https://api.github.com/repos/PaxelK/ThesisWorkCybercom | closed | Describe the RL approach | High Priority | Main Issue #37
- State the Q-learning optimization problem with initial cost function and it's strategy to reaching the optimal policy (how it is going to be done mathematically)
- Explore the Markov Decision Process. The Q-learning problem reduces to a MDP in the end right?
- Also propose the research interest of seeing which penalty factor (discount) and lambda (if it is used) that gives the best result | 1.0 | Describe the RL approach - Main Issue #37
- State the Q-learning optimization problem with initial cost function and it's strategy to reaching the optimal policy (how it is going to be done mathematically)
- Explore the Markov Decision Process. The Q-learning problem reduces to a MDP in the end right?
- Also propose the research interest of seeing which penalty factor (discount) and lambda (if it is used) that gives the best result | non_test | describe the rl approach main issue state the q learning optimization problem with initial cost function and it s strategy to reaching the optimal policy how it is going to be done mathematically explore the markov decision process the q learning problem reduces to a mdp in the end right also propose the research interest of seeing which penalty factor discount and lambda if it is used that gives the best result | 0 |
146,364 | 11,734,867,876 | IssuesEvent | 2020-03-11 10:06:42 | chameleon-system/chameleon-system | https://api.github.com/repos/chameleon-system/chameleon-system | closed | Backend: Layout inconsistency - missing content top padding | Status: Test | **Describe the bug**
The content modules have no top margin/padding to the breadcrumb.
**Affected version(s)**
6.3.x
**To Reproduce**
Steps to reproduce the behavior:
1. Go to backend home and any record edit view
2. See the difference in top padding
**Expected behavior**
There should always be a top border in the content container.
**Screenshots**

**Desktop (please complete the following information):**
- OS: Linux
- Browser: Chrome + FF
**Technical details**
I'm not sure where to add the padding. Below the header toolbar/breadcrumb or for the content container. This would lead to additional space above the record edit button toolbar and would need a minus margin or something.
| 1.0 | Backend: Layout inconsistency - missing content top padding - **Describe the bug**
The content modules have no top margin/padding to the breadcrumb.
**Affected version(s)**
6.3.x
**To Reproduce**
Steps to reproduce the behavior:
1. Go to backend home and any record edit view
2. See the difference in top padding
**Expected behavior**
There should always be a top border in the content container.
**Screenshots**

**Desktop (please complete the following information):**
- OS: Linux
- Browser: Chrome + FF
**Technical details**
I'm not sure where to add the padding. Below the header toolbar/breadcrumb or for the content container. This would lead to additional space above the record edit button toolbar and would need a minus margin or something.
| test | backend layout inconsistency missing content top padding describe the bug the content modules have no top margin padding to the breadcrumb affected version s x to reproduce steps to reproduce the behavior go to backend home and any record edit view see the difference in top padding expected behavior there should always be a top border in the content container screenshots desktop please complete the following information os linux browser chrome ff technical details i m not sure where to add the padding below the header toolbar breadcrumb or for the content container this would lead to additional space above the record edit button toolbar and would need a minus margin or something | 1 |
3,332 | 5,774,585,826 | IssuesEvent | 2017-04-28 07:39:46 | parksandwildlife/biosys-turtles | https://api.github.com/repos/parksandwildlife/biosys-turtles | opened | Role-based interface functionality | Functional Requirement must have | ### Source
Name requesting role or stakeholder from which this requirement was sourced:
* Core stakeholders
* Turtle tagging stakeholders
* Turtle track count stakeholders
* Marine wildlife stranding stakeholders
* Departmental stakeholders
#14
### Requirement
The solution shall provide appropriate functionality based on defined roles that a user is assigned to.
### Use cases
* System administrators have full access to all functionality, e.g. user management and allocating roles
* Curators have full access to data-related functionality
* Field team leaders have permissions and functionality to change or annotate data captured by their teams
* Data collectors have permissions and functionality to add or change their own data
| 1.0 | Role-based interface functionality - ### Source
Name requesting role or stakeholder from which this requirement was sourced:
* Core stakeholders
* Turtle tagging stakeholders
* Turtle track count stakeholders
* Marine wildlife stranding stakeholders
* Departmental stakeholders
#14
### Requirement
The solution shall provide appropriate functionality based on defined roles that a user is assigned to.
### Use cases
* System administrators have full access to all functionality, e.g. user management and allocating roles
* Curators have full access to data-related functionality
* Field team leaders have permissions and functionality to change or annotate data captured by their teams
* Data collectors have permissions and functionality to add or change their own data
| non_test | role based interface functionality source name requesting role or stakeholder from which this requirement was sourced core stakeholders turtle tagging stakeholders turtle track count stakeholders marine wildlife stranding stakeholders departmental stakeholders requirement the solution shall provide appropriate functionality based on defined roles that a user is assigned to use cases system administrators have full access to all functionality e g user management and allocating roles curators have full access to data related functionality field team leaders have permissions and functionality to change or annotate data captured by their teams data collectors have permissions and functionality to add or change their own data | 0 |
277,233 | 24,055,505,903 | IssuesEvent | 2022-09-16 16:26:41 | gradle/gradle | https://api.github.com/repos/gradle/gradle | closed | TestSuiteExecutionException: Could not execute test class 'MyApp$delayedInit$body'. | in:testing a:bug stale | <!---
Please follow the instructions below. We receive dozens of issues every week, so to stay productive, we will close issues that don't provide enough information.
Please open Android-related issues on the Android Issue Tracker at https://source.android.com/source/report-bugs
Please open Kotlin DSL-related issues at https://github.com/gradle/kotlin-dsl/issues
Please open Gradle Native-related issues at https://github.com/gradle/gradle-native/issues
-->
<!--- Provide a brief summary of the issue in the title above -->
Objects extending scala.App that happen to hang around in a test source set seem to get discovered when doing "gradle test" and then delayedInit of scala.App causes the exception.
```
org.gradle.api.internal.tasks.testing.TestSuiteExecutionException: Could not execute test class 'MyApp$delayedInit$body'.
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:53)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
...
Caused by: java.lang.IncompatibleClassChangeError: MyApp and MyApp$delayedInit$body disagree on InnerClasses attribute
at java.lang.Class.getDeclaringClass0(Native Method)
at java.lang.Class.getDeclaringClass(Class.java:1235)
at java.lang.Class.getEnclosingClass(Class.java:1277)
at org.gradle.api.internal.tasks.testing.junitplatform.JUnitPlatformTestClassProcessor.isInnerClass(JUnitPlatformTestClassProcessor.java:97)
```
### Expected Behavior
scala.App objects should be discovered and excluded
### Current Behavior
They are discovered.
### Context
Have to work around it by rewriting those objects to not use scala.App. Behaviour is surprising.
### Steps to Reproduce (for bugs)
Just add following class to your test source set next to a unit test fixture and do gradle test.
```
object MyApp extends App {
println("Initialize")
}
```
### Your Environment
gradle 4.7
junit 5.2
| 1.0 | TestSuiteExecutionException: Could not execute test class 'MyApp$delayedInit$body'. - <!---
Please follow the instructions below. We receive dozens of issues every week, so to stay productive, we will close issues that don't provide enough information.
Please open Android-related issues on the Android Issue Tracker at https://source.android.com/source/report-bugs
Please open Kotlin DSL-related issues at https://github.com/gradle/kotlin-dsl/issues
Please open Gradle Native-related issues at https://github.com/gradle/gradle-native/issues
-->
<!--- Provide a brief summary of the issue in the title above -->
Objects extending scala.App that happen to hang around in a test source set seem to get discovered when doing "gradle test" and then delayedInit of scala.App causes the exception.
```
org.gradle.api.internal.tasks.testing.TestSuiteExecutionException: Could not execute test class 'MyApp$delayedInit$body'.
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:53)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
...
Caused by: java.lang.IncompatibleClassChangeError: MyApp and MyApp$delayedInit$body disagree on InnerClasses attribute
at java.lang.Class.getDeclaringClass0(Native Method)
at java.lang.Class.getDeclaringClass(Class.java:1235)
at java.lang.Class.getEnclosingClass(Class.java:1277)
at org.gradle.api.internal.tasks.testing.junitplatform.JUnitPlatformTestClassProcessor.isInnerClass(JUnitPlatformTestClassProcessor.java:97)
```
### Expected Behavior
scala.App objects should be discovered and excluded
### Current Behavior
They are discovered.
### Context
Have to work around it by rewriting those objects to not use scala.App. Behaviour is surprising.
### Steps to Reproduce (for bugs)
Just add following class to your test source set next to a unit test fixture and do gradle test.
```
object MyApp extends App {
println("Initialize")
}
```
### Your Environment
gradle 4.7
junit 5.2
| test | testsuiteexecutionexception could not execute test class myapp delayedinit body please follow the instructions below we receive dozens of issues every week so to stay productive we will close issues that don t provide enough information please open android related issues on the android issue tracker at please open kotlin dsl related issues at please open gradle native related issues at objects extending scala app that happen to hang around in a test source set seem to get discovered when doing gradle test and then delayedinit of scala app causes the exception org gradle api internal tasks testing testsuiteexecutionexception could not execute test class myapp delayedinit body at org gradle api internal tasks testing suitetestclassprocessor processtestclass suitetestclassprocessor java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java caused by java lang incompatibleclasschangeerror myapp and myapp delayedinit body disagree on innerclasses attribute at java lang class native method at java lang class getdeclaringclass class java at java lang class getenclosingclass class java at org gradle api internal tasks testing junitplatform junitplatformtestclassprocessor isinnerclass junitplatformtestclassprocessor java expected behavior scala app objects should be discovered and excluded current behavior they are discovered context have to work around it by rewriting those objects to not use scala app behaviour is surprising steps to reproduce for bugs just add following class to your test source set next to a unit test fixture and do gradle test object myapp extends app println initialize your environment gradle junit | 1 |
198,462 | 14,981,169,200 | IssuesEvent | 2021-01-28 14:31:56 | HydrolienF/Formiko | https://api.github.com/repos/HydrolienF/Formiko | closed | fined a way to have a %age of covered test | done test tools | J'aimerai avoir un test indice de couverture de test.
Peut-etre qu'on peu faire nous meme une liste de l'ensemble des fonctions du projet et la comparรฉ avec les fonctions testรฉ :
On stock les fonctions du projet dans une table de hachage String,int avec en String le nom de la fonction et en int 0.
On peu crรฉer un fichier avec tout le contenu des test, le charger dans un GString.
1a traitement on retire les lignes de commentaire et les ligne qui commence par "import" ou "package" & on remplace les '(',')' et '.' par des espaces. (comme ca les fonctions seront considรฉrรฉ comme des mots.
2a traitement on stock tout les mots sรฉparรฉment
3a traitement on retire tout les mots du langage java.
4a traitement on supprime tout les mots en doublons. (ca fera moins de travail aprรจs).
5a traitement on retire les mots qui ne commence pas par une lettre. (commence par @ ou par un nombre ou un char spรฉcial)
On pourrait mรชme idรฉalement la triรฉe pour pouvoir chercher juste parmi les fonctions qui commence par une lettre en question.
A ce stade on a une liste de mot qui contient toutes les fonctions testรฉe.
Pour chaqu'1 des fonctions trouvรฉ dans le projet :
On parcours le GString en cherchant la fonction. (On s'arrรชte a la 1a occurence et on revoie true (cad fonction bien testรฉ)
Si true on ajoute 1 dans la table de hackage.
Sinon on fait rien.
Dรฉfaut de ce pseudo code : la signature d'une fonction n'est pas prise en compte. max(int, int) peut etre traitรฉ sans que max(double, double) le soit.
Peut-รชtre qu'on peu obtenir plus facilement la liste des mรฉthodes utilisรฉes et leurs signatures.
| 1.0 | fined a way to have a %age of covered test - J'aimerai avoir un test indice de couverture de test.
Peut-etre qu'on peu faire nous meme une liste de l'ensemble des fonctions du projet et la comparรฉ avec les fonctions testรฉ :
On stock les fonctions du projet dans une table de hachage String,int avec en String le nom de la fonction et en int 0.
On peu crรฉer un fichier avec tout le contenu des test, le charger dans un GString.
1a traitement on retire les lignes de commentaire et les ligne qui commence par "import" ou "package" & on remplace les '(',')' et '.' par des espaces. (comme ca les fonctions seront considรฉrรฉ comme des mots.
2a traitement on stock tout les mots sรฉparรฉment
3a traitement on retire tout les mots du langage java.
4a traitement on supprime tout les mots en doublons. (ca fera moins de travail aprรจs).
5a traitement on retire les mots qui ne commence pas par une lettre. (commence par @ ou par un nombre ou un char spรฉcial)
On pourrait mรชme idรฉalement la triรฉe pour pouvoir chercher juste parmi les fonctions qui commence par une lettre en question.
A ce stade on a une liste de mot qui contient toutes les fonctions testรฉe.
Pour chaqu'1 des fonctions trouvรฉ dans le projet :
On parcours le GString en cherchant la fonction. (On s'arrรชte a la 1a occurence et on revoie true (cad fonction bien testรฉ)
Si true on ajoute 1 dans la table de hackage.
Sinon on fait rien.
Dรฉfaut de ce pseudo code : la signature d'une fonction n'est pas prise en compte. max(int, int) peut etre traitรฉ sans que max(double, double) le soit.
Peut-รชtre qu'on peu obtenir plus facilement la liste des mรฉthodes utilisรฉes et leurs signatures.
| test | fined a way to have a age of covered test j aimerai avoir un test indice de couverture de test peut etre qu on peu faire nous meme une liste de l ensemble des fonctions du projet et la comparรฉ avec les fonctions testรฉ on stock les fonctions du projet dans une table de hachage string int avec en string le nom de la fonction et en int on peu crรฉer un fichier avec tout le contenu des test le charger dans un gstring traitement on retire les lignes de commentaire et les ligne qui commence par import ou package on remplace les et par des espaces comme ca les fonctions seront considรฉrรฉ comme des mots traitement on stock tout les mots sรฉparรฉment traitement on retire tout les mots du langage java traitement on supprime tout les mots en doublons ca fera moins de travail aprรจs traitement on retire les mots qui ne commence pas par une lettre commence par ou par un nombre ou un char spรฉcial on pourrait mรชme idรฉalement la triรฉe pour pouvoir chercher juste parmi les fonctions qui commence par une lettre en question a ce stade on a une liste de mot qui contient toutes les fonctions testรฉe pour chaqu des fonctions trouvรฉ dans le projet on parcours le gstring en cherchant la fonction on s arrรชte a la occurence et on revoie true cad fonction bien testรฉ si true on ajoute dans la table de hackage sinon on fait rien dรฉfaut de ce pseudo code la signature d une fonction n est pas prise en compte max int int peut etre traitรฉ sans que max double double le soit peut รชtre qu on peu obtenir plus facilement la liste des mรฉthodes utilisรฉes et leurs signatures | 1 |
307,894 | 26,569,183,155 | IssuesEvent | 2023-01-21 00:29:58 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | closed | Fix jax_numpy_logic.test_jax_numpy_logical_and | JAX Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3946026848/jobs/6753425284" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3946026848/jobs/6753425284" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3946026848/jobs/6753425284" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_logical_and[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-18T05:37:47.8395826Z E AssertionError: False != True
2023-01-18T05:37:47.8396191Z E Falsifying example: test_jax_numpy_logical_and(
2023-01-18T05:37:47.8396680Z E dtypes_values=(['complex128', 'complex64'],
2023-01-18T05:37:47.8397057Z E [array(-5.87747175e-39-5.87747175e-39j),
2023-01-18T05:37:47.8397477Z E array(-1.-1.j, dtype=complex64)]),
2023-01-18T05:37:47.8397972Z E test_flags=num_positional_args=2. with_out=False. inplace=False. native_arrays=[False]. as_variable=[False]. ,
2023-01-18T05:37:47.8398555Z E fn_tree='ivy.functional.frontends.jax.numpy.logical_and',
2023-01-18T05:37:47.8398971Z E on_device='cpu',
2023-01-18T05:37:47.8399317Z E frontend='jax',
2023-01-18T05:37:47.8399622Z E )
2023-01-18T05:37:47.8399874Z E
2023-01-18T05:37:47.8400511Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2cAAsbjDxhgAIWNRDOiiAEAjqoDXA==') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_logical_and[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-18T05:37:47.8395826Z E AssertionError: False != True
2023-01-18T05:37:47.8396191Z E Falsifying example: test_jax_numpy_logical_and(
2023-01-18T05:37:47.8396680Z E dtypes_values=(['complex128', 'complex64'],
2023-01-18T05:37:47.8397057Z E [array(-5.87747175e-39-5.87747175e-39j),
2023-01-18T05:37:47.8397477Z E array(-1.-1.j, dtype=complex64)]),
2023-01-18T05:37:47.8397972Z E test_flags=num_positional_args=2. with_out=False. inplace=False. native_arrays=[False]. as_variable=[False]. ,
2023-01-18T05:37:47.8398555Z E fn_tree='ivy.functional.frontends.jax.numpy.logical_and',
2023-01-18T05:37:47.8398971Z E on_device='cpu',
2023-01-18T05:37:47.8399317Z E frontend='jax',
2023-01-18T05:37:47.8399622Z E )
2023-01-18T05:37:47.8399874Z E
2023-01-18T05:37:47.8400511Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2cAAsbjDxhgAIWNRDOiiAEAjqoDXA==') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_logical_and[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-18T05:37:47.8395826Z E AssertionError: False != True
2023-01-18T05:37:47.8396191Z E Falsifying example: test_jax_numpy_logical_and(
2023-01-18T05:37:47.8396680Z E dtypes_values=(['complex128', 'complex64'],
2023-01-18T05:37:47.8397057Z E [array(-5.87747175e-39-5.87747175e-39j),
2023-01-18T05:37:47.8397477Z E array(-1.-1.j, dtype=complex64)]),
2023-01-18T05:37:47.8397972Z E test_flags=num_positional_args=2. with_out=False. inplace=False. native_arrays=[False]. as_variable=[False]. ,
2023-01-18T05:37:47.8398555Z E fn_tree='ivy.functional.frontends.jax.numpy.logical_and',
2023-01-18T05:37:47.8398971Z E on_device='cpu',
2023-01-18T05:37:47.8399317Z E frontend='jax',
2023-01-18T05:37:47.8399622Z E )
2023-01-18T05:37:47.8399874Z E
2023-01-18T05:37:47.8400511Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2cAAsbjDxhgAIWNRDOiiAEAjqoDXA==') as a decorator on your test case
</details>
| 1.0 | Fix jax_numpy_logic.test_jax_numpy_logical_and - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3946026848/jobs/6753425284" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3946026848/jobs/6753425284" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3946026848/jobs/6753425284" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_logical_and[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-18T05:37:47.8395826Z E AssertionError: False != True
2023-01-18T05:37:47.8396191Z E Falsifying example: test_jax_numpy_logical_and(
2023-01-18T05:37:47.8396680Z E dtypes_values=(['complex128', 'complex64'],
2023-01-18T05:37:47.8397057Z E [array(-5.87747175e-39-5.87747175e-39j),
2023-01-18T05:37:47.8397477Z E array(-1.-1.j, dtype=complex64)]),
2023-01-18T05:37:47.8397972Z E test_flags=num_positional_args=2. with_out=False. inplace=False. native_arrays=[False]. as_variable=[False]. ,
2023-01-18T05:37:47.8398555Z E fn_tree='ivy.functional.frontends.jax.numpy.logical_and',
2023-01-18T05:37:47.8398971Z E on_device='cpu',
2023-01-18T05:37:47.8399317Z E frontend='jax',
2023-01-18T05:37:47.8399622Z E )
2023-01-18T05:37:47.8399874Z E
2023-01-18T05:37:47.8400511Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2cAAsbjDxhgAIWNRDOiiAEAjqoDXA==') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_logical_and[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-18T05:37:47.8395826Z E AssertionError: False != True
2023-01-18T05:37:47.8396191Z E Falsifying example: test_jax_numpy_logical_and(
2023-01-18T05:37:47.8396680Z E dtypes_values=(['complex128', 'complex64'],
2023-01-18T05:37:47.8397057Z E [array(-5.87747175e-39-5.87747175e-39j),
2023-01-18T05:37:47.8397477Z E array(-1.-1.j, dtype=complex64)]),
2023-01-18T05:37:47.8397972Z E test_flags=num_positional_args=2. with_out=False. inplace=False. native_arrays=[False]. as_variable=[False]. ,
2023-01-18T05:37:47.8398555Z E fn_tree='ivy.functional.frontends.jax.numpy.logical_and',
2023-01-18T05:37:47.8398971Z E on_device='cpu',
2023-01-18T05:37:47.8399317Z E frontend='jax',
2023-01-18T05:37:47.8399622Z E )
2023-01-18T05:37:47.8399874Z E
2023-01-18T05:37:47.8400511Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2cAAsbjDxhgAIWNRDOiiAEAjqoDXA==') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_jax/test_jax_numpy_logic.py::test_jax_numpy_logical_and[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-18T05:37:47.8395826Z E AssertionError: False != True
2023-01-18T05:37:47.8396191Z E Falsifying example: test_jax_numpy_logical_and(
2023-01-18T05:37:47.8396680Z E dtypes_values=(['complex128', 'complex64'],
2023-01-18T05:37:47.8397057Z E [array(-5.87747175e-39-5.87747175e-39j),
2023-01-18T05:37:47.8397477Z E array(-1.-1.j, dtype=complex64)]),
2023-01-18T05:37:47.8397972Z E test_flags=num_positional_args=2. with_out=False. inplace=False. native_arrays=[False]. as_variable=[False]. ,
2023-01-18T05:37:47.8398555Z E fn_tree='ivy.functional.frontends.jax.numpy.logical_and',
2023-01-18T05:37:47.8398971Z E on_device='cpu',
2023-01-18T05:37:47.8399317Z E frontend='jax',
2023-01-18T05:37:47.8399622Z E )
2023-01-18T05:37:47.8399874Z E
2023-01-18T05:37:47.8400511Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2cAAsbjDxhgAIWNRDOiiAEAjqoDXA==') as a decorator on your test case
</details>
| test | fix jax numpy logic test jax numpy logical and tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test jax test jax numpy logic py test jax numpy logical and e assertionerror false true e falsifying example test jax numpy logical and e dtypes values e array e array j dtype e test flags num positional args with out false inplace false native arrays as variable e fn tree ivy functional frontends jax numpy logical and e on device cpu e frontend jax e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test jax test jax numpy logic py test jax numpy logical and e assertionerror false true e falsifying example test jax numpy logical and e dtypes values e array e array j dtype e test flags num positional args with out false inplace false native arrays as variable e fn tree ivy functional frontends jax numpy logical and e on device cpu e frontend jax e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test jax test jax numpy logic py test jax numpy logical and e assertionerror false true e falsifying example test jax numpy logical and e dtypes values e array e array j dtype e test flags num positional args with out false inplace false native arrays as variable e fn tree ivy functional frontends jax numpy logical and e on device cpu e frontend jax e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case | 1 |
77,566 | 15,569,810,871 | IssuesEvent | 2021-03-17 01:02:48 | TechnoConserve/ml-practice | https://api.github.com/repos/TechnoConserve/ml-practice | opened | CVE-2021-25293 (Medium) detected in Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl | security vulnerability | ## CVE-2021-25293 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/40/45/cd1000f1c474136236c5105c882d8e1e40bd94ae939b5ca53bf724967514/Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/40/45/cd1000f1c474136236c5105c882d8e1e40bd94ae939b5ca53bf724967514/Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: ml-practice/requirements.txt</p>
<p>Path to vulnerable library: ml-practice/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security issue was found in python-pillow before version 8.1.1. There is an out of bounds read in SGIRleDecode.c, since pillow 4.3.0.
<p>Publish Date: 2021-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25293>CVE-2021-25293</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pillow.readthedocs.io/en/stable/releasenotes/8.1.1.html">https://pillow.readthedocs.io/en/stable/releasenotes/8.1.1.html</a></p>
<p>Release Date: 2021-01-18</p>
<p>Fix Resolution: 8.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-25293 (Medium) detected in Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-25293 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/40/45/cd1000f1c474136236c5105c882d8e1e40bd94ae939b5ca53bf724967514/Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/40/45/cd1000f1c474136236c5105c882d8e1e40bd94ae939b5ca53bf724967514/Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: ml-practice/requirements.txt</p>
<p>Path to vulnerable library: ml-practice/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-4.3.0-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security issue was found in python-pillow before version 8.1.1. There is an out of bounds read in SGIRleDecode.c, since pillow 4.3.0.
<p>Publish Date: 2021-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25293>CVE-2021-25293</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pillow.readthedocs.io/en/stable/releasenotes/8.1.1.html">https://pillow.readthedocs.io/en/stable/releasenotes/8.1.1.html</a></p>
<p>Release Date: 2021-01-18</p>
<p>Fix Resolution: 8.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in pillow whl cve medium severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file ml practice requirements txt path to vulnerable library ml practice requirements txt dependency hierarchy x pillow whl vulnerable library vulnerability details a security issue was found in python pillow before version there is an out of bounds read in sgirledecode c since pillow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
377,067 | 11,162,900,231 | IssuesEvent | 2019-12-26 19:46:35 | AllanWang/Frost-for-Facebook | https://api.github.com/repos/AllanWang/Frost-for-Facebook | closed | Does Not Save Log In Credentials | feature request low priority | - [ โ] I have verified that the bug does not occur in my mobile browser (at https://touch.facebook.com/)
**Describe the bug**
A clear and concise description of what the bug is.
When logging out, options appear.
One of them is to save the info and then log out.
This is what I choose.
The app does not remember any login info.
**To Reproduce**
Steps to reproduce the behaviour:
1. Go to '...' LOGOUT
2. Click on '....' LOGOUT
3. Select '....' SAVE INFO AND LOGOUT
4. TRY TO LOG BACK IN
5. See error
**Details (please provide at least the app version):**
- App Version: 2.4.1
- Device: ZTE558VL
- Android Version: 7.1.1
| 1.0 | Does Not Save Log In Credentials - - [ โ] I have verified that the bug does not occur in my mobile browser (at https://touch.facebook.com/)
**Describe the bug**
A clear and concise description of what the bug is.
When logging out, options appear.
One of them is to save the info and then log out.
This is what I choose.
The app does not remember any login info.
**To Reproduce**
Steps to reproduce the behaviour:
1. Go to '...' LOGOUT
2. Click on '....' LOGOUT
3. Select '....' SAVE INFO AND LOGOUT
4. TRY TO LOG BACK IN
5. See error
**Details (please provide at least the app version):**
- App Version: 2.4.1
- Device: ZTE558VL
- Android Version: 7.1.1
| non_test | does not save log in credentials i have verified that the bug does not occur in my mobile browser at describe the bug a clear and concise description of what the bug is when logging out options appear one of them is to save the info and then log out this is what i choose the app does not remember any login info to reproduce steps to reproduce the behaviour go to logout click on logout select save info and logout try to log back in see error details please provide at least the app version app version device android version | 0 |
190,385 | 22,051,122,006 | IssuesEvent | 2022-05-30 08:45:20 | Seagate/cortx-test | https://api.github.com/repos/Seagate/cortx-test | closed | CVE-2022-24302 (Medium) detected in paramiko-2.6.0-py2.py3-none-any.whl - autoclosed | security vulnerability | ## CVE-2022-24302 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-2.6.0-py2.py3-none-any.whl</b></p></summary>
<p>SSH2 protocol library</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/4b/80/74dace9e48b0ef923633dfb5e48798f58a168e4734bca8ecfaf839ba051a/paramiko-2.6.0-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/4b/80/74dace9e48b0ef923633dfb5e48798f58a168e4734bca8ecfaf839ba051a/paramiko-2.6.0-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt,/libs/csm/csm,/switch,/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **paramiko-2.6.0-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-test/commit/2a86195797b4352c780fa88a5a2cf3dfea63af99">2a86195797b4352c780fa88a5a2cf3dfea63af99</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302>CVE-2022-24302</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.paramiko.org/changelog.html">https://www.paramiko.org/changelog.html</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: paramiko - 2.10.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"paramiko","packageVersion":"2.6.0","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"paramiko:2.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"paramiko - 2.10.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-24302","vulnerabilityDetails":"In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2022-24302 (Medium) detected in paramiko-2.6.0-py2.py3-none-any.whl - autoclosed - ## CVE-2022-24302 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-2.6.0-py2.py3-none-any.whl</b></p></summary>
<p>SSH2 protocol library</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/4b/80/74dace9e48b0ef923633dfb5e48798f58a168e4734bca8ecfaf839ba051a/paramiko-2.6.0-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/4b/80/74dace9e48b0ef923633dfb5e48798f58a168e4734bca8ecfaf839ba051a/paramiko-2.6.0-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt,/libs/csm/csm,/switch,/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **paramiko-2.6.0-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-test/commit/2a86195797b4352c780fa88a5a2cf3dfea63af99">2a86195797b4352c780fa88a5a2cf3dfea63af99</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302>CVE-2022-24302</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.paramiko.org/changelog.html">https://www.paramiko.org/changelog.html</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: paramiko - 2.10.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"paramiko","packageVersion":"2.6.0","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"paramiko:2.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"paramiko - 2.10.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-24302","vulnerabilityDetails":"In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_test | cve medium detected in paramiko none any whl autoclosed cve medium severity vulnerability vulnerable library paramiko none any whl protocol library library home page a href path to dependency file requirements txt path to vulnerable library requirements txt libs csm csm switch requirements txt dependency hierarchy x paramiko none any whl vulnerable library found in head commit a href found in base branch main vulnerability details in paramiko before a race condition between creation and chmod in the write private key file function could allow unauthorized information disclosure publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution paramiko rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree paramiko isminimumfixversionavailable true minimumfixversion paramiko isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in paramiko before a race condition between creation and chmod in the write private key file function could allow unauthorized information disclosure vulnerabilityurl | 0 |
209,783 | 23,730,823,203 | IssuesEvent | 2022-08-31 01:25:45 | himanshu010/Portfolio_ | https://api.github.com/repos/himanshu010/Portfolio_ | opened | CVE-2020-11023 (Medium) detected in jquery-3.3.1.slim.min.js, jquery-2.1.1.min.js | security vulnerability | ## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-3.3.1.slim.min.js</b>, <b>jquery-2.1.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-3.3.1.slim.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.slim.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.slim.min.js</a></p>
<p>Path to dependency file: /index.html</p>
<p>Path to vulnerable library: /index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.3.1.slim.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-2.1.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js</a></p>
<p>Path to dependency file: /index.html</p>
<p>Path to vulnerable library: /index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.1.min.js** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-11023 (Medium) detected in jquery-3.3.1.slim.min.js, jquery-2.1.1.min.js - ## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-3.3.1.slim.min.js</b>, <b>jquery-2.1.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-3.3.1.slim.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.slim.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.slim.min.js</a></p>
<p>Path to dependency file: /index.html</p>
<p>Path to vulnerable library: /index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.3.1.slim.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-2.1.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js</a></p>
<p>Path to dependency file: /index.html</p>
<p>Path to vulnerable library: /index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.1.min.js** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in jquery slim min js jquery min js cve medium severity vulnerability vulnerable libraries jquery slim min js jquery min js jquery slim min js javascript library for dom operations library home page a href path to dependency file index html path to vulnerable library index html dependency hierarchy x jquery slim min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file index html path to vulnerable library index html dependency hierarchy x jquery min js vulnerable library vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails step up your open source security game with mend | 0 |
255,841 | 21,960,191,982 | IssuesEvent | 2022-05-24 15:12:14 | rstudio/rstudio | https://api.github.com/repos/rstudio/rstudio | closed | Electron font-selection logic | enhancement electron test | The fonts used by Electron are currently hardcoded, e.g. here: https://github.com/rstudio/rstudio/blob/spotted-wakerobin/src/node/desktop/src/renderer/desktop-info-bridge.ts
Need to implement logic for selecting these; at a minimum replicate what we have in the Qt version, but also consider #4134 and #8659.
| 1.0 | Electron font-selection logic - The fonts used by Electron are currently hardcoded, e.g. here: https://github.com/rstudio/rstudio/blob/spotted-wakerobin/src/node/desktop/src/renderer/desktop-info-bridge.ts
Need to implement logic for selecting these; at a minimum replicate what we have in the Qt version, but also consider #4134 and #8659.
| test | electron font selection logic the fonts used by electron are currently hardcoded e g here need to implement logic for selecting these at a minimum replicate what we have in the qt version but also consider and | 1 |
323,570 | 27,736,902,626 | IssuesEvent | 2023-03-15 11:52:31 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | closed | LOGICAL_ERROR: RWLockImpl::getLock(): Cannot acquire exclusive lock while RWLock is already locked | bug comp-joins testing | https://s3.amazonaws.com/clickhouse-test-reports/44879/cc02de69d2cff6b36b9c62bc676f56bdb878f4c8/upgrade_check__debug_.html
Test that caused it: `01502_long_log_tinylog_deadlock_race`
| 1.0 | LOGICAL_ERROR: RWLockImpl::getLock(): Cannot acquire exclusive lock while RWLock is already locked - https://s3.amazonaws.com/clickhouse-test-reports/44879/cc02de69d2cff6b36b9c62bc676f56bdb878f4c8/upgrade_check__debug_.html
Test that caused it: `01502_long_log_tinylog_deadlock_race`
| test | logical error rwlockimpl getlock cannot acquire exclusive lock while rwlock is already locked test that caused it long log tinylog deadlock race | 1 |
5,303 | 2,770,837,775 | IssuesEvent | 2015-05-01 17:24:47 | SunstriderProject/Sunstrider-Issues | https://api.github.com/repos/SunstriderProject/Sunstrider-Issues | closed | SMART_ACTION_TELEPORT | enhancement need test smartai | Make SMART_ACTION_TELEPORT possible to use more target types when Map id (param1) = 0 :
* Param2 : Visual (0/1) (spell 41232)
* SMART_TARGET_VICTIM
* SMART_TARGET_HOSTILE_SECOND_AGGRO
* SMART_TARGET_HOSTILE_LAST_AGGRO
* SMART_TARGET_HOSTILE_RANDOM
* SMART_TARGET_HOSTILE_RANDOM_NOT_TOP
* SMART_TARGET_POSITION | 1.0 | SMART_ACTION_TELEPORT - Make SMART_ACTION_TELEPORT possible to use more target types when Map id (param1) = 0 :
* Param2 : Visual (0/1) (spell 41232)
* SMART_TARGET_VICTIM
* SMART_TARGET_HOSTILE_SECOND_AGGRO
* SMART_TARGET_HOSTILE_LAST_AGGRO
* SMART_TARGET_HOSTILE_RANDOM
* SMART_TARGET_HOSTILE_RANDOM_NOT_TOP
* SMART_TARGET_POSITION | test | smart action teleport make smart action teleport possible to use more target types when map id visual spell smart target victim smart target hostile second aggro smart target hostile last aggro smart target hostile random smart target hostile random not top smart target position | 1 |
7,372 | 6,925,807,773 | IssuesEvent | 2017-11-30 17:00:32 | lqdi/busca-ativa-escolar-web | https://api.github.com/repos/lqdi/busca-ativa-escolar-web | closed | Deletar alerta teste criado no ambiente de produรงรฃo | infrastructure pending-time-tracking | **Deletar alerta teste criado no ambiente de produรงรฃo no municรญpio de Sรฃo Paulo:**

| 1.0 | Deletar alerta teste criado no ambiente de produรงรฃo - **Deletar alerta teste criado no ambiente de produรงรฃo no municรญpio de Sรฃo Paulo:**

| non_test | deletar alerta teste criado no ambiente de produรงรฃo deletar alerta teste criado no ambiente de produรงรฃo no municรญpio de sรฃo paulo | 0 |
398,733 | 27,210,619,425 | IssuesEvent | 2023-02-20 16:14:49 | dirty-cat/dirty_cat | https://api.github.com/repos/dirty-cat/dirty_cat | opened | Get rid of the "scalability considerations..." example | Documentation | We should say that the SimilarityEncoder is not very scalable, and move a few take-home messages of this example elsewhere (in the notes section of the SimilarityEncoder ?).
Maybe we shoulld also remove a few of the corresponding options of the SimilarityEncoder to simplify the code (the value of the SimilarityEncoder is its simplicity, so let's keep it simple) | 1.0 | Get rid of the "scalability considerations..." example - We should say that the SimilarityEncoder is not very scalable, and move a few take-home messages of this example elsewhere (in the notes section of the SimilarityEncoder ?).
Maybe we shoulld also remove a few of the corresponding options of the SimilarityEncoder to simplify the code (the value of the SimilarityEncoder is its simplicity, so let's keep it simple) | non_test | get rid of the scalability considerations example we should say that the similarityencoder is not very scalable and move a few take home messages of this example elsewhere in the notes section of the similarityencoder maybe we shoulld also remove a few of the corresponding options of the similarityencoder to simplify the code the value of the similarityencoder is its simplicity so let s keep it simple | 0 |
255,607 | 8,125,822,992 | IssuesEvent | 2018-08-16 22:29:58 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Add tests for 2D and 3D Line annotations | Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal | Subject says it all.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2337
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Add tests for 2D and 3D Line annotations
Assigned to: Kathleen Biagas
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 07/14/2015
Due date:
% Done: 100
Estimated time:
Created: 07/14/2015 12:53 am
Updated: 08/27/2015 08:22 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
Subject says it all.
Comments:
SVN Revision 26740A tests/rendering/line2d.pyA baseline/rendering/line2d/line2d00.pngA baseline/rendering/line2d/line2d10.pngA baseline/rendering/line2d/line2d01.pngA baseline/rendering/line2d/line2d02.pngA baseline/rendering/line2d/line2d03.pngA baseline/rendering/line2d/line2d04.pngA baseline/rendering/line2d/line2d05.pngA baseline/rendering/line2d/line2d06.pngA baseline/rendering/line2d/line2d07.pngA baseline/rendering/line2d/line2d08.pngA baseline/rendering/line2d/line2d09.pngSVN Revision 27056A baseline/rendering/line3d/line3d00.pngA baseline/rendering/line3d/Line3d01.pngA baseline/rendering/line3d/line3d10.pngA baseline/rendering/line3d/Line3d02.pngA baseline/rendering/line3d/line3d11.pngA baseline/rendering/line3d/Line3d03.pngA baseline/rendering/line3d/line3d12.pngA baseline/rendering/line3d/Line3d04.pngA baseline/rendering/line3d/line3d13.pngA baseline/rendering/line3d/line3d05.pngA baseline/rendering/line3d/line3d06.pngA baseline/rendering/line3d/line3d07.pngA baseline/rendering/line3d/line3d08.pngA baseline/rendering/line3d/line3d09.png
| 1.0 | Add tests for 2D and 3D Line annotations - Subject says it all.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2337
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Add tests for 2D and 3D Line annotations
Assigned to: Kathleen Biagas
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 07/14/2015
Due date:
% Done: 100
Estimated time:
Created: 07/14/2015 12:53 am
Updated: 08/27/2015 08:22 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
Subject says it all.
Comments:
SVN Revision 26740A tests/rendering/line2d.pyA baseline/rendering/line2d/line2d00.pngA baseline/rendering/line2d/line2d10.pngA baseline/rendering/line2d/line2d01.pngA baseline/rendering/line2d/line2d02.pngA baseline/rendering/line2d/line2d03.pngA baseline/rendering/line2d/line2d04.pngA baseline/rendering/line2d/line2d05.pngA baseline/rendering/line2d/line2d06.pngA baseline/rendering/line2d/line2d07.pngA baseline/rendering/line2d/line2d08.pngA baseline/rendering/line2d/line2d09.pngSVN Revision 27056A baseline/rendering/line3d/line3d00.pngA baseline/rendering/line3d/Line3d01.pngA baseline/rendering/line3d/line3d10.pngA baseline/rendering/line3d/Line3d02.pngA baseline/rendering/line3d/line3d11.pngA baseline/rendering/line3d/Line3d03.pngA baseline/rendering/line3d/line3d12.pngA baseline/rendering/line3d/Line3d04.pngA baseline/rendering/line3d/line3d13.pngA baseline/rendering/line3d/line3d05.pngA baseline/rendering/line3d/line3d06.pngA baseline/rendering/line3d/line3d07.pngA baseline/rendering/line3d/line3d08.pngA baseline/rendering/line3d/line3d09.png
| non_test | add tests for and line annotations subject says it all redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject add tests for and line annotations assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created am updated pm likelihood severity found in version impact medium expected use occasional os all support group any description subject says it all comments svn revision tests rendering pya baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pngsvn revision baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering pnga baseline rendering png | 0 |
303,703 | 26,224,409,842 | IssuesEvent | 2023-01-04 17:19:42 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | `Scroll for Brave News` is NOT shown in NTP for users in OFAC list countries | feature/rewards QA/Yes QA/Test-Plan-Specified OS/Desktop | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
`Scroll for Brave News` is NOT shown in NTP for users in OFAC list countries
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Follow the steps 1 to 10 from Test case 1 in https://docs.google.com/document/d/1zlWGAhk0X5nFM5vwwZ21LpDfgPX-6m0-oGcCilsGtXk/edit#
2. `Scroll for Brave News` is NOT shown in NTP for users in OFAC list countries
**Note:** User can enable Brave news via `Customize` in NTP
## Actual result:
<!--Please add screenshots if needed-->
`Scroll for Brave News` is NOT shown in NTP for users in OFAC list countries
## Expected result:
`Scroll for Brave News` should be shown in NTP for users in OFAC list countries
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easy
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave 1.47.161 Chromium: 109.0.5414.61 (Official Build) beta (64-bit)
Revision 846c6eff1c374cde417a9cdc44193545cc07f354-refs/branch-heads/5414@{#1030}
OS Windows 10 Version 21H2 (Build 19044.1826)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? NA
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc : @brave/qa-team @mkarolin @Miyayes | 1.0 | `Scroll for Brave News` is NOT shown in NTP for users in OFAC list countries - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
`Scroll for Brave News` is NOT shown in NTP for users in OFAC list countries
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Follow the steps 1 to 10 from Test case 1 in https://docs.google.com/document/d/1zlWGAhk0X5nFM5vwwZ21LpDfgPX-6m0-oGcCilsGtXk/edit#
2. `Scroll for Brave News` is NOT shown in NTP for users in OFAC list countries
**Note:** User can enable Brave news via `Customize` in NTP
## Actual result:
<!--Please add screenshots if needed-->
`Scroll for Brave News` is NOT shown in NTP for users in OFAC list countries
## Expected result:
`Scroll for Brave News` should be shown in NTP for users in OFAC list countries
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easy
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave 1.47.161 Chromium: 109.0.5414.61 (Official Build) beta (64-bit)
Revision 846c6eff1c374cde417a9cdc44193545cc07f354-refs/branch-heads/5414@{#1030}
OS Windows 10 Version 21H2 (Build 19044.1826)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? NA
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc : @brave/qa-team @mkarolin @Miyayes | test | scroll for brave news is not shown in ntp for users in ofac list countries have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description scroll for brave news is not shown in ntp for users in ofac list countries steps to reproduce follow the steps to from test case in scroll for brave news is not shown in ntp for users in ofac list countries note user can enable brave news via customize in ntp actual result scroll for brave news is not shown in ntp for users in ofac list countries expected result scroll for brave news should be shown in ntp for users in ofac list countries reproduces how often easy brave version brave version info brave chromium official build beta bit revision refs branch heads os windows version build version channel information can you reproduce this issue with the current release na can you reproduce this issue with the beta channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc brave qa team mkarolin miyayes | 1 |
78,058 | 7,613,524,848 | IssuesEvent | 2018-05-01 21:34:06 | infiniteautomation/ma-core-public | https://api.github.com/repos/infiniteautomation/ma-core-public | reopened | Cleanup of SystemSettingsDao | Ready for Testing | Some methods are static, some are not - make consistent, all methods should be called on an instance of SystemSettingsDao.
Mark get() methods with default parameters as deprecated to ensure that defaults are added to the DEFAULT_VALUES map. | 1.0 | Cleanup of SystemSettingsDao - Some methods are static, some are not - make consistent, all methods should be called on an instance of SystemSettingsDao.
Mark get() methods with default parameters as deprecated to ensure that defaults are added to the DEFAULT_VALUES map. | test | cleanup of systemsettingsdao some methods are static some are not make consistent all methods should be called on an instance of systemsettingsdao mark get methods with default parameters as deprecated to ensure that defaults are added to the default values map | 1 |
272,821 | 23,705,731,629 | IssuesEvent | 2022-08-30 00:45:45 | PathmindAI/nativerl | https://api.github.com/repos/PathmindAI/nativerl | closed | NativeRL pr build tests interfere with each other | tests | The jenkins builds all copy their files to s3 at this location s3://dev-training-static-files.pathmind.com/nativerl/pr_test/nativerl-1.8.1-SNAPSHOT-bin.zip
so if there are multiple tests going, it's not clear which is being tested. | 1.0 | NativeRL pr build tests interfere with each other - The jenkins builds all copy their files to s3 at this location s3://dev-training-static-files.pathmind.com/nativerl/pr_test/nativerl-1.8.1-SNAPSHOT-bin.zip
so if there are multiple tests going, it's not clear which is being tested. | test | nativerl pr build tests interfere with each other the jenkins builds all copy their files to at this location dev training static files pathmind com nativerl pr test nativerl snapshot bin zip so if there are multiple tests going it s not clear which is being tested | 1 |
246,566 | 20,884,210,495 | IssuesEvent | 2022-03-23 02:00:20 | kal-aalrajhi/flashcards-starter | https://api.github.com/repos/kal-aalrajhi/flashcards-starter | opened | Check that every test has a happy and sad path | enhancement test high priority | Make sure that every test that has a possible sad path is accounted for. | 1.0 | Check that every test has a happy and sad path - Make sure that every test that has a possible sad path is accounted for. | test | check that every test has a happy and sad path make sure that every test that has a possible sad path is accounted for | 1 |
44,368 | 7,106,954,658 | IssuesEvent | 2018-01-16 18:16:01 | Caoimhinmg/PySkew | https://api.github.com/repos/Caoimhinmg/PySkew | opened | Example Jupyter Notebooks | Documentation New Feature | I want to make some example Jupyter notebooks on how to generate pole space plots, how to use the built-in functions that make up the scripts, and how to use the object oriented rotation code Rot which I plan to integrate soon. | 1.0 | Example Jupyter Notebooks - I want to make some example Jupyter notebooks on how to generate pole space plots, how to use the built-in functions that make up the scripts, and how to use the object oriented rotation code Rot which I plan to integrate soon. | non_test | example jupyter notebooks i want to make some example jupyter notebooks on how to generate pole space plots how to use the built in functions that make up the scripts and how to use the object oriented rotation code rot which i plan to integrate soon | 0 |
403,273 | 11,838,395,485 | IssuesEvent | 2020-03-23 15:35:38 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | Block submission pages are using the entire Submission logs as page locators | component: admin tools priority: p4 state: pull request ready | Follow up for #13201
### Describe the problem and steps to reproduce it:
1. Log into the Admin Tolls
2. Open the Block submissions module and select one of the entries to inspect - see https://addons-internal.dev.mozaws.net/en-US/admin/models/blocklist/blocksubmission/20/change/
3. Look in the navigation bar (breadcrumbs) and identify the current page locator
### What happened?
The page locator uses the entire 'Submission Logs' data to identify the current page, which makes the title extremely long
### What did you expect to happen?
Block submission pages should use a different locator
### Anything else we should know?
- the same locator is used in success messages, in block history etc
[breadcrumbs]

[success message]

[block history]

| 1.0 | Block submission pages are using the entire Submission logs as page locators - Follow up for #13201
### Describe the problem and steps to reproduce it:
1. Log into the Admin Tolls
2. Open the Block submissions module and select one of the entries to inspect - see https://addons-internal.dev.mozaws.net/en-US/admin/models/blocklist/blocksubmission/20/change/
3. Look in the navigation bar (breadcrumbs) and identify the current page locator
### What happened?
The page locator uses the entire 'Submission Logs' data to identify the current page, which makes the title extremely long
### What did you expect to happen?
Block submission pages should use a different locator
### Anything else we should know?
- the same locator is used in success messages, in block history etc
[breadcrumbs]

[success message]

[block history]

| non_test | block submission pages are using the entire submission logs as page locators follow up for describe the problem and steps to reproduce it log into the admin tolls open the block submissions module and select one of the entries to inspect see look in the navigation bar breadcrumbs and identify the current page locator what happened the page locator uses the entire submission logs data to identify the current page which makes the title extremely long what did you expect to happen block submission pages should use a different locator anything else we should know the same locator is used in success messages in block history etc | 0 |
159,092 | 20,036,638,202 | IssuesEvent | 2022-02-02 12:37:50 | kapseliboi/webuild | https://api.github.com/repos/kapseliboi/webuild | opened | CVE-2021-23343 (High) detected in path-parse-1.0.5.tgz | security vulnerability | ## CVE-2021-23343 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.5.tgz</b></p></summary>
<p>Node.js path.parse() ponyfill</p>
<p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.5.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/path-parse/package.json</p>
<p>
Dependency Hierarchy:
- grunt-jscs-3.0.1.tgz (Root Library)
- jscs-3.0.7.tgz
- resolve-1.4.0.tgz
- :x: **path-parse-1.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/webuild/commit/cfe337e0888ede79847ee4201c96d43ab06dc89e">cfe337e0888ede79847ee4201c96d43ab06dc89e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: path-parse - 1.0.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23343 (High) detected in path-parse-1.0.5.tgz - ## CVE-2021-23343 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.5.tgz</b></p></summary>
<p>Node.js path.parse() ponyfill</p>
<p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.5.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/path-parse/package.json</p>
<p>
Dependency Hierarchy:
- grunt-jscs-3.0.1.tgz (Root Library)
- jscs-3.0.7.tgz
- resolve-1.4.0.tgz
- :x: **path-parse-1.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/webuild/commit/cfe337e0888ede79847ee4201c96d43ab06dc89e">cfe337e0888ede79847ee4201c96d43ab06dc89e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: path-parse - 1.0.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in path parse tgz cve high severity vulnerability vulnerable library path parse tgz node js path parse ponyfill library home page a href path to dependency file package json path to vulnerable library node modules path parse package json dependency hierarchy grunt jscs tgz root library jscs tgz resolve tgz x path parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution path parse step up your open source security game with whitesource | 0 |
223,354 | 7,452,800,026 | IssuesEvent | 2018-03-29 09:36:57 | Nashet/EconomicSimulation | https://api.github.com/repos/Nashet/EconomicSimulation | closed | Help player to do foreign investments | enhancement in priority | - [ ] Show all factories button
- [ ] keep some money for player after conquering | 1.0 | Help player to do foreign investments - - [ ] Show all factories button
- [ ] keep some money for player after conquering | non_test | help player to do foreign investments show all factories button keep some money for player after conquering | 0 |
9,898 | 11,903,203,752 | IssuesEvent | 2020-03-30 15:00:55 | ionic-team/capacitor | https://api.github.com/repos/ionic-team/capacitor | closed | bug: add cordova plugin "jpush-phonegap-plugin", IOS app crashed | known incompatible cordova plugin | # Bug Report
Hi, there is still a problem with plugin "jpush-phonegap-plugin", since I download this commit change to replace my local files which is "https://github.com/ionic-team/capacitor/pull/2636", Xcode build success width no errors, but it can not running on my phone, the app crashed
# Capacitor Version
<!-- Provide the version of Capacitor and related installed dependencies. You can use the command `npx capacitor doctor` or `npx cap doctor` for the output from the root directory of your project. -->
`npx cap doctor` output:
```
Latest Dependencies:
@capacitor/cli: 1.5.2
@capacitor/core: 1.5.2
@capacitor/android: 1.5.2
@capacitor/electron: 1.5.2
@capacitor/ios: 1.5.2
Installed Dependencies:
@capacitor/android not installed
@capacitor/electron not installed
@capacitor/cli 1.5.2
@capacitor/ios 1.5.2
@capacitor/core 1.5.2
Found 3 Capacitor plugins for ios:
cordova-plugin-device (2.0.3)
cordova-plugin-jcore (1.3.1)
jpush-phonegap-plugin (3.7.3)
[success] iOS looking great! ๐
```
## Affected Platform(s)
<!-- Leave these checkboxes EMPTY until after the issue is created. Once the issue is created, check one or more boxes for the platform(s) this bug affects. -->
- [ ] Android
- [x] iOS
- [ ] Electron
- [ ] Web
## Current Behavior
<!-- Describe how the bug manifests. Be specific. -->
Xcode build success, but app crashed.
## Expected Behavior
app can running normal
<!-- Describe what the behavior would be without the bug. -->
## Sample Code or Sample Application Repo
there is sample code https://github.com/AlwaysLoveme/demo_.git
<!--
NOTE: Issues with sample projects are prioritized higher than ones without because they are easier to reproduce and fix.
If you are able to illustrate the bug or feature request with an example, please provide sample code snippets or a sample application via a public GitHub or Bitbucket repo.
-->
## Reproduction Steps
1. create a blank project with "@ionic/cli"
2. add cordova plugin "npm i jpush-phonegap-plugin"
3. download "https://github.com/ionic-team/capacitor.git" files which is included ""fix(cordova): Add lib prefix to .a library names (#2636)", and build @capacitor/cli, then copy dist to replace local file
4. run "ionic build --prod && ionic cap add ios && ionic cap open ios"
5. Xcode running
<!-- Please explain the steps required to duplicate the issue, especially if you are able to provide a sample application. -->
## Other Technical Details
<!-- Please provide the following information with your request, and any other relevant technical details (versions of IDEs, local environment info, plugin information or links, etc). -->
`npm --version` output: 6.14.4
`node --version` output: 12.16.1
`pod --version` output (iOS issues only): 1.9.1
## Other Information
<!-- List any other information that is relevant to your issue. Stack traces, related issues, suggestions on how to fix, Stack Overflow links, forum links, etc. -->
#2632 | True | bug: add cordova plugin "jpush-phonegap-plugin", IOS app crashed - # Bug Report
Hi, there is still a problem with plugin "jpush-phonegap-plugin", since I download this commit change to replace my local files which is "https://github.com/ionic-team/capacitor/pull/2636", Xcode build success width no errors, but it can not running on my phone, the app crashed
# Capacitor Version
<!-- Provide the version of Capacitor and related installed dependencies. You can use the command `npx capacitor doctor` or `npx cap doctor` for the output from the root directory of your project. -->
`npx cap doctor` output:
```
Latest Dependencies:
@capacitor/cli: 1.5.2
@capacitor/core: 1.5.2
@capacitor/android: 1.5.2
@capacitor/electron: 1.5.2
@capacitor/ios: 1.5.2
Installed Dependencies:
@capacitor/android not installed
@capacitor/electron not installed
@capacitor/cli 1.5.2
@capacitor/ios 1.5.2
@capacitor/core 1.5.2
Found 3 Capacitor plugins for ios:
cordova-plugin-device (2.0.3)
cordova-plugin-jcore (1.3.1)
jpush-phonegap-plugin (3.7.3)
[success] iOS looking great! ๐
```
## Affected Platform(s)
<!-- Leave these checkboxes EMPTY until after the issue is created. Once the issue is created, check one or more boxes for the platform(s) this bug affects. -->
- [ ] Android
- [x] iOS
- [ ] Electron
- [ ] Web
## Current Behavior
<!-- Describe how the bug manifests. Be specific. -->
Xcode build success, but app crashed.
## Expected Behavior
app can running normal
<!-- Describe what the behavior would be without the bug. -->
## Sample Code or Sample Application Repo
there is sample code https://github.com/AlwaysLoveme/demo_.git
<!--
NOTE: Issues with sample projects are prioritized higher than ones without because they are easier to reproduce and fix.
If you are able to illustrate the bug or feature request with an example, please provide sample code snippets or a sample application via a public GitHub or Bitbucket repo.
-->
## Reproduction Steps
1. create a blank project with "@ionic/cli"
2. add cordova plugin "npm i jpush-phonegap-plugin"
3. download "https://github.com/ionic-team/capacitor.git" files which is included ""fix(cordova): Add lib prefix to .a library names (#2636)", and build @capacitor/cli, then copy dist to replace local file
4. run "ionic build --prod && ionic cap add ios && ionic cap open ios"
5. Xcode running
<!-- Please explain the steps required to duplicate the issue, especially if you are able to provide a sample application. -->
## Other Technical Details
<!-- Please provide the following information with your request, and any other relevant technical details (versions of IDEs, local environment info, plugin information or links, etc). -->
`npm --version` output: 6.14.4
`node --version` output: 12.16.1
`pod --version` output (iOS issues only): 1.9.1
## Other Information
<!-- List any other information that is relevant to your issue. Stack traces, related issues, suggestions on how to fix, Stack Overflow links, forum links, etc. -->
#2632 | non_test | bug add cordova plugin jpush phonegap plugin ios app crashed bug report hi there is still a problem with plugin jpush phonegap plugin since i download this commit change to replace my local files which is xcode build success width no errors but it can not running on my phone the app crashed capacitor version npx cap doctor output latest dependencies capacitor cli capacitor core capacitor android capacitor electron capacitor ios installed dependencies capacitor android not installed capacitor electron not installed capacitor cli capacitor ios capacitor core found capacitor plugins for ios cordova plugin device cordova plugin jcore jpush phonegap plugin ios looking great ๐ affected platform s android ios electron web current behavior xcode build success but app crashed expected behavior app can running normal sample code or sample application repo there is sample code note issues with sample projects are prioritized higher than ones without because they are easier to reproduce and fix if you are able to illustrate the bug or feature request with an example please provide sample code snippets or a sample application via a public github or bitbucket repo reproduction steps create a blank project with ionic cli add cordova plugin npm i jpush phonegap plugin download files which is included fix cordova add lib prefix to a library names and build capacitor cli then copy dist to replace local file run ionic build prod ionic cap add ios ionic cap open ios xcode running other technical details npm version output node version output pod version output ios issues only other information | 0 |
288,605 | 21,714,160,823 | IssuesEvent | 2022-05-10 16:13:46 | amosproj/amos2022ss03-turtlebot-fleet-management | https://api.github.com/repos/amosproj/amos2022ss03-turtlebot-fleet-management | opened | Evaluate & specify FMS Backend - Frontend interface | documentation | ## User story
1. As a SD
2. I want to brainstorm on the interface between the backend and frontend of the fleet management software
3. So that have a better idea on how these components communicate.
## Acceptance criteria
* Good understanding of the interface between specified components.
* Documenting a standard on how the components communicate, aka create API docs.
## Definition of done (DoD)
* Added only after week 5
* The same for all features
* Here goes the project specific part
## DoD general criteria
* Feature has been fully implemented
* Feature has been merged into the mainline
* All acceptance criteria were met
* Product owner approved features
* All tests are passing
* Developers agreed to release
| 1.0 | Evaluate & specify FMS Backend - Frontend interface - ## User story
1. As a SD
2. I want to brainstorm on the interface between the backend and frontend of the fleet management software
3. So that have a better idea on how these components communicate.
## Acceptance criteria
* Good understanding of the interface between specified components.
* Documenting a standard on how the components communicate, aka create API docs.
## Definition of done (DoD)
* Added only after week 5
* The same for all features
* Here goes the project specific part
## DoD general criteria
* Feature has been fully implemented
* Feature has been merged into the mainline
* All acceptance criteria were met
* Product owner approved features
* All tests are passing
* Developers agreed to release
| non_test | evaluate specify fms backend frontend interface user story as a sd i want to brainstorm on the interface between the backend and frontend of the fleet management software so that have a better idea on how these components communicate acceptance criteria good understanding of the interface between specified components documenting a standard on how the components communicate aka create api docs definition of done dod added only after week the same for all features here goes the project specific part dod general criteria feature has been fully implemented feature has been merged into the mainline all acceptance criteria were met product owner approved features all tests are passing developers agreed to release | 0 |
52,305 | 6,226,772,249 | IssuesEvent | 2017-07-10 19:12:15 | ProjectSidewalk/SidewalkWebpage | https://api.github.com/repos/ProjectSidewalk/SidewalkWebpage | closed | FAQ Page Off Center | EasyFix! Priority: Low pull-request-submitted Relaunch Testing | Not sure if this is a big problem or not, but the bottom of the FAQ page seems to be a little off center, this should be a quick fix though.

| 1.0 | FAQ Page Off Center - Not sure if this is a big problem or not, but the bottom of the FAQ page seems to be a little off center, this should be a quick fix though.

| test | faq page off center not sure if this is a big problem or not but the bottom of the faq page seems to be a little off center this should be a quick fix though | 1 |
101,800 | 21,787,758,226 | IssuesEvent | 2022-05-14 12:13:49 | cactool/cactool | https://api.github.com/repos/cactool/cactool | opened | Refactor server-side embed rendering | code design | The code for rendering embeds on the server needs to be refactored, mainly due to the code reuse | 1.0 | Refactor server-side embed rendering - The code for rendering embeds on the server needs to be refactored, mainly due to the code reuse | non_test | refactor server side embed rendering the code for rendering embeds on the server needs to be refactored mainly due to the code reuse | 0 |
147,488 | 13,208,583,419 | IssuesEvent | 2020-08-15 05:48:06 | TesseractCoding/NeoAlgo | https://api.github.com/repos/TesseractCoding/NeoAlgo | opened | Readme for Java folder | documentation | ## ๐ Documentation
Adding content into Readme file in Java folder
### Have you read the [Contributing Guidelines on Pull Requests](https://github.com/TesseractCoding/NeoAlgo/blob/master/CONTRIBUTING.md#reporting-new-issues)?
Yes
| 1.0 | Readme for Java folder - ## ๐ Documentation
Adding content into Readme file in Java folder
### Have you read the [Contributing Guidelines on Pull Requests](https://github.com/TesseractCoding/NeoAlgo/blob/master/CONTRIBUTING.md#reporting-new-issues)?
Yes
| non_test | readme for java folder ๐ documentation adding content into readme file in java folder have you read the yes | 0 |
198,995 | 15,730,707,601 | IssuesEvent | 2021-03-29 16:13:08 | dfinity/docs | https://api.github.com/repos/dfinity/docs | opened | CLI: no-wallet description and examples | documentation | ## Describe the command or option
## Add a usage scenario and example**
| 1.0 | CLI: no-wallet description and examples - ## Describe the command or option
## Add a usage scenario and example**
| non_test | cli no wallet description and examples describe the command or option add a usage scenario and example | 0 |
249,468 | 21,162,328,455 | IssuesEvent | 2022-04-07 10:29:37 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [Security Solution] The search bar have different placeholder values for Blocklist tab and Blocklist page under Policy tab | bug triage_needed impact:low Team: SecuritySolution Team:Onboarding and Lifecycle Mgt QA:Ready for Testing OLM Sprint v8.2.0 | **Description:**
The search bar have different placeholder values for Blocklist tab and Blocklist page under Policy tab
**Build Details:**
```
VERSION: 8.2.0 SNAPSHOT
BUILD: 51336
COMMIT: 354cd01ecbfbac7b38a98ded6e15babdbae79a16
```
**Browser Details:**
All
**Preconditions:**
1. Kibana user should be logged in.
**Steps to Reproduce:**
1. Navigate to Endpoints Tab from the left hand side navigation
2. Go to the Blocklist tab
**Impacted Test case:**
N/A
**Actual Result:**
The search bar have different placeholder values for Blocklist tab and Blocklist page under Policy tab
**Expected Result:**
The placeholder values for Blocklist tab and Blocklist page under Policy tab should be consistent
**What's working:**
N/A
**What's not working:**
N/A
**Screenshots:**


**Screen Recording:**
https://user-images.githubusercontent.com/60252716/159623309-da37eb05-c523-4ed2-a49b-5ce2eca79c56.mp4
**Logs:**
N/A | 1.0 | [Security Solution] The search bar have different placeholder values for Blocklist tab and Blocklist page under Policy tab - **Description:**
The search bar have different placeholder values for Blocklist tab and Blocklist page under Policy tab
**Build Details:**
```
VERSION: 8.2.0 SNAPSHOT
BUILD: 51336
COMMIT: 354cd01ecbfbac7b38a98ded6e15babdbae79a16
```
**Browser Details:**
All
**Preconditions:**
1. Kibana user should be logged in.
**Steps to Reproduce:**
1. Navigate to Endpoints Tab from the left hand side navigation
2. Go to the Blocklist tab
**Impacted Test case:**
N/A
**Actual Result:**
The search bar have different placeholder values for Blocklist tab and Blocklist page under Policy tab
**Expected Result:**
The placeholder values for Blocklist tab and Blocklist page under Policy tab should be consistent
**What's working:**
N/A
**What's not working:**
N/A
**Screenshots:**


**Screen Recording:**
https://user-images.githubusercontent.com/60252716/159623309-da37eb05-c523-4ed2-a49b-5ce2eca79c56.mp4
**Logs:**
N/A | test | the search bar have different placeholder values for blocklist tab and blocklist page under policy tab description the search bar have different placeholder values for blocklist tab and blocklist page under policy tab build details version snapshot build commit browser details all preconditions kibana user should be logged in steps to reproduce navigate to endpoints tab from the left hand side navigation go to the blocklist tab impacted test case n a actual result the search bar have different placeholder values for blocklist tab and blocklist page under policy tab expected result the placeholder values for blocklist tab and blocklist page under policy tab should be consistent what s working n a what s not working n a screenshots screen recording logs n a | 1 |
95,600 | 8,564,521,914 | IssuesEvent | 2018-11-09 16:58:03 | hammerlab/genspio | https://api.github.com/repos/hammerlab/genspio | closed | Example: git meta-status | tests |
- find all `.git/`s
- get status, untracked files, unpushed branches
- display in a nice way
| 1.0 | Example: git meta-status -
- find all `.git/`s
- get status, untracked files, unpushed branches
- display in a nice way
| test | example git meta status find all git s get status untracked files unpushed branches display in a nice way | 1 |
7,613 | 7,018,862,966 | IssuesEvent | 2017-12-21 15:14:13 | procxx/kepka | https://api.github.com/repos/procxx/kepka | opened | Fix warnings | enhancement good first task help wanted infrastructure refactoring | Currently, we have a ton of warnings. On some distros it interprets like errors. We need to fix these warnings.
Some warnings are simple and could be fixed automatically with brand new C++17 features like [[maybe-unused]], [[fallthrough]], some warnings related to Qt/stdint types collision (see bug #18 , for example). | 1.0 | Fix warnings - Currently, we have a ton of warnings. On some distros it interprets like errors. We need to fix these warnings.
Some warnings are simple and could be fixed automatically with brand new C++17 features like [[maybe-unused]], [[fallthrough]], some warnings related to Qt/stdint types collision (see bug #18 , for example). | non_test | fix warnings currently we have a ton of warnings on some distros it interprets like errors we need to fix these warnings some warnings are simple and could be fixed automatically with brand new c features like some warnings related to qt stdint types collision see bug for example | 0 |
257,872 | 22,240,993,373 | IssuesEvent | 2022-06-09 05:14:10 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | opened | [Flaky Test] creates a group from multiple blocks of different types via block transforms | [Type] Flaky Test | <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
creates a group from multiple blocks of different types via block transforms
## Test path
`specs/editor/various/block-grouping.test.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2022-06-09T05:14:09.913Z"><code>[2022-06-09T05:14:09.913Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/2466014484"><code>fix/accessibility-template-buttons</code></a>.
</summary>
```
โ Block Grouping โบ Group creation โบ creates a group from multiple blocks of different types via block transforms
expect(jest.fn()).not.toHaveErrored(expected)
Expected mock function not to be called but it was called with:
["Failed to load resource: the server responded with a status of 500 (Internal Server Error)"]
at Object.assertExpectedCalls (../jest-console/build/@wordpress/jest-console/src/index.js:36:4)
at runMicrotasks (<anonymous>)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| 1.0 | [Flaky Test] creates a group from multiple blocks of different types via block transforms - <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
creates a group from multiple blocks of different types via block transforms
## Test path
`specs/editor/various/block-grouping.test.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2022-06-09T05:14:09.913Z"><code>[2022-06-09T05:14:09.913Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/2466014484"><code>fix/accessibility-template-buttons</code></a>.
</summary>
```
โ Block Grouping โบ Group creation โบ creates a group from multiple blocks of different types via block transforms
expect(jest.fn()).not.toHaveErrored(expected)
Expected mock function not to be called but it was called with:
["Failed to load resource: the server responded with a status of 500 (Internal Server Error)"]
at Object.assertExpectedCalls (../jest-console/build/@wordpress/jest-console/src/index.js:36:4)
at runMicrotasks (<anonymous>)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| test | creates a group from multiple blocks of different types via block transforms flaky test detected this is an auto generated issue by github actions please do not edit this manually test title creates a group from multiple blocks of different types via block transforms test path specs editor various block grouping test js errors test passed after failed attempt on a href โ block grouping โบ group creation โบ creates a group from multiple blocks of different types via block transforms expect jest fn not tohaveerrored expected expected mock function not to be called but it was called with at object assertexpectedcalls jest console build wordpress jest console src index js at runmicrotasks | 1 |
225,651 | 24,881,052,404 | IssuesEvent | 2022-10-28 01:08:34 | TERMINALSERVERORDERLY/github-services | https://api.github.com/repos/TERMINALSERVERORDERLY/github-services | closed | WS-2022-0334 (Medium) detected in nokogiri-1.8.1.gem - autoclosed | security vulnerability | ## WS-2022-0334 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.8.1.gem</b></p></summary>
<p>Nokogiri (้ธ) is an HTML, XML, SAX, and Reader parser. Among
Nokogiri's many features is the ability to search documents via XPath
or CSS3 selectors.</p>
<p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.8.1.gem">https://rubygems.org/gems/nokogiri-1.8.1.gem</a></p>
<p>Path to vulnerable library: /vendor/cache/nokogiri-1.8.1.gem</p>
<p>
Dependency Hierarchy:
- :x: **nokogiri-1.8.1.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
nokogiri up to and including 1.13.8 is affected by several vulnerabilities (CVE-2022-40303, CVE-2022-40304 and CVE-2022-2309) in the dependency bundled libxml2 library. Version 1.13.9 of nokogiri contains a patch where the dependency is upgraded with the patches as well.
<p>Publish Date: 2022-10-18
<p>URL: <a href=https://github.com/sparklemotion/nokogiri/commit/e8cfe13953c63099f879d8a25ca70a909e19fb96>WS-2022-0334</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-2qc6-mcvw-92cw">https://github.com/advisories/GHSA-2qc6-mcvw-92cw</a></p>
<p>Release Date: 2022-10-18</p>
<p>Fix Resolution: nokogiri - 1.13.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2022-0334 (Medium) detected in nokogiri-1.8.1.gem - autoclosed - ## WS-2022-0334 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.8.1.gem</b></p></summary>
<p>Nokogiri (้ธ) is an HTML, XML, SAX, and Reader parser. Among
Nokogiri's many features is the ability to search documents via XPath
or CSS3 selectors.</p>
<p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.8.1.gem">https://rubygems.org/gems/nokogiri-1.8.1.gem</a></p>
<p>Path to vulnerable library: /vendor/cache/nokogiri-1.8.1.gem</p>
<p>
Dependency Hierarchy:
- :x: **nokogiri-1.8.1.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
nokogiri up to and including 1.13.8 is affected by several vulnerabilities (CVE-2022-40303, CVE-2022-40304 and CVE-2022-2309) in the dependency bundled libxml2 library. Version 1.13.9 of nokogiri contains a patch where the dependency is upgraded with the patches as well.
<p>Publish Date: 2022-10-18
<p>URL: <a href=https://github.com/sparklemotion/nokogiri/commit/e8cfe13953c63099f879d8a25ca70a909e19fb96>WS-2022-0334</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-2qc6-mcvw-92cw">https://github.com/advisories/GHSA-2qc6-mcvw-92cw</a></p>
<p>Release Date: 2022-10-18</p>
<p>Fix Resolution: nokogiri - 1.13.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws medium detected in nokogiri gem autoclosed ws medium severity vulnerability vulnerable library nokogiri gem nokogiri ้ธ is an html xml sax and reader parser among nokogiri s many features is the ability to search documents via xpath or selectors library home page a href path to vulnerable library vendor cache nokogiri gem dependency hierarchy x nokogiri gem vulnerable library vulnerability details nokogiri up to and including is affected by several vulnerabilities cve cve and cve in the dependency bundled library version of nokogiri contains a patch where the dependency is upgraded with the patches as well publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution nokogiri step up your open source security game with mend | 0 |
652,897 | 21,564,454,360 | IssuesEvent | 2022-05-01 17:01:05 | canonical/hotsos | https://api.github.com/repos/canonical/hotsos | closed | Not clear what action needs to be taken for `unbalanced crush buckets` warning | plugin:storage priority:HIGH | We've got the following warning from hotsos. However, the description wasn't crystal clear what needs to be checked in the crush map nor what action needs to be taken to remediate the situation (if it's not a false positive).
```
potential-issues:
CephCrushWarnings:
- unbalanced crush buckets identified in CRUSH root 'default' using failure
domain 'rack'. Affected CRUSH rule id is '0'. This can cause data distribution
to become skewed - please check crush map (origin=storage.ceph_cluster_checks)
``` | 1.0 | Not clear what action needs to be taken for `unbalanced crush buckets` warning - We've got the following warning from hotsos. However, the description wasn't crystal clear what needs to be checked in the crush map nor what action needs to be taken to remediate the situation (if it's not a false positive).
```
potential-issues:
CephCrushWarnings:
- unbalanced crush buckets identified in CRUSH root 'default' using failure
domain 'rack'. Affected CRUSH rule id is '0'. This can cause data distribution
to become skewed - please check crush map (origin=storage.ceph_cluster_checks)
``` | non_test | not clear what action needs to be taken for unbalanced crush buckets warning we ve got the following warning from hotsos however the description wasn t crystal clear what needs to be checked in the crush map nor what action needs to be taken to remediate the situation if it s not a false positive potential issues cephcrushwarnings unbalanced crush buckets identified in crush root default using failure domain rack affected crush rule id is this can cause data distribution to become skewed please check crush map origin storage ceph cluster checks | 0 |
251,571 | 21,511,072,132 | IssuesEvent | 2022-04-28 04:36:19 | stores-cedcommerce/Karan-Patel-wholesale---Internal--May-4th | https://api.github.com/repos/stores-cedcommerce/Karan-Patel-wholesale---Internal--May-4th | closed | In product page, the title You may also like section the font size have to be small other titles font size is used in the store in product page. | Product page Desktop Ready to test fixed | **Actual result:**
In product page, the title You may also like section the font size have to be small other titles font size is used in the store in product page
The title you may also like have to come in the centre it will be much better.

**Expected result:**
Font size of the title have to improved and the title have to come in the centre For You may also like section it will much better.
| 1.0 | In product page, the title You may also like section the font size have to be small other titles font size is used in the store in product page. - **Actual result:**
In product page, the title You may also like section the font size have to be small other titles font size is used in the store in product page
The title you may also like have to come in the centre it will be much better.

**Expected result:**
Font size of the title have to improved and the title have to come in the centre For You may also like section it will much better.
| test | in product page the title you may also like section the font size have to be small other titles font size is used in the store in product page actual result in product page the title you may also like section the font size have to be small other titles font size is used in the store in product page the title you may also like have to come in the centre it will be much better expected result font size of the title have to improved and the title have to come in the centre for you may also like section it will much better | 1 |
202,337 | 15,281,282,113 | IssuesEvent | 2021-02-23 07:52:47 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Flaky Test [Build&Test / auditbeat-macos-macosx / github.com/elastic/beats/v7/auditbeat/module/file_integrity.TestActions] | Team:SIEM flaky-test | This one fails for `macos` too:
## Flaky Test
* **Test Name:** auditbeat/module/file_integrity.TestActions
* **Link:** https://github.com/elastic/beats/blob/master/auditbeat/module/file_integrity/metricset_test.go#L173
* **Branch:** master
* **Artifact Link:** https://beats-ci.elastic.co/job/Beats/job/beats/job/master/lastFailedBuild/testReport/junit/github/com_elastic_beats_v7_auditbeat_module_file_integrity/Build_Test___auditbeat_macos_macosx___TestActions/
* **Notes:** macosx
### Stack Trace
```
=== RUN TestActions
metricset_test.go:173:
Error Trace: metricset_test.go:173
Error: []string{"created", "updated"} does not contain "attributes_modified"
Test: TestActions
--- FAIL: TestActions (0.13s)
```
Related to https://github.com/elastic/beats/issues/22518 | 1.0 | Flaky Test [Build&Test / auditbeat-macos-macosx / github.com/elastic/beats/v7/auditbeat/module/file_integrity.TestActions] - This one fails for `macos` too:
## Flaky Test
* **Test Name:** auditbeat/module/file_integrity.TestActions
* **Link:** https://github.com/elastic/beats/blob/master/auditbeat/module/file_integrity/metricset_test.go#L173
* **Branch:** master
* **Artifact Link:** https://beats-ci.elastic.co/job/Beats/job/beats/job/master/lastFailedBuild/testReport/junit/github/com_elastic_beats_v7_auditbeat_module_file_integrity/Build_Test___auditbeat_macos_macosx___TestActions/
* **Notes:** macosx
### Stack Trace
```
=== RUN TestActions
metricset_test.go:173:
Error Trace: metricset_test.go:173
Error: []string{"created", "updated"} does not contain "attributes_modified"
Test: TestActions
--- FAIL: TestActions (0.13s)
```
Related to https://github.com/elastic/beats/issues/22518 | test | flaky test this one fails for macos too flaky test test name auditbeat module file integrity testactions link branch master artifact link notes macosx stack trace run testactions metricset test go error trace metricset test go error string created updated does not contain attributes modified test testactions fail testactions related to | 1 |
70,486 | 15,085,816,132 | IssuesEvent | 2021-02-05 19:17:25 | mthbernardes/shaggy-rogers | https://api.github.com/repos/mthbernardes/shaggy-rogers | closed | CVE-2019-13990 (High) detected in quartz-2.2.0.jar - autoclosed | security vulnerability | ## CVE-2019-13990 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>quartz-2.2.0.jar</b></p></summary>
<p>Enterprise Job Scheduler</p>
<p>Library home page: <a href="http://www.quartz-scheduler.org">http://www.quartz-scheduler.org</a></p>
<p>Path to dependency file: shaggy-rogers/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/quartz-scheduler/quartz/2.2.0/quartz-2.2.0.jar</p>
<p>
Dependency Hierarchy:
- pantomime-2.11.0.jar (Root Library)
- tika-parsers-1.19.1.jar
- cdm-4.5.5.jar
- :x: **quartz-2.2.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mthbernardes/shaggy-rogers/commit/f72a5cb259e01c0ac208ba3a95eee5232c30fe6c">f72a5cb259e01c0ac208ba3a95eee5232c30fe6c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
initDocumentParser in xml/XMLSchedulingDataProcessor.java in Terracotta Quartz Scheduler through 2.3.0 allows XXE attacks via a job description.
<p>Publish Date: 2019-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13990>CVE-2019-13990</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/quartz-scheduler/quartz/releases/tag/v2.3.2">https://github.com/quartz-scheduler/quartz/releases/tag/v2.3.2</a></p>
<p>Release Date: 2019-12-05</p>
<p>Fix Resolution: Replace or update the following file: org.quartz-scheduler.internal:quartz-core:2.3.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.quartz-scheduler","packageName":"quartz","packageVersion":"2.2.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.novemberain:pantomime:2.11.0;org.apache.tika:tika-parsers:1.19.1;edu.ucar:cdm:4.5.5;org.quartz-scheduler:quartz:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Replace or update the following file: org.quartz-scheduler.internal:quartz-core:2.3.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-13990","vulnerabilityDetails":"initDocumentParser in xml/XMLSchedulingDataProcessor.java in Terracotta Quartz Scheduler through 2.3.0 allows XXE attacks via a job description.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13990","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-13990 (High) detected in quartz-2.2.0.jar - autoclosed - ## CVE-2019-13990 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>quartz-2.2.0.jar</b></p></summary>
<p>Enterprise Job Scheduler</p>
<p>Library home page: <a href="http://www.quartz-scheduler.org">http://www.quartz-scheduler.org</a></p>
<p>Path to dependency file: shaggy-rogers/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/quartz-scheduler/quartz/2.2.0/quartz-2.2.0.jar</p>
<p>
Dependency Hierarchy:
- pantomime-2.11.0.jar (Root Library)
- tika-parsers-1.19.1.jar
- cdm-4.5.5.jar
- :x: **quartz-2.2.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mthbernardes/shaggy-rogers/commit/f72a5cb259e01c0ac208ba3a95eee5232c30fe6c">f72a5cb259e01c0ac208ba3a95eee5232c30fe6c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
initDocumentParser in xml/XMLSchedulingDataProcessor.java in Terracotta Quartz Scheduler through 2.3.0 allows XXE attacks via a job description.
<p>Publish Date: 2019-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13990>CVE-2019-13990</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/quartz-scheduler/quartz/releases/tag/v2.3.2">https://github.com/quartz-scheduler/quartz/releases/tag/v2.3.2</a></p>
<p>Release Date: 2019-12-05</p>
<p>Fix Resolution: Replace or update the following file: org.quartz-scheduler.internal:quartz-core:2.3.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.quartz-scheduler","packageName":"quartz","packageVersion":"2.2.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.novemberain:pantomime:2.11.0;org.apache.tika:tika-parsers:1.19.1;edu.ucar:cdm:4.5.5;org.quartz-scheduler:quartz:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Replace or update the following file: org.quartz-scheduler.internal:quartz-core:2.3.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-13990","vulnerabilityDetails":"initDocumentParser in xml/XMLSchedulingDataProcessor.java in Terracotta Quartz Scheduler through 2.3.0 allows XXE attacks via a job description.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13990","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in quartz jar autoclosed cve high severity vulnerability vulnerable library quartz jar enterprise job scheduler library home page a href path to dependency file shaggy rogers pom xml path to vulnerable library home wss scanner repository org quartz scheduler quartz quartz jar dependency hierarchy pantomime jar root library tika parsers jar cdm jar x quartz jar vulnerable library found in head commit a href found in base branch master vulnerability details initdocumentparser in xml xmlschedulingdataprocessor java in terracotta quartz scheduler through allows xxe attacks via a job description publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file org quartz scheduler internal quartz core isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com novemberain pantomime org apache tika tika parsers edu ucar cdm org quartz scheduler quartz isminimumfixversionavailable true minimumfixversion replace or update the following file org quartz scheduler internal quartz core basebranches vulnerabilityidentifier cve vulnerabilitydetails initdocumentparser in xml xmlschedulingdataprocessor java in terracotta quartz scheduler through allows xxe attacks via a job description vulnerabilityurl | 0 |
225,231 | 17,835,747,200 | IssuesEvent | 2021-09-03 00:37:09 | docking-org/AMIS | https://api.github.com/repos/docking-org/AMIS | closed | Website Modfication Requests | ready to test | 

-
See Image or contact me for Clarifications:
โข Adjust png Size at bottom, rescale images to 100%,
โข Move Buttons etc.
Move TdTomato, tdTomato-RI,Dapi, Invert color, Change background color, Download Tiff file, and Sharable URL
To bottom left and add some type of frame around them
โข Move Navigation Buttons
From Middle to Bottom Center
โข Add Controls in font
Add text for โpositive controlsโ (top) and โnegative controlsโ (bottom)
โข Hide MD5 value tag from Web Access but not from Me
Add an Editor or Secret Mode?
โข Move LUT Bar
Move from Right Side to adjacent to tdTomato RI image
| 1.0 | Website Modfication Requests - 

-
See Image or contact me for Clarifications:
โข Adjust png Size at bottom, rescale images to 100%,
โข Move Buttons etc.
Move TdTomato, tdTomato-RI,Dapi, Invert color, Change background color, Download Tiff file, and Sharable URL
To bottom left and add some type of frame around them
โข Move Navigation Buttons
From Middle to Bottom Center
โข Add Controls in font
Add text for โpositive controlsโ (top) and โnegative controlsโ (bottom)
โข Hide MD5 value tag from Web Access but not from Me
Add an Editor or Secret Mode?
โข Move LUT Bar
Move from Right Side to adjacent to tdTomato RI image
| test | website modfication requests see image or contact me for clarifications โข adjust png size at bottom rescale images to โข move buttons etc move tdtomato tdtomato ri dapi invert color change background color download tiff file and sharable url to bottom left and add some type of frame around them โข move navigation buttons from middle to bottom center โข add controls in font add text for โpositive controlsโ top and โnegative controlsโ bottom โข hide value tag from web access but not from me add an editor or secret mode โข move lut bar move from right side to adjacent to tdtomato ri image | 1 |
76,392 | 7,527,274,005 | IssuesEvent | 2018-04-13 16:28:45 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | All e2e tests which create CustomResourceDefinitions are flaky. | kind/bug kind/flake priority/failing-test sig/api-machinery | **Is this a BUG REPORT or FEATURE REQUEST?**:
/kind bug
/sig api-machinery
/kind flake
/priority failing-test
**What happened**:
All the tests that rely on creating a CustomResourceDefinition are [flaky](https://k8s-testgrid.appspot.com/sig-release-master-blocking#gci-gce&width=5&include-filter-by-regex=api-machinery&include-filter-by-regex=custom%20resource) (about 6-7 flakes per day in ci). This is because the test "AdmissionWebhook Should be able to deny custom resource creation" creates a validating webhook, called ```deny-crd.k8s.io```, which denies all crd creations. This is an issue for any other test which will create a crd that gets run in parallel with that webhook test, causing them to fail with the message ```failed to create CustomResourceDefinition: admission webhook "deny-crd.k8s.io" denied the request: this webhook denies all requests```
I tried to fix this in https://github.com/kubernetes/kubernetes/pull/61609#discussion_r178959024 but it didn't work because I used a namespace selector. The tests are still flaky. I think changing the test webhook image to support handling a new "/crd" path which would accept and deny crd creations based on their labels would work (instead of a namespace selector since the crds aren't namespaced anyway). | 1.0 | All e2e tests which create CustomResourceDefinitions are flaky. - **Is this a BUG REPORT or FEATURE REQUEST?**:
/kind bug
/sig api-machinery
/kind flake
/priority failing-test
**What happened**:
All the tests that rely on creating a CustomResourceDefinition are [flaky](https://k8s-testgrid.appspot.com/sig-release-master-blocking#gci-gce&width=5&include-filter-by-regex=api-machinery&include-filter-by-regex=custom%20resource) (about 6-7 flakes per day in ci). This is because the test "AdmissionWebhook Should be able to deny custom resource creation" creates a validating webhook, called ```deny-crd.k8s.io```, which denies all crd creations. This is an issue for any other test which will create a crd that gets run in parallel with that webhook test, causing them to fail with the message ```failed to create CustomResourceDefinition: admission webhook "deny-crd.k8s.io" denied the request: this webhook denies all requests```
I tried to fix this in https://github.com/kubernetes/kubernetes/pull/61609#discussion_r178959024 but it didn't work because I used a namespace selector. The tests are still flaky. I think changing the test webhook image to support handling a new "/crd" path which would accept and deny crd creations based on their labels would work (instead of a namespace selector since the crds aren't namespaced anyway). | test | all tests which create customresourcedefinitions are flaky is this a bug report or feature request kind bug sig api machinery kind flake priority failing test what happened all the tests that rely on creating a customresourcedefinition are about flakes per day in ci this is because the test admissionwebhook should be able to deny custom resource creation creates a validating webhook called deny crd io which denies all crd creations this is an issue for any other test which will create a crd that gets run in parallel with that webhook test causing them to fail with the message failed to create customresourcedefinition admission webhook deny crd io denied the request this webhook denies all requests i tried to fix this in but it didn t work because i used a namespace selector the tests are still flaky i think changing the test webhook image to support handling a new crd path which would accept and deny crd creations based on their labels would work instead of a namespace selector since the crds aren t namespaced anyway | 1 |
148,462 | 11,853,093,301 | IssuesEvent | 2020-03-24 21:14:34 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | sql: race in the txn start time recording code | C-test-failure O-robot branch-master | [(server).TestSSLEnforcement failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1823707&tab=buildLog) on [master@7d762f83659245c4f14965366f0a1716c86f0f0c](https://github.com/cockroachdb/cockroach/commits/7d762f83659245c4f14965366f0a1716c86f0f0c):
```
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).recordTransactionStart()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:1286 +0x70
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).txnStateTransitionsApplyWrapper()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:2127 +0x778
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execCmd()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1523 +0xa76
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1323 +0x483
github.com/cockroachdb/cockroach/pkg/sql.(*InternalExecutor).initConnEx.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/internal.go:158 +0x8e
Goroutine 101 (running) created at:
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:715 +0xb8
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/internal/transport.(*http2Server).operateHeaders()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/internal/transport/http2_server.go:421 +0x1596
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/internal/transport.(*http2Server).HandleStreams()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/internal/transport/http2_server.go:461 +0x39f
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:713 +0x19a
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).handleRawConn.func1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:675 +0x50
Goroutine 105 (running) created at:
github.com/cockroachdb/cockroach/pkg/sql.(*InternalExecutor).initConnEx()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/internal.go:157 +0x5c9
github.com/cockroachdb/cockroach/pkg/sql.(*InternalExecutor).execInternal()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/internal.go:427 +0x950
github.com/cockroachdb/cockroach/pkg/sql.(*InternalExecutor).ExecEx()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/internal.go:295 +0x14f
github.com/cockroachdb/cockroach/pkg/sqlmigrations.runner.execAsRootWithRetry()
/go/src/github.com/cockroachdb/cockroach/pkg/sqlmigrations/migrations.go:371 +0x37e
github.com/cockroachdb/cockroach/pkg/sqlmigrations.optInToDiagnosticsStatReporting()
/go/src/github.com/cockroachdb/cockroach/pkg/sqlmigrations/migrations.go:862 +0x111
github.com/cockroachdb/cockroach/pkg/sqlmigrations.(*Manager).EnsureMigrations()
/go/src/github.com/cockroachdb/cockroach/pkg/sqlmigrations/migrations.go:613 +0xc3c
github.com/cockroachdb/cockroach/pkg/server.(*Server).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1776 +0x4c69
github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:384 +0x2e8
github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServerRaw()
/go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:228 +0x15f
github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServer()
/go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:196 +0x65
github.com/cockroachdb/cockroach/pkg/server.TestSSLEnforcement()
/go/src/github.com/cockroachdb/cockroach/pkg/server/authentication_test.go:79 +0xef
testing.tRunner()
/usr/local/go/src/testing/testing.go:909 +0x199
==================
FAIL github.com/cockroachdb/cockroach/pkg/server 189.608s
```
<details><summary>More</summary><p>
Parameters:
- GOFLAGS=-json
```
make stressrace TESTS=TestSSLEnforcement PKG=./pkg/server TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestSSLEnforcement.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 1.0 | sql: race in the txn start time recording code - [(server).TestSSLEnforcement failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1823707&tab=buildLog) on [master@7d762f83659245c4f14965366f0a1716c86f0f0c](https://github.com/cockroachdb/cockroach/commits/7d762f83659245c4f14965366f0a1716c86f0f0c):
```
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).recordTransactionStart()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:1286 +0x70
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).txnStateTransitionsApplyWrapper()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:2127 +0x778
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execCmd()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1523 +0xa76
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1323 +0x483
github.com/cockroachdb/cockroach/pkg/sql.(*InternalExecutor).initConnEx.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/internal.go:158 +0x8e
Goroutine 101 (running) created at:
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:715 +0xb8
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/internal/transport.(*http2Server).operateHeaders()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/internal/transport/http2_server.go:421 +0x1596
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/internal/transport.(*http2Server).HandleStreams()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/internal/transport/http2_server.go:461 +0x39f
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:713 +0x19a
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).handleRawConn.func1()
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:675 +0x50
Goroutine 105 (running) created at:
github.com/cockroachdb/cockroach/pkg/sql.(*InternalExecutor).initConnEx()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/internal.go:157 +0x5c9
github.com/cockroachdb/cockroach/pkg/sql.(*InternalExecutor).execInternal()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/internal.go:427 +0x950
github.com/cockroachdb/cockroach/pkg/sql.(*InternalExecutor).ExecEx()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/internal.go:295 +0x14f
github.com/cockroachdb/cockroach/pkg/sqlmigrations.runner.execAsRootWithRetry()
/go/src/github.com/cockroachdb/cockroach/pkg/sqlmigrations/migrations.go:371 +0x37e
github.com/cockroachdb/cockroach/pkg/sqlmigrations.optInToDiagnosticsStatReporting()
/go/src/github.com/cockroachdb/cockroach/pkg/sqlmigrations/migrations.go:862 +0x111
github.com/cockroachdb/cockroach/pkg/sqlmigrations.(*Manager).EnsureMigrations()
/go/src/github.com/cockroachdb/cockroach/pkg/sqlmigrations/migrations.go:613 +0xc3c
github.com/cockroachdb/cockroach/pkg/server.(*Server).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1776 +0x4c69
github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start()
/go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:384 +0x2e8
github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServerRaw()
/go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:228 +0x15f
github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServer()
/go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:196 +0x65
github.com/cockroachdb/cockroach/pkg/server.TestSSLEnforcement()
/go/src/github.com/cockroachdb/cockroach/pkg/server/authentication_test.go:79 +0xef
testing.tRunner()
/usr/local/go/src/testing/testing.go:909 +0x199
==================
FAIL github.com/cockroachdb/cockroach/pkg/server 189.608s
```
<details><summary>More</summary><p>
Parameters:
- GOFLAGS=-json
```
make stressrace TESTS=TestSSLEnforcement PKG=./pkg/server TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestSSLEnforcement.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| test | sql race in the txn start time recording code on github com cockroachdb cockroach pkg sql connexecutor recordtransactionstart go src github com cockroachdb cockroach pkg sql conn executor exec go github com cockroachdb cockroach pkg sql connexecutor txnstatetransitionsapplywrapper go src github com cockroachdb cockroach pkg sql conn executor go github com cockroachdb cockroach pkg sql connexecutor execcmd go src github com cockroachdb cockroach pkg sql conn executor go github com cockroachdb cockroach pkg sql connexecutor run go src github com cockroachdb cockroach pkg sql conn executor go github com cockroachdb cockroach pkg sql internalexecutor initconnex go src github com cockroachdb cockroach pkg sql internal go goroutine running created at github com cockroachdb cockroach vendor google golang org grpc server servestreams go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach vendor google golang org grpc internal transport operateheaders go src github com cockroachdb cockroach vendor google golang org grpc internal transport server go github com cockroachdb cockroach vendor google golang org grpc internal transport handlestreams go src github com cockroachdb cockroach vendor google golang org grpc internal transport server go github com cockroachdb cockroach vendor google golang org grpc server servestreams go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach vendor google golang org grpc server handlerawconn go src github com cockroachdb cockroach vendor google golang org grpc server go goroutine running created at github com cockroachdb cockroach pkg sql internalexecutor initconnex go src github com cockroachdb cockroach pkg sql internal go github com cockroachdb cockroach pkg sql internalexecutor execinternal go src github com cockroachdb cockroach pkg sql internal go github com cockroachdb cockroach pkg sql internalexecutor execex go src github com cockroachdb cockroach pkg sql internal go github com cockroachdb cockroach pkg sqlmigrations runner execasrootwithretry go src github com cockroachdb cockroach pkg sqlmigrations migrations go github com cockroachdb cockroach pkg sqlmigrations optintodiagnosticsstatreporting go src github com cockroachdb cockroach pkg sqlmigrations migrations go github com cockroachdb cockroach pkg sqlmigrations manager ensuremigrations go src github com cockroachdb cockroach pkg sqlmigrations migrations go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg server testserver start go src github com cockroachdb cockroach pkg server testserver go github com cockroachdb cockroach pkg testutils serverutils startserverraw go src github com cockroachdb cockroach pkg testutils serverutils test server shim go github com cockroachdb cockroach pkg testutils serverutils startserver go src github com cockroachdb cockroach pkg testutils serverutils test server shim go github com cockroachdb cockroach pkg server testsslenforcement go src github com cockroachdb cockroach pkg server authentication test go testing trunner usr local go src testing testing go fail github com cockroachdb cockroach pkg server more parameters goflags json make stressrace tests testsslenforcement pkg pkg server testtimeout stressflags timeout powered by | 1 |
133,389 | 29,116,750,825 | IssuesEvent | 2023-05-17 02:06:55 | Jian-Min-Huang/tech-note | https://api.github.com/repos/Jian-Min-Huang/tech-note | reopened | 148. Sort List | leetcode | Sort a linked list in O(n log n) time using constant space complexity.
**Example 1:**
```
Input: 4->2->1->3
Output: 1->2->3->4
```
**Example 2:**
```
Input: -1->5->3->4->0
Output: -1->0->3->4->5
``` | 1.0 | 148. Sort List - Sort a linked list in O(n log n) time using constant space complexity.
**Example 1:**
```
Input: 4->2->1->3
Output: 1->2->3->4
```
**Example 2:**
```
Input: -1->5->3->4->0
Output: -1->0->3->4->5
``` | non_test | sort list sort a linked list in o n log n time using constant space complexity example input output example input output | 0 |
183,768 | 14,952,488,789 | IssuesEvent | 2021-01-26 15:35:46 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | aws_transfer_user not accepting arguments introduced since v3.8.0 | documentation service/transfer | <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a ๐ [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform AWS Provider Version
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
```hcl
Terraform v0.14.3
+ provider registry.terraform.io/hashicorp/aws v3.24.0
+ provider registry.terraform.io/hashicorp/external v1.2.0
+ provider registry.terraform.io/hashicorp/helm v1.2.4
+ provider registry.terraform.io/hashicorp/kubernetes v1.12.0
+ provider registry.terraform.io/hashicorp/local v2.0.0
+ provider registry.terraform.io/hashicorp/null v3.0.0
+ provider registry.terraform.io/hashicorp/random v3.0.1
+ provider registry.terraform.io/hashicorp/tls v3.0.0
Your version of Terraform is out of date! The latest version
is 0.14.4. You can update by downloading from https://www.terraform.io/downloads.html
```
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* aws_transfer_user
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
Please include all Terraform configurations required to reproduce the bug. Bug reports without a functional reproduction may be closed without investigation.
```hcl
### Before: working fine
resource "aws_transfer_user" "sftp_user" {
for_each = local.enable_sftp ? local.user_to_home : {}
server_id = data.aws_transfer_server.existing_sftp_server[0].server_id
user_name = each.key
role = data.aws_iam_role.sftp_s3_access_role[0].arn
policy = data.aws_iam_policy_document.sftp_scoped_down_user_policy[0].json
home_directory = "/${local.sftp_bucket}/${each.value}"
tags = local.common_tags
provider = aws.sftp_provider
}
### After: BROKEN
resource "aws_transfer_user" "sftp_user" {
for_each = local.enable_sftp ? local.user_to_home : {}
server_id = data.aws_transfer_server.existing_sftp_server[0].server_id
user_name = each.key
role = data.aws_iam_role.sftp_s3_access_role[0].arn
policy = data.aws_iam_policy_document.sftp_scoped_down_user_policy[0].json
home_directory = "/${local.sftp_bucket}/${each.value}"
tags = local.common_tags
provider = aws.sftp_provider
home_directory_mappings = [] <-- this is the problem
}
```
### Debug Output
`terraform apply` fails during code validation step, so no plan is generated
I believe in this case the Debug output is irrelevant, so I copied the normal output below:
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
```hcl
Error: Unsupported argument
on modules/ec2_v2/jobs_sftp.tf line 87, in resource "aws_transfer_user" "sftp_user":
87: home_directory_mappings = []
An argument named "home_directory_mappings" is not expected here. Did you mean
to define a block of type "home_directory_mappings"?
```
### Expected Behavior
<!--- What should have happened? --->
Terraform plan should show no changes as the argument is valid since aws provider v3.8.0 and I specify an empty list which should cause no changes
### Actual Behavior
<!--- What actually happened? --->
Terraform throws an error stating that the argument is not expected
```hcl
Error: Unsupported argument
on modules/ec2_v2/jobs_sftp.tf line 87, in resource "aws_transfer_user" "sftp_user":
87: home_directory_mappings = []
An argument named "home_directory_mappings" is not expected here. Did you mean
to define a block of type "home_directory_mappings"?
```
I already had some suspicions when I saw that my IntelliJ tooltip suggestion wasn't showing the argument in the list (see below image):

### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. given a working sftp user configuration with the same structure as in the "Before" snippet
2. add the `home_directory_mappings` argument with any value
3. `terraform apply`
### Important Factoids
My steps:
1. This feature has been merged in v3.8.0 - PR https://github.com/hashicorp/terraform-provider-aws/pull/13591
2. The first time I saw this error, my codebase was configured to use AWS Provider version v3.19.0
3. I then retried after I upgraded to the current version v3.24.0, to make sure it was not some regression bug specific to v3.19.0
4. Still experiencing the same issue, I tried deleting the entire project folder and re-initializing from scratch
5. After another failure, I decided to open this issue
Worth noting that my sftp code is inside my custom module (see path mentioned in the error), but everything else works fine
<!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
* PR #13591
| 1.0 | aws_transfer_user not accepting arguments introduced since v3.8.0 - <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a ๐ [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform AWS Provider Version
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
```hcl
Terraform v0.14.3
+ provider registry.terraform.io/hashicorp/aws v3.24.0
+ provider registry.terraform.io/hashicorp/external v1.2.0
+ provider registry.terraform.io/hashicorp/helm v1.2.4
+ provider registry.terraform.io/hashicorp/kubernetes v1.12.0
+ provider registry.terraform.io/hashicorp/local v2.0.0
+ provider registry.terraform.io/hashicorp/null v3.0.0
+ provider registry.terraform.io/hashicorp/random v3.0.1
+ provider registry.terraform.io/hashicorp/tls v3.0.0
Your version of Terraform is out of date! The latest version
is 0.14.4. You can update by downloading from https://www.terraform.io/downloads.html
```
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* aws_transfer_user
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
Please include all Terraform configurations required to reproduce the bug. Bug reports without a functional reproduction may be closed without investigation.
```hcl
### Before: working fine
resource "aws_transfer_user" "sftp_user" {
for_each = local.enable_sftp ? local.user_to_home : {}
server_id = data.aws_transfer_server.existing_sftp_server[0].server_id
user_name = each.key
role = data.aws_iam_role.sftp_s3_access_role[0].arn
policy = data.aws_iam_policy_document.sftp_scoped_down_user_policy[0].json
home_directory = "/${local.sftp_bucket}/${each.value}"
tags = local.common_tags
provider = aws.sftp_provider
}
### After: BROKEN
resource "aws_transfer_user" "sftp_user" {
for_each = local.enable_sftp ? local.user_to_home : {}
server_id = data.aws_transfer_server.existing_sftp_server[0].server_id
user_name = each.key
role = data.aws_iam_role.sftp_s3_access_role[0].arn
policy = data.aws_iam_policy_document.sftp_scoped_down_user_policy[0].json
home_directory = "/${local.sftp_bucket}/${each.value}"
tags = local.common_tags
provider = aws.sftp_provider
home_directory_mappings = [] <-- this is the problem
}
```
### Debug Output
`terraform apply` fails during code validation step, so no plan is generated
I believe in this case the Debug output is irrelevant, so I copied the normal output below:
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
```hcl
Error: Unsupported argument
on modules/ec2_v2/jobs_sftp.tf line 87, in resource "aws_transfer_user" "sftp_user":
87: home_directory_mappings = []
An argument named "home_directory_mappings" is not expected here. Did you mean
to define a block of type "home_directory_mappings"?
```
### Expected Behavior
<!--- What should have happened? --->
Terraform plan should show no changes as the argument is valid since aws provider v3.8.0 and I specify an empty list which should cause no changes
### Actual Behavior
<!--- What actually happened? --->
Terraform throws an error stating that the argument is not expected
```hcl
Error: Unsupported argument
on modules/ec2_v2/jobs_sftp.tf line 87, in resource "aws_transfer_user" "sftp_user":
87: home_directory_mappings = []
An argument named "home_directory_mappings" is not expected here. Did you mean
to define a block of type "home_directory_mappings"?
```
I already had some suspicions when I saw that my IntelliJ tooltip suggestion wasn't showing the argument in the list (see below image):

### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. given a working sftp user configuration with the same structure as in the "Before" snippet
2. add the `home_directory_mappings` argument with any value
3. `terraform apply`
### Important Factoids
My steps:
1. This feature has been merged in v3.8.0 - PR https://github.com/hashicorp/terraform-provider-aws/pull/13591
2. The first time I saw this error, my codebase was configured to use AWS Provider version v3.19.0
3. I then retried after I upgraded to the current version v3.24.0, to make sure it was not some regression bug specific to v3.19.0
4. Still experiencing the same issue, I tried deleting the entire project folder and re-initializing from scratch
5. After another failure, I decided to open this issue
Worth noting that my sftp code is inside my custom module (see path mentioned in the error), but everything else works fine
<!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
* PR #13591
| non_test | aws transfer user not accepting arguments introduced since please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a ๐ to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform aws provider version hcl terraform provider registry terraform io hashicorp aws provider registry terraform io hashicorp external provider registry terraform io hashicorp helm provider registry terraform io hashicorp kubernetes provider registry terraform io hashicorp local provider registry terraform io hashicorp null provider registry terraform io hashicorp random provider registry terraform io hashicorp tls your version of terraform is out of date the latest version is you can update by downloading from affected resource s aws transfer user terraform configuration files please include all terraform configurations required to reproduce the bug bug reports without a functional reproduction may be closed without investigation hcl before working fine resource aws transfer user sftp user for each local enable sftp local user to home server id data aws transfer server existing sftp server server id user name each key role data aws iam role sftp access role arn policy data aws iam policy document sftp scoped down user policy json home directory local sftp bucket each value tags local common tags provider aws sftp provider after broken resource aws transfer user sftp user for each local enable sftp local user to home server id data aws transfer server existing sftp server server id user name each key role data aws iam role sftp access role arn policy data aws iam policy document sftp scoped down user policy json home directory local sftp bucket each value tags local common tags provider aws sftp provider home directory mappings this is the problem debug output terraform apply fails during code validation step so no plan is generated i believe in this case the debug output is irrelevant so i copied the normal output below please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the hcl error unsupported argument on modules jobs sftp tf line in resource aws transfer user sftp user home directory mappings an argument named home directory mappings is not expected here did you mean to define a block of type home directory mappings expected behavior terraform plan should show no changes as the argument is valid since aws provider and i specify an empty list which should cause no changes actual behavior terraform throws an error stating that the argument is not expected hcl error unsupported argument on modules jobs sftp tf line in resource aws transfer user sftp user home directory mappings an argument named home directory mappings is not expected here did you mean to define a block of type home directory mappings i already had some suspicions when i saw that my intellij tooltip suggestion wasn t showing the argument in the list see below image steps to reproduce given a working sftp user configuration with the same structure as in the before snippet add the home directory mappings argument with any value terraform apply important factoids my steps this feature has been merged in pr the first time i saw this error my codebase was configured to use aws provider version i then retried after i upgraded to the current version to make sure it was not some regression bug specific to still experiencing the same issue i tried deleting the entire project folder and re initializing from scratch after another failure i decided to open this issue worth noting that my sftp code is inside my custom module see path mentioned in the error but everything else works fine references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example pr | 0 |
74,365 | 14,239,671,827 | IssuesEvent | 2020-11-18 20:30:19 | LittleImprovementsCustom/LittleImprovementsCustom | https://api.github.com/repos/LittleImprovementsCustom/LittleImprovementsCustom | opened | Show the button as unclickable when no packs are selected | code enhancement | should use `--secondary-font-color` font colour and `not-allowed` cursor | 1.0 | Show the button as unclickable when no packs are selected - should use `--secondary-font-color` font colour and `not-allowed` cursor | non_test | show the button as unclickable when no packs are selected should use secondary font color font colour and not allowed cursor | 0 |
40,085 | 5,173,772,572 | IssuesEvent | 2017-01-18 16:52:15 | oppia/oppia | https://api.github.com/repos/oppia/oppia | closed | Allow learners to download a transcript of their playthrough | loc: full-stack team: learner view (@maitbayev) TODO: tech (design doc) type: feature (minor) | ```
In as much detail as possible, please describe what you would like to see.
After an exploration is played (or in the middle of an exploration), it would
be nice to be able to download a transcript of the playthrough for future
reference, in some reasonable format (HTML/PDF/plaintext).
```
Original issue reported on code.google.com by `s...@seanlip.org` on 30 May 2015 at 2:46
| 1.0 | Allow learners to download a transcript of their playthrough - ```
In as much detail as possible, please describe what you would like to see.
After an exploration is played (or in the middle of an exploration), it would
be nice to be able to download a transcript of the playthrough for future
reference, in some reasonable format (HTML/PDF/plaintext).
```
Original issue reported on code.google.com by `s...@seanlip.org` on 30 May 2015 at 2:46
| non_test | allow learners to download a transcript of their playthrough in as much detail as possible please describe what you would like to see after an exploration is played or in the middle of an exploration it would be nice to be able to download a transcript of the playthrough for future reference in some reasonable format html pdf plaintext original issue reported on code google com by s seanlip org on may at | 0 |
20,870 | 27,657,188,695 | IssuesEvent | 2023-03-12 04:21:41 | pentium3/sys_reading | https://api.github.com/repos/pentium3/sys_reading | opened | CrystalPerf: Learning to Characterize the Performance of Dataflow Computation through Code Analysis | stream processing | https://www.usenix.org/system/files/atc21-tian.pdf | 1.0 | CrystalPerf: Learning to Characterize the Performance of Dataflow Computation through Code Analysis - https://www.usenix.org/system/files/atc21-tian.pdf | non_test | crystalperf learning to characterize the performance of dataflow computation through code analysis | 0 |
314,833 | 23,538,032,841 | IssuesEvent | 2022-08-20 00:58:04 | hue891/git_web_practice | https://api.github.com/repos/hue891/git_web_practice | closed | Un commit que no sigue la convenciรณn de cรณdigo o FIX a realizar | documentation | La convenciรณn del mensaje del รบltimo commit no es la esperada:
`quitar archivos swap de vim`
Recuerde que debe tener el siguiente formato: `<Identificador de la correcciรณn>: <Comentario>`
Para realizar la correcciรณn del mensaje de commit ejecute los comandos `git commit --amend` y `git push -f`
Este issue es solo un recordatorio de la convenciรณn de comentarios en los commits y puede ser cerrado. | 1.0 | Un commit que no sigue la convenciรณn de cรณdigo o FIX a realizar - La convenciรณn del mensaje del รบltimo commit no es la esperada:
`quitar archivos swap de vim`
Recuerde que debe tener el siguiente formato: `<Identificador de la correcciรณn>: <Comentario>`
Para realizar la correcciรณn del mensaje de commit ejecute los comandos `git commit --amend` y `git push -f`
Este issue es solo un recordatorio de la convenciรณn de comentarios en los commits y puede ser cerrado. | non_test | un commit que no sigue la convenciรณn de cรณdigo o fix a realizar la convenciรณn del mensaje del รบltimo commit no es la esperada quitar archivos swap de vim recuerde que debe tener el siguiente formato para realizar la correcciรณn del mensaje de commit ejecute los comandos git commit amend y git push f este issue es solo un recordatorio de la convenciรณn de comentarios en los commits y puede ser cerrado | 0 |
153,606 | 12,153,152,231 | IssuesEvent | 2020-04-25 00:54:45 | gisellemartel/CONPASS | https://api.github.com/repos/gisellemartel/CONPASS | closed | AT-20 : (US4D - As a user, I want to find specific amenities in each building) | Acceptance Test (SPRINT 3) | **Acceptance Criteria**
In the application. I want to see bathrooms. Elevators, ramps
1. Click on the Hall building
2. The indoor map shows elevators/escalators, classrooms etc. | 1.0 | AT-20 : (US4D - As a user, I want to find specific amenities in each building) - **Acceptance Criteria**
In the application. I want to see bathrooms. Elevators, ramps
1. Click on the Hall building
2. The indoor map shows elevators/escalators, classrooms etc. | test | at as a user i want to find specific amenities in each building acceptance criteria in the application i want to see bathrooms elevators ramps click on the hall building the indoor map shows elevators escalators classrooms etc | 1 |
301,916 | 26,108,796,223 | IssuesEvent | 2022-12-27 16:37:04 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | Rewrite password policy Cypress test to permit upgrade to v4.0.0 | Automated testing โญ๏ธ Sitewide CMS Quality Assurance | ## Description
Password Policy 4.0.0 includes a complete rewrite of the configuration form, so I need to rewrite the above test to handle it correctly.
## Acceptance Criteria
- [ ] Tests pass on #11834. | 1.0 | Rewrite password policy Cypress test to permit upgrade to v4.0.0 - ## Description
Password Policy 4.0.0 includes a complete rewrite of the configuration form, so I need to rewrite the above test to handle it correctly.
## Acceptance Criteria
- [ ] Tests pass on #11834. | test | rewrite password policy cypress test to permit upgrade to description password policy includes a complete rewrite of the configuration form so i need to rewrite the above test to handle it correctly acceptance criteria tests pass on | 1 |
106,167 | 11,472,595,557 | IssuesEvent | 2020-02-09 18:11:19 | SemanticMediaWiki/SemanticMediaWiki | https://api.github.com/repos/SemanticMediaWiki/SemanticMediaWiki | closed | Elasticsearch, file ingestion, and the "Content keyword" property | documentation | As outlined in [0] we are able to fetch certain information from Elasticsearch/Tika as part of the file ingestion and make them available in the `File attachment` container.
Now, one of the properties is the `Content keyword` property but the documentation [1] isn't really forthcoming on how the data are expected to be stored in the Elasticsearch `keywords` field whether with `,`, `;`, or just separated by a space. [2] shows what is possible and based on these examples we have to conclude that it is "just" a long text field without a predetermined method (for example always uses `,` or `;` etc.) to split those terms into individual representations in SMW.
refs #4488
[0] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/src/Elastic/docs/replication.md#file-attachment
[1] https://www.elastic.co/guide/en/elasticsearch/plugins/master/using-ingest-attachment.html
[2] https://sandbox.semantic-mediawiki.org/wiki/Attribut:Content_keyword | 1.0 | Elasticsearch, file ingestion, and the "Content keyword" property - As outlined in [0] we are able to fetch certain information from Elasticsearch/Tika as part of the file ingestion and make them available in the `File attachment` container.
Now, one of the properties is the `Content keyword` property but the documentation [1] isn't really forthcoming on how the data are expected to be stored in the Elasticsearch `keywords` field whether with `,`, `;`, or just separated by a space. [2] shows what is possible and based on these examples we have to conclude that it is "just" a long text field without a predetermined method (for example always uses `,` or `;` etc.) to split those terms into individual representations in SMW.
refs #4488
[0] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/src/Elastic/docs/replication.md#file-attachment
[1] https://www.elastic.co/guide/en/elasticsearch/plugins/master/using-ingest-attachment.html
[2] https://sandbox.semantic-mediawiki.org/wiki/Attribut:Content_keyword | non_test | elasticsearch file ingestion and the content keyword property as outlined in we are able to fetch certain information from elasticsearch tika as part of the file ingestion and make them available in the file attachment container now one of the properties is the content keyword property but the documentation isn t really forthcoming on how the data are expected to be stored in the elasticsearch keywords field whether with or just separated by a space shows what is possible and based on these examples we have to conclude that it is just a long text field without a predetermined method for example always uses or etc to split those terms into individual representations in smw refs | 0 |
121,500 | 10,170,652,206 | IssuesEvent | 2019-08-08 06:11:16 | NSW-OEH-EMS-KST/grid-garage | https://api.github.com/repos/NSW-OEH-EMS-KST/grid-garage | closed | raster - to ascii | tested and working | I didn't have time to set up a test branch.. let me know if you can't replicate. When using **i268** gg code branch.. Getting the following error message when running tool using sample data grids as inputs:
_['Traceback (most recent call last):\n', ' File "D:\\gg\\grid-garage\\base\\base_tool.py", line 667, in do_iteration\n res = func(row)\n', ' File "D:\\gg\\grid-garage\\tools\\raster\\to_ascii.py", line 73, in copy\n arcpy.RasterToASCII_conversion(ras, ras_out)\n', ' File "c:\\program files (x86)\\arcgis\\desktop10.4\\arcpy\\arcpy\\conversion.py", line 255, in RasterToASCII\n raise e\n', "ExecuteError: Failed to execute. Parameters are not valid.\nERROR 000885: Output ASCII raster file: R:\\TomBarrett\\Data\\grid-garage-sample-data\\OUTPUT_ASCII_\\T0825_04_SPOT5_WoodyExtent_Armidale_sml_asc's does not have a file extension.\nFailed to execute (RasterToASCII).\n\n"] | 1.0 | raster - to ascii - I didn't have time to set up a test branch.. let me know if you can't replicate. When using **i268** gg code branch.. Getting the following error message when running tool using sample data grids as inputs:
_['Traceback (most recent call last):\n', ' File "D:\\gg\\grid-garage\\base\\base_tool.py", line 667, in do_iteration\n res = func(row)\n', ' File "D:\\gg\\grid-garage\\tools\\raster\\to_ascii.py", line 73, in copy\n arcpy.RasterToASCII_conversion(ras, ras_out)\n', ' File "c:\\program files (x86)\\arcgis\\desktop10.4\\arcpy\\arcpy\\conversion.py", line 255, in RasterToASCII\n raise e\n', "ExecuteError: Failed to execute. Parameters are not valid.\nERROR 000885: Output ASCII raster file: R:\\TomBarrett\\Data\\grid-garage-sample-data\\OUTPUT_ASCII_\\T0825_04_SPOT5_WoodyExtent_Armidale_sml_asc's does not have a file extension.\nFailed to execute (RasterToASCII).\n\n"] | test | raster to ascii i didn t have time to set up a test branch let me know if you can t replicate when using gg code branch getting the following error message when running tool using sample data grids as inputs | 1 |
376,195 | 11,139,607,727 | IssuesEvent | 2019-12-21 06:55:21 | celo-org/celo-monorepo | https://api.github.com/repos/celo-org/celo-monorepo | closed | Entered amount is hidden on โAmount Goldโ field from โBuy and Sell Goldโ screen | Priority: P1 applications bug ios qa wallet | **Frequency:** 100%
**App version:** IOS test flight build v1.5.2 (17)
**Repro on:** iPhone 7 (13.3), iPhone XS Max (13.2), iPhone 7+ (12.4)
**Repro Steps:**
1) Launch the app.
2) Tap on the Gold tab.
3) Tap on either the Bug or sell button.
4) Enter an amount on the โAmount Goldโ field.
5) Observed.
**Impact** User may
**Current Behavior:** Entered amount is hidden.
**Expected Behavior:** Enter the amount should be shown.
**Attachment:** [IOS_AmuntNotShownGoldTab.mov ](https://drive.google.com/file/d/1AskhGi-oVyYYshg1--hAwrBvVLM-jH3N/view?usp=sharing) (Bug observed at 07 sec)
| 1.0 | Entered amount is hidden on โAmount Goldโ field from โBuy and Sell Goldโ screen - **Frequency:** 100%
**App version:** IOS test flight build v1.5.2 (17)
**Repro on:** iPhone 7 (13.3), iPhone XS Max (13.2), iPhone 7+ (12.4)
**Repro Steps:**
1) Launch the app.
2) Tap on the Gold tab.
3) Tap on either the Bug or sell button.
4) Enter an amount on the โAmount Goldโ field.
5) Observed.
**Impact** User may
**Current Behavior:** Entered amount is hidden.
**Expected Behavior:** Enter the amount should be shown.
**Attachment:** [IOS_AmuntNotShownGoldTab.mov ](https://drive.google.com/file/d/1AskhGi-oVyYYshg1--hAwrBvVLM-jH3N/view?usp=sharing) (Bug observed at 07 sec)
| non_test | entered amount is hidden on โamount goldโ field from โbuy and sell goldโ screen frequency app version ios test flight build repro on iphone iphone xs max iphone repro steps launch the app tap on the gold tab tap on either the bug or sell button enter an amount on the โamount goldโ field observed impact user may current behavior entered amount is hidden expected behavior enter the amount should be shown attachment bug observed at sec | 0 |
14,674 | 8,664,957,570 | IssuesEvent | 2018-11-28 21:45:17 | keras-team/keras | https://api.github.com/repos/keras-team/keras | closed | TensorBoard Callback write_images | type:bug/performance type:tensorFlow | I want to use the TensorBoard callback to visualize my conv layer kernels. But i can only see the first conv layer kernel in TensorBoard and my Dense layers at the end. For the other conv layers i can just see the bias values and not the kernels.
Here is my sample code for the Keras model.
```
# Imports
import tensorflow as tf
import numpy as np
import os
from os import makedirs
from os.path import exists, join
from keras.datasets import mnist
import time
from keras.layers import *
from keras.activations import *
from keras.models import *
from keras.optimizers import *
from keras.initializers import *
from keras.callbacks import TensorBoard
from keras.callbacks import ModelCheckpoint
from keras.utils.np_utils import to_categorical
from plotting import *
log_dir = '"./"
# Load MNIST dataset
(x_train, y_train), (x_test, y_test) = mnist.load_data()
batch_size = 128
epochs = 10
width = 28
height = 28
depth = 1
num_classes = 10
train_size = x_train.shape[0]
test_size = x_test.shape[0]
x_train = x_train.reshape(train_size, width, height, depth)
y_train = to_categorical(y_train, num_classes=num_classes)
x_test = x_test.reshape(test_size, width, height, depth)
y_test = to_categorical(y_test, num_classes=num_classes)
tb = TensorBoard(
log_dir=log_dir,
histogram_freq=1,
write_graph=True,
write_images=True)
# Define the DNN
model = Sequential()
model.add(Conv2D(filters=16, kernel_size=3, input_shape=(width, height, depth), name="conv1"))
model.add(Activation("relu"))
model.add(Conv2D(filters=20, kernel_size=3, name="conv2"))
model.add(Activation("relu"))
model.add(MaxPool2D())
model.add(Conv2D(filters=24, kernel_size=3, name="conv3"))
model.add(Activation("relu"))
model.add(Conv2D(filters=28, kernel_size=3, name="conv4"))
model.add(Activation("relu"))
model.add(MaxPool2D())
model.add(Flatten())
model.add(Dense(128))
model.add(Activation("relu"))
model.add(Dense(num_classes, name="features"))
model.add(Activation("softmax"))
# Print the DNN layers
model.summary()
# Train the DNN
lr = 1e-3
optimizer = Adam(lr=lr)
model.compile(loss="categorical_crossentropy", optimizer=optimizer, metrics=["accuracy"])
model.fit(x_train, y_train, verbose=1, batch_size=batch_size, epochs=epochs, validation_data=(x_test, y_test), callbacks=[tb])
# Test the DNN
score = model.evaluate(x_test, y_test, batch_size=batch_size)
print("Test performance: ", score)
```
Here is the resulting screenshot from TensorBoard.

| True | TensorBoard Callback write_images - I want to use the TensorBoard callback to visualize my conv layer kernels. But i can only see the first conv layer kernel in TensorBoard and my Dense layers at the end. For the other conv layers i can just see the bias values and not the kernels.
Here is my sample code for the Keras model.
```
# Imports
import tensorflow as tf
import numpy as np
import os
from os import makedirs
from os.path import exists, join
from keras.datasets import mnist
import time
from keras.layers import *
from keras.activations import *
from keras.models import *
from keras.optimizers import *
from keras.initializers import *
from keras.callbacks import TensorBoard
from keras.callbacks import ModelCheckpoint
from keras.utils.np_utils import to_categorical
from plotting import *
log_dir = '"./"
# Load MNIST dataset
(x_train, y_train), (x_test, y_test) = mnist.load_data()
batch_size = 128
epochs = 10
width = 28
height = 28
depth = 1
num_classes = 10
train_size = x_train.shape[0]
test_size = x_test.shape[0]
x_train = x_train.reshape(train_size, width, height, depth)
y_train = to_categorical(y_train, num_classes=num_classes)
x_test = x_test.reshape(test_size, width, height, depth)
y_test = to_categorical(y_test, num_classes=num_classes)
tb = TensorBoard(
log_dir=log_dir,
histogram_freq=1,
write_graph=True,
write_images=True)
# Define the DNN
model = Sequential()
model.add(Conv2D(filters=16, kernel_size=3, input_shape=(width, height, depth), name="conv1"))
model.add(Activation("relu"))
model.add(Conv2D(filters=20, kernel_size=3, name="conv2"))
model.add(Activation("relu"))
model.add(MaxPool2D())
model.add(Conv2D(filters=24, kernel_size=3, name="conv3"))
model.add(Activation("relu"))
model.add(Conv2D(filters=28, kernel_size=3, name="conv4"))
model.add(Activation("relu"))
model.add(MaxPool2D())
model.add(Flatten())
model.add(Dense(128))
model.add(Activation("relu"))
model.add(Dense(num_classes, name="features"))
model.add(Activation("softmax"))
# Print the DNN layers
model.summary()
# Train the DNN
lr = 1e-3
optimizer = Adam(lr=lr)
model.compile(loss="categorical_crossentropy", optimizer=optimizer, metrics=["accuracy"])
model.fit(x_train, y_train, verbose=1, batch_size=batch_size, epochs=epochs, validation_data=(x_test, y_test), callbacks=[tb])
# Test the DNN
score = model.evaluate(x_test, y_test, batch_size=batch_size)
print("Test performance: ", score)
```
Here is the resulting screenshot from TensorBoard.

| non_test | tensorboard callback write images i want to use the tensorboard callback to visualize my conv layer kernels but i can only see the first conv layer kernel in tensorboard and my dense layers at the end for the other conv layers i can just see the bias values and not the kernels here is my sample code for the keras model imports import tensorflow as tf import numpy as np import os from os import makedirs from os path import exists join from keras datasets import mnist import time from keras layers import from keras activations import from keras models import from keras optimizers import from keras initializers import from keras callbacks import tensorboard from keras callbacks import modelcheckpoint from keras utils np utils import to categorical from plotting import log dir load mnist dataset x train y train x test y test mnist load data batch size epochs width height depth num classes train size x train shape test size x test shape x train x train reshape train size width height depth y train to categorical y train num classes num classes x test x test reshape test size width height depth y test to categorical y test num classes num classes tb tensorboard log dir log dir histogram freq write graph true write images true define the dnn model sequential model add filters kernel size input shape width height depth name model add activation relu model add filters kernel size name model add activation relu model add model add filters kernel size name model add activation relu model add filters kernel size name model add activation relu model add model add flatten model add dense model add activation relu model add dense num classes name features model add activation softmax print the dnn layers model summary train the dnn lr optimizer adam lr lr model compile loss categorical crossentropy optimizer optimizer metrics model fit x train y train verbose batch size batch size epochs epochs validation data x test y test callbacks test the dnn score model evaluate x test y test batch size batch size print test performance score here is the resulting screenshot from tensorboard | 0 |
434,181 | 30,445,660,541 | IssuesEvent | 2023-07-15 16:26:46 | Alarm-Siren/6502-kicad-library | https://api.github.com/repos/Alarm-Siren/6502-kicad-library | closed | Make library compatible with Kicad's Package & Content Manager (PCM) | enhancement compatibility documentation | Make library compatible with Kicad's Package & Content Manager (PCM)
Update README file with new installation instructions accordingly. | 1.0 | Make library compatible with Kicad's Package & Content Manager (PCM) - Make library compatible with Kicad's Package & Content Manager (PCM)
Update README file with new installation instructions accordingly. | non_test | make library compatible with kicad s package content manager pcm make library compatible with kicad s package content manager pcm update readme file with new installation instructions accordingly | 0 |
234,328 | 19,143,716,566 | IssuesEvent | 2021-12-02 03:53:35 | boostcampwm-2021/iOS06-MateRunner | https://api.github.com/repos/boostcampwm-2021/iOS06-MateRunner | closed | [๋จ์ ํ
์คํธ] LoginUseCase | ์ ์ test | ## ๐ฃ ์ค๋ช
- ๋น์ฆ๋์ค ๋ก์ง์ด ์ ์์ ์ผ๋ก ๋์ํ๋์ง ํ์ธํฉ๋๋ค.
```swift
protocol LoginUseCase {
var isRegistered: PublishSubject<Bool> { get set }
var isSaved: PublishSubject<Bool> { get set }
func checkRegistration(uid: String)
func saveLoginInfo(uid: String)
func saveFCMToken(of nickname: String)
}
```
## ๐ ์ฒดํฌ๋ฆฌ์คํธ
> ๊ตฌํํด์ผํ๋ ์ด์ ์ฒดํฌ๋ฆฌ์คํธ
- [ ] ๋จ์ ํ
์คํธ ์์ฑ
| 1.0 | [๋จ์ ํ
์คํธ] LoginUseCase - ## ๐ฃ ์ค๋ช
- ๋น์ฆ๋์ค ๋ก์ง์ด ์ ์์ ์ผ๋ก ๋์ํ๋์ง ํ์ธํฉ๋๋ค.
```swift
protocol LoginUseCase {
var isRegistered: PublishSubject<Bool> { get set }
var isSaved: PublishSubject<Bool> { get set }
func checkRegistration(uid: String)
func saveLoginInfo(uid: String)
func saveFCMToken(of nickname: String)
}
```
## ๐ ์ฒดํฌ๋ฆฌ์คํธ
> ๊ตฌํํด์ผํ๋ ์ด์ ์ฒดํฌ๋ฆฌ์คํธ
- [ ] ๋จ์ ํ
์คํธ ์์ฑ
| test | loginusecase ๐ฃ ์ค๋ช
๋น์ฆ๋์ค ๋ก์ง์ด ์ ์์ ์ผ๋ก ๋์ํ๋์ง ํ์ธํฉ๋๋ค swift protocol loginusecase var isregistered publishsubject get set var issaved publishsubject get set func checkregistration uid string func savelogininfo uid string func savefcmtoken of nickname string ๐ ์ฒดํฌ๋ฆฌ์คํธ ๊ตฌํํด์ผํ๋ ์ด์ ์ฒดํฌ๋ฆฌ์คํธ ๋จ์ ํ
์คํธ ์์ฑ | 1 |
88,926 | 8,180,919,542 | IssuesEvent | 2018-08-28 21:00:52 | reeFridge/spiner | https://api.github.com/repos/reeFridge/spiner | opened | Tests | test-cov | # Issue for tracking test coverage progress:
* So, we haven't any tests, yet.
## Current objective:
- [ ] Create fixtures for atlases/skeletons.
- [ ] Write tests for structures which lives/dying in Rust (e.g. Skeleton, AnimationState etc.) | 1.0 | Tests - # Issue for tracking test coverage progress:
* So, we haven't any tests, yet.
## Current objective:
- [ ] Create fixtures for atlases/skeletons.
- [ ] Write tests for structures which lives/dying in Rust (e.g. Skeleton, AnimationState etc.) | test | tests issue for tracking test coverage progress so we haven t any tests yet current objective create fixtures for atlases skeletons write tests for structures which lives dying in rust e g skeleton animationstate etc | 1 |
135,986 | 12,696,814,143 | IssuesEvent | 2020-06-22 10:41:47 | OpenMined/Threepio | https://api.github.com/repos/OpenMined/Threepio | closed | Add support Slack channel links to main readme file | Priority: 3 - Medium :unamused: Severity: 4 - Low :sunglasses: Status: Available :wave: Type: Documentation :books: | ## Where?
Main readme
## Who?
All contributors
## What?
Users of this library should have a good idea of where to get support within Slack so that the general channel does not become a place for people dropping software implementation problems. Please add the following section somewhere in your readme:
```
## Support
For support in using this library, please join the **#lib_threepio** Slack channel. If youโd like to follow along with any code changes to the library, please join the **#code_threepio** Slack channel. [Click here to join our Slack community!](https://slack.openmined.org)
```
When finished, it should look exactly as it does here: https://github.com/OpenMined/.github/blob/master/README-TEMPLATE.md#support
## Additional Context
None
| 1.0 | Add support Slack channel links to main readme file - ## Where?
Main readme
## Who?
All contributors
## What?
Users of this library should have a good idea of where to get support within Slack so that the general channel does not become a place for people dropping software implementation problems. Please add the following section somewhere in your readme:
```
## Support
For support in using this library, please join the **#lib_threepio** Slack channel. If youโd like to follow along with any code changes to the library, please join the **#code_threepio** Slack channel. [Click here to join our Slack community!](https://slack.openmined.org)
```
When finished, it should look exactly as it does here: https://github.com/OpenMined/.github/blob/master/README-TEMPLATE.md#support
## Additional Context
None
| non_test | add support slack channel links to main readme file where main readme who all contributors what users of this library should have a good idea of where to get support within slack so that the general channel does not become a place for people dropping software implementation problems please add the following section somewhere in your readme support for support in using this library please join the lib threepio slack channel if youโd like to follow along with any code changes to the library please join the code threepio slack channel when finished it should look exactly as it does here additional context none | 0 |
118,546 | 9,994,964,527 | IssuesEvent | 2019-07-11 19:00:51 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Google Auth - Copy to clipboard issues under 'OAuth client ID' setup section | [zube]: To Test kind/bug-qa team/ui | **What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
1. Login to Rancher UI
2. Go to Global --> Authentication --> Google
3. Under section `1. Configure your Google Application settings`, pointrancher/rancher#3 - `Navigate to the "Credentials" tab to create your OAuth client ID:`
**Issue#1**
The copy to clipboard for `Authorized Javascript origins:` -- On copying the contents and entering this value in Google's `Create OAuth client ID` page-- gives this error **`Invalid Origin: URIs must not contain a path or end with "/".`**
**Expected Result:**
The text present to be copied should not have "/" in the end.
**Issue#2**
The copy to clipboard for `Authorized redirect URIs:` does not copy the whole content shown on the UI.
**Expected Result:**
Copy to clipboard action must copy all of the text shown on the UI.
**Other details that may be helpful:**
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master-head
- Installation option (single install/HA): single
| 1.0 | Google Auth - Copy to clipboard issues under 'OAuth client ID' setup section - **What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
1. Login to Rancher UI
2. Go to Global --> Authentication --> Google
3. Under section `1. Configure your Google Application settings`, pointrancher/rancher#3 - `Navigate to the "Credentials" tab to create your OAuth client ID:`
**Issue#1**
The copy to clipboard for `Authorized Javascript origins:` -- On copying the contents and entering this value in Google's `Create OAuth client ID` page-- gives this error **`Invalid Origin: URIs must not contain a path or end with "/".`**
**Expected Result:**
The text present to be copied should not have "/" in the end.
**Issue#2**
The copy to clipboard for `Authorized redirect URIs:` does not copy the whole content shown on the UI.
**Expected Result:**
Copy to clipboard action must copy all of the text shown on the UI.
**Other details that may be helpful:**
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master-head
- Installation option (single install/HA): single
| test | google auth copy to clipboard issues under oauth client id setup section what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible login to rancher ui go to global authentication google under section configure your google application settings pointrancher rancher navigate to the credentials tab to create your oauth client id issue the copy to clipboard for authorized javascript origins on copying the contents and entering this value in google s create oauth client id page gives this error invalid origin uris must not contain a path or end with expected result the text present to be copied should not have in the end issue the copy to clipboard for authorized redirect uris does not copy the whole content shown on the ui expected result copy to clipboard action must copy all of the text shown on the ui other details that may be helpful environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui master head installation option single install ha single | 1 |
285,533 | 24,674,924,621 | IssuesEvent | 2022-10-18 16:13:13 | dotnet/source-build | https://api.github.com/repos/dotnet/source-build | closed | Produce RHEL8 tarballs in CI | area-ci-testing arpow-relevant | We don't currently produce RHEL8 tarballs - there are some differences here like the lack of system libunwind. We should use a supported RHEL8 Docker image to do these builds in CI - not like the snapshot RHEL7 image we use. | 1.0 | Produce RHEL8 tarballs in CI - We don't currently produce RHEL8 tarballs - there are some differences here like the lack of system libunwind. We should use a supported RHEL8 Docker image to do these builds in CI - not like the snapshot RHEL7 image we use. | test | produce tarballs in ci we don t currently produce tarballs there are some differences here like the lack of system libunwind we should use a supported docker image to do these builds in ci not like the snapshot image we use | 1 |
442,866 | 12,752,007,507 | IssuesEvent | 2020-06-27 14:14:32 | wso2/kubernetes-apim | https://api.github.com/repos/wso2/kubernetes-apim | opened | [3.2.0] Test Helm Resources With Custom Key Stores | Priority/High Type/Task | **Description:**
WSO2 product Helm Charts are made available with the default product key stores.
Since, WSO2 Helm Charts are recommended for production grade deployments, it is essential to test these resources with custom key stores, as well.
**Affected Product Version:**
Helm Resources for WSO2 API Management version `3.1.0.1` and below | 1.0 | [3.2.0] Test Helm Resources With Custom Key Stores - **Description:**
WSO2 product Helm Charts are made available with the default product key stores.
Since, WSO2 Helm Charts are recommended for production grade deployments, it is essential to test these resources with custom key stores, as well.
**Affected Product Version:**
Helm Resources for WSO2 API Management version `3.1.0.1` and below | non_test | test helm resources with custom key stores description product helm charts are made available with the default product key stores since helm charts are recommended for production grade deployments it is essential to test these resources with custom key stores as well affected product version helm resources for api management version and below | 0 |
181,648 | 14,072,698,449 | IssuesEvent | 2020-11-04 02:35:18 | red/red | https://api.github.com/repos/red/red | closed | crash on reading a URL without scheme | status.built status.tested type.bug | **Describe the bug**
`read` and `load` crash when called on a `url!` that denotes an address with a non-existent scheme.
**To reproduce**
```red
read to url! "www.example.org"
load make url! 0
```
```red
*** Runtime Error 1: access violation
*** in file: .../runtime/datatypes/object.reds
*** at line: 109
***
*** stack: red/object/get-values 02D32574h
*** stack: red/port/get-actors 02D32574h
*** stack: red/port/open 02D32574h false false false false 00541FA4h
*** stack: red/port/read 02D32574h 02D32564h 02D32564h false false false 02D32564h
*** stack: red/url/read 02D32574h 02D32564h 02D32564h false false false 02D32564h
*** stack: red/actions/read 02D32574h 02D32564h 02D32564h false false false 02D32564h
*** stack: red/actions/read* -1 -1 -1 -1 -1 -1
```
**Expected behavior**
An error; perhaps in the former case HTTPS should be assumed as a reasonable default.
**Platform version**
df54dcd, W10. | 1.0 | crash on reading a URL without scheme - **Describe the bug**
`read` and `load` crash when called on a `url!` that denotes an address with a non-existent scheme.
**To reproduce**
```red
read to url! "www.example.org"
load make url! 0
```
```red
*** Runtime Error 1: access violation
*** in file: .../runtime/datatypes/object.reds
*** at line: 109
***
*** stack: red/object/get-values 02D32574h
*** stack: red/port/get-actors 02D32574h
*** stack: red/port/open 02D32574h false false false false 00541FA4h
*** stack: red/port/read 02D32574h 02D32564h 02D32564h false false false 02D32564h
*** stack: red/url/read 02D32574h 02D32564h 02D32564h false false false 02D32564h
*** stack: red/actions/read 02D32574h 02D32564h 02D32564h false false false 02D32564h
*** stack: red/actions/read* -1 -1 -1 -1 -1 -1
```
**Expected behavior**
An error; perhaps in the former case HTTPS should be assumed as a reasonable default.
**Platform version**
df54dcd, W10. | test | crash on reading a url without scheme describe the bug read and load crash when called on a url that denotes an address with a non existent scheme to reproduce red read to url load make url red runtime error access violation in file runtime datatypes object reds at line stack red object get values stack red port get actors stack red port open false false false false stack red port read false false false stack red url read false false false stack red actions read false false false stack red actions read expected behavior an error perhaps in the former case https should be assumed as a reasonable default platform version | 1 |
451,038 | 13,023,367,811 | IssuesEvent | 2020-07-27 09:52:54 | aeternity/jitsi-meet | https://api.github.com/repos/aeternity/jitsi-meet | closed | Update texts for Superhero Meet | priority/high | <img width="571" alt="Screenshot 2020-07-27 at 10 27 03" src="https://user-images.githubusercontent.com/13139371/88515071-c528a800-cff3-11ea-93eb-2c3511c278d0.png">
This should be: "Social good is even better when done together" | 1.0 | Update texts for Superhero Meet - <img width="571" alt="Screenshot 2020-07-27 at 10 27 03" src="https://user-images.githubusercontent.com/13139371/88515071-c528a800-cff3-11ea-93eb-2c3511c278d0.png">
This should be: "Social good is even better when done together" | non_test | update texts for superhero meet img width alt screenshot at src this should be social good is even better when done together | 0 |
453,334 | 13,068,066,256 | IssuesEvent | 2020-07-31 02:26:35 | dr4fters/dr4ft | https://api.github.com/repos/dr4fters/dr4ft | opened | Feature Request: support drafting N cards per pick | feature request high priority | ## Detailed Description
<!-- Explain what you exactly miss and how the feature could look like -->
With Double Masters, wizards has added that for the first pick of each pack, you get two cards. So it would be 2,1,1,1,...,1 totaling 15 cards selected in 14 picks.
Goal is to support this draft pardim for double masters packs that get drafted.
Stretch goal is A way to allow users to set how many cards per pick (input field would be a list of numbs like I had above)
## Further Information
<!-- Feel free to provide e.g. screenshots or mockups -->
| 1.0 | Feature Request: support drafting N cards per pick - ## Detailed Description
<!-- Explain what you exactly miss and how the feature could look like -->
With Double Masters, wizards has added that for the first pick of each pack, you get two cards. So it would be 2,1,1,1,...,1 totaling 15 cards selected in 14 picks.
Goal is to support this draft pardim for double masters packs that get drafted.
Stretch goal is A way to allow users to set how many cards per pick (input field would be a list of numbs like I had above)
## Further Information
<!-- Feel free to provide e.g. screenshots or mockups -->
| non_test | feature request support drafting n cards per pick detailed description with double masters wizards has added that for the first pick of each pack you get two cards so it would be totaling cards selected in picks goal is to support this draft pardim for double masters packs that get drafted stretch goal is a way to allow users to set how many cards per pick input field would be a list of numbs like i had above further information | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.